Neural Networks and Physical Systems with Emergent Collective Computational Abilities¶
Motivation(s)¶
All modeling is based on details. Current neuroscience theories focus on the dynamic electrochemical properties neurons and synapses at the scale of a few neurons to analyze elementary biological behavior. However, in many physical systems (e.g. a magnetic system), the nature of collective properties is insensitive to the details inserted in the model. This makes one wonder whether analogous collective phenomena in a system of simple interacting neurons have useful computational correlates.
Proposed Solution(s)¶
The author observed that any physical system whose dynamics in phase space is dominated by a substantial number of locally stable states to which it is attracted can be regarded as a general content-addressable memory. The system will be useful if any prescribed set of states can be made the stable states of the system.
To store a set of states \(\left\{ V^s \in \{0, 1\} \right\}_{s = 1}^n\), the author defines a complete graph whose weights use the storage prescription
with \(T_{ii} = 0\). Notice that these particular weights can be interpreted as a form of Hebbian learning. Unlike previous neural nets that emphasize relative timing of action potential spikes, each neuron in this network updates asynchronously.
Evaluation(s)¶
Monte Carlo calculations were made on systems of \(N = 30\) and \(N = 100\) neurons; both exhibited qualitative similarities.
The flow in phase space produced by this neural network has the properties necessary for a physical content-addressable memory whether or not \(T_{ij}\) is symmetric. Setting \(T_{ij} \neq 0\) and \(T_{ji} = 0\) did increase the probability of making errors, but the algorithm continued to generate stable minima. However, asymmetric \(T_{ij}\) can lead to a metastable minimum.
Furthermore, the phase space flow is dominated by attractors which are the nominally assigned memories, each of which dominates a substantial region around it. Memories too close to each other are confused and tend to merge. The flow is not entirely deterministic, and the system responds to an ambiguous starting state by a statistical choice between the memory states it most resembles.
A serious limitation of this model is its capacity. About \(\frac{N}{4 \log N}\) states can be simultaneously remembered before severe recall error.
Future Direction(s)¶
How to applying Hebbian learning to the hidden units within a single layer to avoid co-adaptations of neurons?
Maxout networks is capable of learning the desired activation function, but may still encountered the exploding gradient problem. Would a neuron reset scheme be an appropriate mechanism to simulate cell death and respawn?
Question(s)¶
Is the analysis of the effect of the noise terms equivalent to the derivation of model capacity in [Snab]?
The pseudo-orthogonality condition is not obvious by inspection.
Analysis¶
The spontaneous emergence of new computational capabilities from the collective behavior of large numbers of simple processing elements have been shown to be feasible. While this and asynchronous parallel processing are attractive ideas, [Snaa] proved that the time till convergence is too long if each neuron is sequentially updated. The idea is interesting and deserves further research, but this network model is not practical compared to existing techniques that are resilient to failures.
One of the interesting points is the mention of Hebbian learning: neurons that fire together wire together. This particular learning paradigm has been ignored with respect to the hidden units in multilayer feedforward networks.
Notes¶
Ising Model¶
Section 13.2.3 of [Roj13] states that the Ising model can be used to describe systems made of particles capable of adopting one of two states. In the case of ferromagnetic materials, their atoms can be modeled as particles of spin up (\(s = +1 / 2\)) or spin down (\(s = -1 / 2\)). The spin points in the direction of the magnetic field. All tiny magnets interact with each other. This causes some of the atoms to flip their spin until equilibrium is reached and the total magnetization of the material reaches a constant level, which is the sum of the individual spins.
The total magnetic field \(h_i\) sensed by the atom \(i\) in an ensemble of particles is the sum of the fields induced by each atom and the external field \(h^*\) (if present). Formally,
where \(w_{ij}\) represents the magnitude of the magnetic coupling between the atoms labeled \(i\) and \(j\). The magnetic coupling changes according to the distance between atoms and the magnetic permeability of the environment. The potential energy \(E\) of a certain state \((s_1, s_2, \ldots, s_n)\) of an Ising material has the form
In paramagnetic materials, the coupling constants are zero. In ferromagnetic materials, the constants \(w_{ij}\) are all positive. At zero temperature, both systems behave deterministically. Notice that this potential energy at zero temperature is isomorphic to the energy function of Hopfield networks. By inspection, a Hopfield network can be seen as a microscopic model of magnetism [Kru]:
physical |
neural |
---|---|
atom |
neuron |
magnetic spin |
activation state |
strength of external field |
threshold value |
magnetic coupling between atoms |
connection weights |
Hopfield Network¶
The input to a particular neuron arises from the current leaks of the synapses to that neuron. The synapses in turn are activated by arriving action potentials that neurons emit. Neurons are capable of propagating pulses of action potentials—electrochemical activity—when the average potential across their membrane is held well above its normal resting value. The mean rate at which action potentials are generated for a typical neuron \(i\) can be modeled as
where the pre-synaptic sum represents the input signal to the neuron, \(T_{ij}\) is the effectiveness of a synapse, and \(V_\cdot\) is the membrane potential (Volts).
A Hopfield network uses a simpler model of a neuron called the McCulloch-Pitts linear threshold unit. \(V_\cdot \in \{0, 1\}\) now denotes whether the neuron fired or not and is defined as
Notice that all neurons are input as well as output neurons i.e. there are no hidden units. To account for the delays in synaptic transmission and in the transmission of impulses along axons and dendrites, each neuron asynchronously readjusts its state randomly in time but with a mean attempt rate \(W\).
The energy function of a Hopfield network is
Convergence of Hopfield Network¶
By inspection, the change in energy \(\Delta E\) due to \(\Delta V_i = V_i(t + 1) - V_i(t)\) is
Observe that if \(V_i(t + 1) = V_i(t)\), the \(\Delta E = 0\).
To show the convergence of the network, assume \(T_{ij} = T_{ji}\), \(T_{ii} = 0\), and only one update occurs at a time.
When \(\Delta V_i = 1\),
because \(\sum_j T_{ij} V_j(t) > 0\).
When \(\Delta V_i = -1\),
because \(\sum_j T_{ij} V_j(t) \leq 0\).
References
- Hop82
John J Hopfield. Neural networks and physical systems with emergent collective computational abilities. Proceedings of the national academy of sciences, 79(8):2554–2558, 1982.
- Kru
Rudolf Kruse. Neural networks: hopfield networks. http://fuzzy.cs.ovgu.de/ci/nn/v07_hopfield_en.pdf. Accessed slide 15 on 2017-02-15.
- Orh
Emin Orhan. The hopfield model. http://www.cns.nyu.edu/ eorhan/notes/hopfield.pdf. Accessed on 2017-02-15.
- Roj13
Raúl Rojas. Neural Networks: A Systematic Introduction. Springer Science & Business Media, 2013.
- Snaa
Robert Snapp. Cs 256: neural computation lecture notes on hopfield networks. http://www.cems.uvm.edu/ rsnapp/teaching/cs256/lectures/hopfield.pdf. Accessed on 2017-02-15.
- Snab
Robert Snapp. Cs 256: neural computation lecture notes on hopfield networks (ii). http://www.cems.uvm.edu/ rsnapp/teaching/cs256/lectures/hopfield2.pdf. Accessed on 2017-02-15.