- Morphogenic network/Inspiration for morphogenic network
=Inspiration for morphogenic network=
Genetic algorithm with branching
This theory was inspired by an attempt to apply the
genetic algorithm to aclassifier system . The traditional genetic algorithm fails to take into account that thegene pool is not fully mixed; that a cat won’t mate with a dog, indeed, can’t mate with a dog because thephenotype s are incompatible. In general, the more “local” the phenotypes are, the higher the chance of sexual reproduction. This is apt enough, as mating a refrigerator with a car won’t do much good, but mating roller skates with ice skates might produce roller blades.Morphogenesis
The
morphospace ofevolution of a given gene is not “flat”;invariant , it is approximately defined by the genes of it’s nearest neighbors. That is, if a bit string is 111010, and there are fivebit strings of 111000, two of 111011, and one of each of the other four strings onehamming distance away (differing by one bit), the a posteri probability of the bit evolving to 111000 is 5/(4+2+5=11). The difference in information this represents is –log(5/11), (not –log(1/6)). A string should evolve by taking arandom walk , at a constant bit rate, through a posteri (not a priori)probability space.In a fixed-
topology neural network , the problem is posed differently: neurons cannot die or reproduce. However, it is still entirely possible for each neuron to evolve according to an a posteri probability space. The surrounding neurons represent different directions instate space . The uncorrelated variance, or “information deviation”, of their output signals represents their hamming distance from each other in a priori state space. Their relative “health” determines the population growth rate of strings in that state.Far-from equilibrium systems ;non-equilibrium thermodynamics Later contributing to the idea was the concept of
dissipative structure s in non-equilibrium thermodynamics. This concept includesGibbs free energy of formation, which, is related, aptly enough, to information entropy, throughthermodynamic entropy. As the ensemble of neurons evolves towards the a priori probability density, it thereby evolves towards a lower-energy state, towards “thermodynamic equilibrium ”.Self-organizing map s.This theory can be viewed as an extension to Kohonen’s self-organizing maps, where said extension is addition of a cofactor of information deviation to the lateral feedback.
Two advantages that this cofactor presents are:
*It allows information to flow faster: It allows neurons with different output signals to exist physically close to each-other without interfering with each-other. A neuron “on top” of them would take the outputs of the neurons as input. If the received signals are more disparate, the receiving neuron is ipso facto receiving more information.
*It allows neurons to cooperatively explore state-space: If neurons physically close to each-other have similar output signals, they will follow the local energy field among them to lower energy-states, like a swarm – this I owe also to the concept ofemergence andswarm intelligence .Nonlinear Dynamics To nonlinear dynamics, this idea owes much: the concept of
phase space ,attractor s, the contraction of volume within phase space. Nonlinear dynamics also helped me figure out what assumptions to make about the nature of the signals processed by the neural network. Also important is an insight fromJohn Nash : nonlinear phase space can be locally linearized. How do you solve a nonlinear problem with a neural network? By solving an ensemble of locally linear problems!Nonequilibrium statistical mechanics
This idea is, in retrospect, inspired by the belief that an information-theoretic approach to
non-equilibrium statistical mechanics (the empirical physics of complex systems such as ecologies, after dispensing with the distinction between "living" and "not living" matter) should provide the necessary formalism for an ideal artificial neural network.The reasoning for applying principles of nonequilibrium statistical mechanics to the design of artificial neural networks is thus: Insofar as the neural network is coupled with a "physical" system, and processes "information" from that system, and is "itself" a physical system, an "information" theoretic formalism of "physics" is the appropriate formalism to use when designing a neural network.
Wikimedia Foundation. 2010.