Models of neural computation

Models of neural computation

Models of neural computation are attempts to elucidate, in an abstract and mathematical fashion, the core principles that underlie information processing in biological nervous systems, or functional components thereof. This article aims to provide an overview of the most definitive models of neuro-biological computation as well as the tools commonly used to construct and analyze them.

Contents

Introduction

Due to the complexity of nervous system behavior, the associated experimental error bounds are ill-defined, but the relative merit of the different models of a particular subsystem can be compared according to how closely they reproduce real-world behaviors or respond to specific input signals. In the closely related field of computational neuroethology, the practice is to include the environment in the model in such a way that the loop is closed. In the cases where competing models are unavailable, or where only gross responses have been measured or quantified, a clearly formulated model can guide the scientist in designing experiments to probe biochemical mechanisms or network connectivity.

In all but the simplest cases, the mathematical equations that form the basis of a model cannot be solved exactly. Nevertheless, computer technology, sometimes in the form of specialized software or hardware architectures, allow scientists to perform iterative calculations and search for plausible solutions. A computer chip or a robot that can interact with the natural environment in ways akin to the original organism is one embodiment of a useful model. The ultimate measure of success is however the ability to make testable predictions.

General criteria for evaluating models

Speed of information processing

The rate of information processing in biological neural systems are constrained by the speed at which an action potential can propagate down a nerve fibre. This conduction velocity ranges from 1 m/s to over 100 m/s, and generally increases with the diameter of the neuronal process. Slow in the timescales of biologically-relevant events dictated by the speed of sound or the force of gravity, the nervous system overwhelmingly prefers parallel computations over serial ones in time-critical applications.

Robustness

A model is robust if it continues to produce the same computational results under small variations in inputs or operating parameters. For example, the direction of motion as computed by a robust motion detector would not change under small changes of luminance, contrast or velocity jitter.

Gain control

This refers to the principle that the response of a nervous system should stay within certain bounds even as the inputs from the environment change drastically. For example, when adjusting between a sunny day and a moonless night, the retina changes the relationship between light level and neuronal output by a factor of more than 106 so that the signals sent to later stages of the visual system always remain within a much narrower range of amplitudes.[1].

Linearity versus nonlinearity

A linear system is one whose response in a specified unit of measure, to a set of inputs considered at once, is the sum of its responses due to the inputs considered individually.

Linear systems are easier to analyze mathematically. Linearity may occur in the basic elements of a neural circuit such as the response of a postsynaptic neuron, or as an emergent property of a combination of nonlinear subcircuits [2].

Examples

A computational neural model may be constrained to the level of biochemical signalling in individual neurons or it may describe an entire organism in its environment. The examples here are grouped according to their scope.

Models of information transfer in neurons

The most widely used models of information transfer in biological neurons are based on analogies with electrical circuits. The equations to be solved are time-dependent differential equations with electro-dynamical variables such as current, conductance or resistance, capacitance and voltage.

Hodgkin–Huxley model and its derivatives

The Hodgkin–Huxley model, widely regarded as one of the great achievements of 20th-century biophysics, describes how action potentials in neurons are initiated and propagated in axons via voltage-gated ion channels. It is a set of nonlinear ordinary differential equations that were introduced by Alan Lloyd Hodgkin and Andrew Huxley in 1952 to explain the results of voltage clamp experiments on the squid giant axon. Analytic solutions do not exist, but the Levenberg–Marquardt algorithm, a modified Gauss–Newton algorithm, is often used to fit these equations to voltage-clamp data.

The FitzHugh–Nagumo model is a simplication of the Hodgkin–Huxley model. The Hindmarsh–Rose model is an extension which describes neuronal spike bursts. The Morris–Lecar model is a modification which does not generate spikes, but describes slow-wave propagation, which is implicated in the inhibitory synaptic mechanisms of central pattern generators.

Solitons

Transfer functions and linear filters

This approach, influenced by control theory and signal processing, treats neurons and synapses as time-invariant entities that produce outputs that are linear combinations of input signals, often depicted as sine waves with a well-defined temporal or spatial frequencies.

The entire behavior of a neuron or synapse are encoded in a transfer function, lack of knowledge concerning the exact underlying mechanism notwithstanding. This brings a highly developed mathematics to bear on the problem of information transfer.

The accompanying taxonomy of linear filters turns out to be useful in characterizing neural circuitry. Both low- and high-pass filters are postulated to exist in some form in sensory systems, as they act to prevent information loss in high and low contrast environments, respectively.

Indeed, measurements of the transfer functions of neurons in the horseshoe crab retina according to linear systems analysis show that they remove short-term fluctuations in input signals leaving only the long-term trends, in the manner of low-pass filters. These animals are unable to see low-contrast objects without the help of optical distortions caused by underwater currents [2][3].

Models of computations in sensory systems

Lateral inhibition in the retina: Hartline–Ratliff equations

In the retina, an excited neural receptor can suppress the activity of surrounding neurons within an area called the inhibitory field. This effect, known as lateral inhibition, increases the contrast and sharpness in visual response, but leads to the epiphenomenon of Mach bands. This is often illustrated by the optical illusion of light or dark stripes next to a sharp boundary between two regions in an image of different luminance.

The Hartline-Ratliff model describes interactions within a group of p photoreceptor cells[4]. Assuming these interactions to be linear, they proposed the following relationship for the steady-state response rate rp of the given p-th photoreceptor in terms of the steady-state response rates rj of the j surrounding receptors:

r_{p}=\left|\left[e_{p}-\sum_{j=1,j\ne p}^{n}k_{pj}\left|r_{j}-r_{pj}^{o}\right|\right]\right|.

Here,

ep is the excitation of the target p-th receptor from sensory transduction

r_{pj}^o is the associated threshold of the firing cell, and

kpj is the coefficient of inhibitory interaction between the p-th and the jth receptor. The inhibitory interaction decreases with distance from the target p-th receptor.

Cross-correlation in sound localization: Jeffress model

According to Jeffress[5], in order to compute the location of a sound source in space from interaural time differences, an auditory system relies on delay lines: the induced signal from an ipsilateral auditory receptor to a particular neuron is delayed for the same time as it takes for the original sound to go in space from that ear to the other. Each postsynaptic cell is differently delayed and thus specific for a particular inter-aural time difference. This theory is equivalent to the mathematical procedure of cross-correlation.

Following Fischer and Anderson[6], the response of the postsynaptic neuron to the signals from the left and right ears is given by

y_{R}\left(t\right) - y_{L}\left(t\right)

where

y_{L}\left(t\right)=\int_{0}^{\tau}u_{L}\left(\sigma\right)w\left(t-\sigma\right)d\sigma

y_{R}\left(t\right)=\int_{0}^{\tau}u_{R}\left(\sigma\right)w\left(t-\sigma\right)d\sigma

and

w\left(t-\sigma\right) represents the delay function. This is not entirely correct and a clear eye is needed to put the symbols in order.

Structures have been located in the barn owl which are consistent with Jeffress-type mechanisms.[7]

Cross-correlation for motion detection: Hassenstein–Reichardt model

A motion detector needs to satisfy three general requirements: pair-inputs, asymmetry and nonlinearity.[8] The cross-correlation operation implemented asymmetrically on the responses from a pair of photoreceptors satisfies these minimal criteria, and furthermore, predicts features which have been observed in the response of neurons of the lobula plate in bi-wing insects [9].

The master equation for response is

R = A1(t − τ)B2(t) − A2(t − τ)B1(t)

The HR model predicts a peaking of the response at a particular input temporal frequency. The conceptually similar Barlow–Levick model is deficient in the sense that a stimulus presented to only one receptor of the pair is sufficient to generate a response. This is unlike the HR model, which requires two correlated signals delivered in a time ordered fashion. However the HR model does not show a saturation of response at high contrasts, which is observed in experiment. Extensions of the Barlow-Levick model can provide for this discrepancy[10].

Watson–Ahumada model for motion estimation in humans

This uses a cross-correlation in both the spatial and temporal directions, and is related to the concept of optical flow. [11]

Neurophysiological metronomes: neural circuits for pattern generation

Mutually inhibitory processes are an unifying motif of all central pattern generators. This has been demonstrated in the stomatogastric (STG) nervous system of crayfish and lobsters[12]. Two and three-cell oscillating networks based on the STG have been constructed which are amenable to mathematical analysis, and which depend in a simple way on on synaptic strengths and overall activity, presumably the knobs on these things.[13]. The mathematics involved is the theory of dynamical systems.

Anti-Hebbian adaptation: spike-timing dependent plasticity

[14]

[15]

Feedback and control: models of flight control in the fly

Flight control in the fly is believed to be mediated by inputs from the visual system and also the halteres, a pair of knob-like organs which measure angular velocity. Integrated computer models of Drosophila, short on neuronal circuitry but based on the general guidelines given by control theory and data from the tethered flights of flies, have been constructed to investigate the details of flight control [1][2].

Software modelling approaches and tools

Neural networks

In this approach the strength and type, excitatory or inhibitory, of synaptic connections are represented by the magnitude and sign of weights, that is, numerical coefficients w' in front of the inputs x to a particular neuron. The response of the j-th neuron is given by a sum of nonlinear, usually "sigmoidal" functions g of the inputs as:

f_{j}=\sum_{i}g\left(w_{ji}'x_{i}+b_{j}\right).

This response is then fed as input into other neurons and so on. The goal is to solve for the weights given a desired response at the output layer to a set given inputs at the input layer.

Genetic algorithms

Genetic algorithms are used to evolve neural (and sometimes body) properties in a model brain-body-environment system so as to exhibit some desired behavioral performance. The evolved agents can then be subjected to a detailed analysis to uncover their principles of operation. Evolutionary approaches are particularly useful for exploring spaces of possible solutions to a given behavioral task because these approaches minimize a priori assumptions about how a given behavior ought to be instantiated. They can also be useful for exploring different ways to complete a computational neuroethology model when only partial neural circuitry is available for a biological system of interest [3].

MATLAB

MATLAB is a programming environment that is used globally in virtually all neuroscience and cognitive psychology laboratories [4]. MATLAB integrates the modelling and experimental processes by bringing together, under the aegis of an intuitive scripting language, powerful data analysis and mathematical modelling tools.

NEURON

NEURON, developed at Duke University, is a simulation environment for modeling individual neurons and networks of neurons [5]. With the NEURON interface, it is possible to generate publication-quality results without having to write any program code at all. The NEURON simulation engine is based on a Hodgkin–Huxley type model with a Borg–Graham formulation.

Embodiment in electronic hardware

Conductance-based silicon neurons

Nervous systems differ from the majority of silicon-based computing devices in that they are analog, not digital, and massively parallel, not sequential. To model nervous systems accurately, in real-time, alternative hardware is required.

The most realistic circuits to date make use of analog properties of existing digital electronics (operated under non-standard conditions) to realize Hodgkin–Huxley-type models in silico[16][6].

Retinomorphic chips

[17]

See also

References

  1. ^ Nicholas J. Priebe and David Ferster (2002). "A New Mechanism for Neuronal Gain Control". 35. Neuron. pp. 602–604. 
  2. ^ a b Alyosha Molnar, Hain-Ann Hsueh, Botond Roska, Frank S. Werblin, (2009). "Crossover inhibition in the retina: circuitry that compensates for nonlinear rectifying synaptic transmission". J Comput Neurosci 27 (3): 569–590. doi:10.1007/s10827-009-0170-6. PMC 2766457. PMID 19636690. http://www.pubmedcentral.nih.gov/articlerender.fcgi?tool=pmcentrez&artid=2766457. 
  3. ^ Robert B. Barlow, James M. Hitt and Frederick A. Dodge (2001). "Limulus Vision in the Marine Environment". The Biological Bulletin 200 (2): 169–176. doi:10.2307/1543311. JSTOR 1543311. PMID 11341579. 
  4. ^ K. P. Hadeler and D. Kuhn (1987). "Stationary States of the Hartline–Ratliff Model". Biological Cybernetics 56 (5–6): 411–417. doi:10.1007/BF00319520. 
  5. ^ Jeffress, L.A., 1948. A place theory of sound localization. Journal of Comparative and Physiological Psychology 41, 35–39.
  6. ^ Brian J. Fischer and Charles H. Anderson, 2004. A computational model of sound localization in the barn owl Neurocomputing" 58–60 (2004) 1007–1012
  7. ^ Catherine E. Carr, 1993. Delay Line Models of Sound Localization in the Barn Owl "American Zoologist" Vol. 33, No. 1 79–85
  8. ^ Borst A, Egelhaaf M., 1989. Principles of visual motion detection. "Trends in Neuroscience" 12(8):297–306
  9. ^ Joesch, M. et al. (2008) Response properties of motion-sensitive visual interneurons in the lobula plate of Drosophila melanogaster. Curr. Biol. 18, 368–374
  10. ^ Gonzalo G. de Polavieja, 2006. Neuronal Algorithms That Detect the Temporal Order of Events "Neural Computation" 18 (2006) 2102–2121
  11. ^ Andrew B. Watson and Albert J. Ahumada, Jr., 1985. Model of human visual-motion sensing "J. Opt. Soc. Am. A" 2(2) 322–341
  12. ^ Michael P. Nusbaum and Mark P. Beenhakker, A small-systems approach to motor pattern generation, Nature 417, 343–350 (16 May 2002)
  13. ^ Cristina Soto-Treviño, Kurt A. Thoroughman and Eve Marder, L. F. Abbott, 2006. Activity-dependent modification of inhibitory synapses in models of rhythmic neural networks Nature Vol 4 No 3 2102–2121
  14. ^ Tzounopoulos T, Kim Y, Oertel D, Trussell LO (2004) Cell-specific, spike timing-dependent plasticities in the dorsal cochlear nucleus. Nat Neurosci 7:719–725
  15. ^ Patrick D. Roberts, Christine V. Portfors (2008) Cell-specific, spike timing-dependent plasticities in the dorsal cochlear nucleus. Biological Cybernetics 98:491–507
  16. ^ L. Alvadoa, J. Tomasa, S. Saghia, S. Renauda, T. Balb, A. Destexheb, G. Le Masson, 2004. Hardware computation of conductance-based neuron models. Neurocomputing 58–60 (2004) 109 – 115
  17. ^ Kwabena Boahen, "A Retinomorphic Chip with Parallel Pathways: Encoding INCREASING, ON, DECREASING, and OFF Visual Signals", Analog Integrated Circuits and Signal Processing, 30, 121–135, 2002

External links


Wikimedia Foundation. 2010.

Игры ⚽ Нужен реферат?

Look at other dictionaries:

  • Interdisciplinary Center for Neural Computation — Infobox University name =The Interdisciplinary Center for Neural Computation native name =המרכז הבינתחומי לחישוביות עצבית established =1992 type =Public endowment = staff = provost = principal = rector = chancellor = vice chancellor = dean =… …   Wikipedia

  • Neural coding — is a neuroscience related field concerned with how sensory and other information is represented in the brain by networks of neurons. The main goal of studying neural coding is to characterize the relationship between the stimulus and the… …   Wikipedia

  • Neural network — For other uses, see Neural network (disambiguation). Simplified view of a feedforward artificial neural network The term neural network was traditionally used to refer to a network or circuit of biological neurons.[1] The modern usage of the term …   Wikipedia

  • Neural ensemble — A neural ensemble is a population of nervous system cells (or cultured neurons) involved in a particular neural computation. Contents 1 Background 2 Encoding 3 Location and function 4 …   Wikipedia

  • Neural modeling fields — Neural modeling field (NMF) is a mathematical framework for machine learning which combines ideas from neural networks, fuzzy logic, and model based recognition. It has also been referred to as modeling fields, modeling fields theory (MFT),… …   Wikipedia

  • Computation — is defined as any type of calculation.[1] Also defined as use of computer technology in Information processing.[2][3]Computation is a process following a well defined model understood and expressed in an algorithm, protocol, network topology, etc …   Wikipedia

  • Neural backpropagation — This article is about the biological process. For the computer algorithm, see Backpropagation. Neural backpropagation is the phenomenon in which the action potential of a neuron creates a voltage spike both at the end of the axon (normal… …   Wikipedia

  • Artificial neural network — An artificial neural network (ANN), usually called neural network (NN), is a mathematical model or computational model that is inspired by the structure and/or functional aspects of biological neural networks. A neural network consists of an… …   Wikipedia

  • Codificación neural — Este artículo o sección necesita referencias que aparezcan en una publicación acreditada, como revistas especializadas, monografías, prensa diaria o páginas de Internet fidedignas. Puedes añadirlas así o avisar …   Wikipedia Español

  • Biological neural network — From Texture of the Nervous System of Man and the Vertebrates by Santiago Ramón y Cajal. The figure illustrates the diversity of neuronal morphologies in the auditory cortex. In neuroscience, a biological neural network describes a population of… …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”