When investigating the brain, models are very helpful. They allow us to explore neural behaviors without sacrificing animals or invading their nervous systems. We would like to understand how behaviors can be precisely timed after just one example. For instance if you're a child learning to play basketball, and someone shows you how to make a basket, you can probably do it all by yourself after one or two attempts. Early Model NeuronsModeling of neurons for computational purposes began with McCulloch & Pitts in 1942. They considered a binary neuron whose state could be either firing, or not firing. The neuron integrates its inputs (as a weighted sum), and if the result is over the threshold the neuron fires, and if it's under the threshold it doesn't. The original threshold function was the Heaviside step function, which is not smooth and can not be easily differentiated. The ideas embodied in the McCulloch-Pitts neuron later became the basis for the early learning machine called the Perceptron (Rosenblatt 1958). In the Perceptron, the synaptic weights are adjusted based on an error that is determined from the data. The Perceptron essentially "fits" the data to a set of linear parameters. Unlike the McCulloch-Pitts neuron, the Perceptron uses neurons with a sigmoidal threshold function, which is differentiable and therefore errors can be passed backward through the network and assigned to their sources in proportion to their contributions.![]() Simple neurons like these have severe limitations, and based on what we've already seen they're inadequate to mimic biological behavior. Nevertheless they still have significant computational abilities when they're wired into populations. The Perceptron is able to perform "linear separability"on a dataset, meaning it finds the slope and intercept of a straight line that partitions the dataset. More neurons simply means more lines, the end result is a combination of linear partitions. Linear regression is probably the single most common activity in all of statistics, so this capability is useful. However it doesn't work on all datasets, and the Perceptron is a bit of a one-trick pony in terms of its neural behavior. Unfortunately Frank Rosenblatt passed away before he could counter Marvin Minsky's points about Perceptron limitations, but before he died he saw his invention being used in real time national security applications. Even binary neurons can be very powerful when wired into populations and given the right kinds of plasticity. ![]() In the above figure the inputs I are multiplied by the synaptic weights W and integrated over the dendritic surface to arrive at a sum S, which is added to a "bias" term b representing noise or a baseline input level. The difference between this neuron and the McCulloch-Pitts neuron is that the threshold function f is sigmoidal (S-shaped) instead of being a step. When this threshold function can be differentiated, an error occurring at the output O can be passed back through the threshold and the contributions of each input to the error can be determined. Then, the synaptic weights can be updated to reflect the new error information. The process of passing the error in the opposite direction through the threshold function is called "back propagation". Simplifying the Hodgkin-Huxley ModelReal neurons have more than two parts, more than just an integrative portion and a spike generating portion. The properties of a neural membrane vary according to its location, for example calcium channels are present in the dendritic tufts of cortical pyramidal cells, and in the basal dendrites, but not in the shafts of the apical dendrites. In general ion channels are carefully localized along the neural membrane, and this is true everywhere, not just in the synapses. Morris-Lecar ModelOne of the variations of Hodgkin-Huxley that students should be familiar with (because it's widely cited and widely used) is the Morris-Lecar model. This is a somewhat simplified version of H-H that still shows bifurcations and spiking activity.![]() (figure from Rowat & Greenwood 2014) The drawback of the Morris-Lecar model is it only uses two conductances (calcium and potassium), and it has some significant computational artifacts, both at high frequencies and at certain particular frequencies.Fitzhugh-Nagumo ModelAnother model neuron variation that is frequently found in the literature is the Fitzhugh-Nagumo model. This model is an abstraction, rather than directly modeling ion channels. It does not display any bursting behavior, nor can it model subthreshold dynamics.![]() (figure from Nagumo et al 1962) Any of the above simplified model neurons can still be useful if we're just trying to replicate simple network or neuron behavior. However the simplified dynamics can create computational artifacts. We would like a simple model neuron that is computationally friendly and can be modified at least to a limited extent, so we can test the effect of various conductances and various geometries.Integrate-and-Fire (Izhikevich) ModelSpike times can be precisely modeled by integrate-and-fire neurons, which are frequently used when modeling TTFS and STDP (please refer to the glossary for the definitions of these terms). Unfortunately this model can create spike trains with arbitrarily high firing frequencies, so one must be careful. There are modifications to this model that allow for setting maximum firing rates (Strack et al 2023).![]() (figure from Izhikevich 2003) A MATLAB tutorial for Izhikevich neurons is given here.Poisson ModelsThere are some models that are "in between" rate codes and spike times, for example they may use a Poisson approximation to estimate spike times from a rate code. Such models are frequently used in oscillator paradigms to conveniently visualize the system attractors. Poisson models are related to gamma distributions, which in turn are related to Bayesian statistics, and gamma functions (which are the integrals of gamma distributions) are related to the fractional calculus which describes past, present, and future events. ![]() (figure from Amin 2006) To adjust the ratio of variance to mean in a Poisson model, the Fano factor can be used. If we're looking at a time series and digesting data as it arrives, the Fano factor can vary with time. Stochastic ModelsThere is a class of models that doesn't depend on the underlying physics and only looks at behavior (these models could be classified as "statistical", or "phenomenological"). The statistical approach uses the inter-spike interval (ISI) as the foundation of its method, and thus is often associated with Poisson dynamics. The subsequent analysis is very much along the lines of data science (using principal component analysis and similar techniques). The advantage of this method is that synaptic modification can be performed on the basis of temporal correlation alone, regardless of the underlying physical mechanisms. This approach has its roots in machine learning, it goes all the way back to Kohonen, Widrow, Hebb, and beyond. CompartmentationIn addition to the timing of action potentials, there are also questions relating to the integration of signals (data) along the dendrites. Among them are the issue of membrane multi-stability, the issue of dendritic spiking, and the issue of sub-threshold membrane oscillations. The latter issue can affect both integration and spike generation, and possibly serve as a control point for one or both. ![]() (figure from Wilfrid Rall - CC BY-SA 2.5) ![]() (figure from Wilfrid Rall - CC BY-SA 2.5) Within a dendrite, compartmentation is further enabled by the architecture related to spiny synapses. The thin stalks of dendritic spines help to restrict biochemical diffusion, essentially creating an independent compartment in each synapse. The communication out of such compartments is often complex, involving multiple stable states in the nearby dendritic membrane. In addition to ordinary graded potentials and plateau potentials, spiny synapses can generate dendritic mini-spikes that travel into the cell body, where they may convert the neuron from from one state to another. Sculpting Population BehaviorAt the end of the day, the behavior of a neuron is determined mainly by its ion channels. These channels can have different kinetics, they can be voltage sensitive or not, and they can have different conditions for activation and deactivation. One can model such behaviors using programs like NEURON or Nengo, but they're very difficult to implement in machine learning situations with TensorFlow or PyTorch, because the latter specialize in matrix multiplication and can't really handle dynamics. So the machine learning community encodes the dynamic in various ways to approximate the desired results. We are just now at the point where the neuroscience and machine learning communities are beginning to inform each other. I started modeling neural networks in 1978, long before they were a thing - and there was no one in the field at that time, only a handful of forward looking researchers who were often perceived as eccentric bordering on crazy. I was at Princeton when John Hopfield wrote his Nobel prize winning papers, and at the time no one took notice, they were considered an interesting oddity in the physics community. (But I noticed - I was fresh off extracting Volterra kernels from the lateral line organs of sharks and skates, and I looked at Hopfield's first 1982 paper and started laughing. "Binary neurons!" I said, and put the journal down. The next day I was back in the library looking at it again, because something about it caught my attention). And then it was a few years before Hinton and Sejnowski and David Tank and others expanded Hopfield's models and invented the Boltzmann machine (and demonstrated NetTalk - the video is still on YouTube). Next: Neurons in Populations |