Location: | House 6, Lecture Hall - Campus Nord | ||
Time: | Tue., 16:00-18:00 | ||
Contact: | https://itb.biologie.hu-berlin.de/~schleimer |
||
Practical: | 1 week end of semester |
Nerve tissue is comprised of excitable cells, supporting cells and a surrounding bath. This seminar lecture reviews the biophysical basics of excitability, including the stochastic processes that govern voltage-gated ion-channels in nerve membranes. The composition of ion channels, among other factors, determines the functional properties of action potential pulses, which are analysed using dynamical systems theory. The statistics of pulse trains is derived from the microscopic level of ion-channel fluctuations. This will also serve the understanding of a self-consistent description of macroscopic neural-network dynamics.
This scriptum shall evolve. It will have flaws (didactic ones and outright errors) but you, the reader, student, search bot, provide the selective pressure to improve it. Please write to the lecturer if you find fault of any kind with it.
Although written in a somewhat formal fashion the book is by no means a text in applies mathematics. It completely lacks the mathematical rigor to meet such standards. The choice of wrapping statements in definitions, observations and propositions arose from the need to be able to easily refer to them and to render the logical buildings blocks of some arguments easier to identify and follow through the text. Let’s see how that worked out …
There are many different lectures on “the biophysics of computation”, or “computational neuroscience”. The topic by now is so vast, that only a small non-exhaustive subset of the subject can be discussed within the temporal confinement of a lecture. This one is no different.
The idea here is to approach questions like “What is a nerve impulse?” or “Are there different types of excitability?” from a mathematical and a biophysical angle.
From the biophysical point of view, one has to addresse the many buidling blocks of nervous systems: There is needs to understand the biophysics of lipid sheets and the chemical gradiente across them. Voltage-gated ion channels swimming in the membrane should be understood, including their stochastic dynamics. What is going on in the extracellular medium might be a topic. And of course what are the consequences of wiring neurons in networks via synapses or gap-junctions. Also the enourmous energetic demands of nerve tissue are important factors in understanding its structure and function.
Many of these components of nervous systems derive from preadaptations, meaning precurses have been arround long before the first nervous system emerged some 600 Myr ago. This is certainly true for excitability, which is found in singe celled Paramecium and even bacteria. Synaptic components like the SNAR complex is present in fungi and some of the transmitters are used in bacterial communication.
Naturally, there is a risc of getting overwelmed by these beautiful details. To explain the complexity of nervous systems ab initio seems a mindbogglingly difficult task. Yet, there are already all-atom simulation of ion channels in the membrane. Still, before and during discussing the structural aspects it is insightful to ask what are the exigences of a nervous system. Not in any teleological sense, theoretical biology is over that, but from an evolutionary standpoint. One may even formulate certain desiderata, without claiming that brains were design by anyone.
If by attending the lecture or goint through its notes the amount of confusion becomes unbearable Tinbergen’s levels of analysis might help
Diachronic view of historical sequence | Static view of current form | |
---|---|---|
Proximate cause | Ontogeny | Mechanism |
Ultimate cause | Phylogeny | Function |
While, categories and schemes like the above are of utility in structuring ones thoughts about nature, they should not be taken as dogmatic fences that do not allow you to think, speculate outside of them. Nature is to beautiful to get involved into dogmatic fights over reified wordings, leave them to the humanities.
Roughly one might associated certain disciplins of applied math that can be utilised in the four categories from above.
Diachronic view of historical sequence | Static view of current form | |
---|---|---|
Proximate cause | Control theory | Dynamical systems, Stochastic processes |
Ultimate cause | Information theory, Optimisation | Filter theory, Triggered ensembles |
Of many natural systems we observe only some drastically obvious events, while the underlying complex evolution of all state variables is hidden. The malatonin falling level below threshold (in the suprachiasmatic nucleus) causes a “waking-up event”. Results the complex regulatory network causing gap phase 1 synthsis gap phase 2 mitosis transition result in clearly visible cell division events. Sadly, extinction events can be the results of changes in complex foodwebs, the dynamics of which is not completely determined. Action potentials are no different. The dynamics involves charging the cell membrane and the nonlinear kinetics of several ion channels, but the electric impulse is an clearly detectable event (with a glass electrode at least).
Why do nervous system use pulses, called action potentials or spikes to communicate? In fact, not all do. The information picked up by your photo receptors is processed by cells in the retina (amacrine cells, horizontal cells) still as graded potentials, which are then recoded into a impulse based representation. Conversely, the visusal system of flys starts out with pulses representation and converts it to graded potentials upstream. Last but not least, the small nervous system of nematode C. elegans (302 neurons) seemed to operate entirely on graded potentials. No spikes had been reportet for a long time. But in 2018 someone found Ca based spikes in one of their neurons.
One of the differences between passive graded potential and an active action potential is that the latter does not suffer from the same kind of siganl degradation over long distances. There could be active more or less graded potentials, but these are less temporally precise as a short pulse and it turns out that many computations rely on on exactly this temporal precession.
A neurobiological “dogma” states that action potentials follow the all-or-nothing principle (E.D. Adrian and Zotterman 1926). Fortunately, the jelly fish Aglanta digitalae does not care much for dogmas and encodes its swimming patterns in different action potential shapes. So its not really a dogma or even a principle. But most action potentials really are very stereotypical events, hence its a common property. The all-or-nothing property can be construed to mean that the exact shape of the action potential does not matter1 and that information is instead stored in the exact spike times or the frequency of their occurence.
This can be formalised to mean that the voltage evolution ought to be represented by a shifted pulse wave form
where are the spike times, which need to be defined later.
It also means that the action potential should be stable under perturbations, but that inputs ought to be able to shift the occurrence of spikes in time. In mathematical terms, these desiderata read
To wit, (ii). just means that stimulus induced phase shifts should neither decay nor blow up.
The mathematical object that suffices the desiderata (i) and (ii) are limit cycles. They are closed orbits in phase space. Naturally, the state space need to be at least two dimensional for it to contain a limit cycle. What are the state variables? The membrane voltage, , typically in milivolts sounds like a reasonable one, at least that is what electrophysiologist measure. The conformational state of the ion channels also changes with time, , and so do ion concentration, say . The exact governing equations are going to be derived in the following chapters. For now, denote a vector of state variables as
A general differential equation could have the following structure
is the time independent (homogeneous) flow field. is a (often assumed small), possibly time dependent (inhomogeneous) perturbation to the system. are system’s parameter. Parameters could be cell size and other fixed properties, but also a constant applied bias current. The change of solutions in response to variations in system’s parameter will be studied later.
In the absence of perturbations, , Eq. (2) becomes a homogeneous ordinary differential equation
We are interested in the cases in which, at least for cerctain parameter values , a -periodic limit cycle solution, with , also known as periodic orbit. This solution is identified as the action potential if it sufices the requirements of amplitude-stability and time-shiftability (i) and (ii).
Stability of is probed by studying small perturbation to an invariant set solution. Our invariant set is the limit cycle (periodic orbit) . Assuming there was a small perturbation to the system the solution can be decomposed as
with some “small” perturbation to the orbit. What small, i.e., is we do not want to say now, maybe later, lets see …
Assuming the perturbation was transient (only in initial conditions) and the system is homogeneous again we plug the Ansatz of Eq. (4) into Eq. (3) and get
The Jacobi matrix evaluated on the limit cycle can be written as a function of time . Note that since the limit cycle solution is -periodic, so is .
Identifying the limit cycle solution above we are left with the first variational equation of Eq. (3)
Hence, one needs to study of linear system with periodic coefficients. One solution of Eq. (5) can be guessed, let us try the time derivative of the orbit,
So it is a solution alright, and it happens to be a -periodic solution. This solution is called the Goldstone mode. But for arbitray intitial conditions not all solutions should be periodic.
For the proof please consult (Chicone 2006), here we are just goint to work with this as an Ansatz. The constant matrix is called the Floquet matrix.
Recall the matrix exponential
Inserting the Floquet Ansatz into Eq. (5) yields
Which by multiplying with results in a dynamics equation for similarity matrix .
Remember that was invertable so one can also derive an equation for the inverse
The Floquet matrix, , is constant, though not necesarily symmetric. Hence it has an orthonormal left and right eigensystem
The are called Floquet exponents, for reasons becomming obvious below. One can now define a “rotated” eigensystem
Def. (“Rotated” eigensystem and Floquet modes) #def:floqmode
The rotated eigensystem of a spike is and . For which ipso facto obeys
- .
and are also called the Floquet modes.
If we project the eigenvectors on the Eqs. (9) and (10) and use this definitions we get
and
As a final test of what has been chiefed by introducing the new time-dependent coordinate system of the right and left Floquet modes, project the general solution from Eq. (7) on the the adjoint Floquet modes, :
.
Along these rotating basis one observes a simple exponential contraction or expansion depending on the sign of the Floquet exponents. The maginitude of the Floquet exponents is the rate of convergence in units of 1/Time. What if a Floquet exponent is zero? In fact, by comparing Eq. (6) and (13) one may observe that the Goldtone mode is alwasys a solution to the
Note that if ,
is the Goldstone mode, and from Eq. (12)
W.l.o.g, let the Floquet exponents of a stable limit cycle be ordert such that
What are left Floquet modes useful for? Here is an example. First, rewrite Eq. (14) for in the form
Say one is interested in the change of the period of a limit cycle oscillator w.r.t. a parameter change. By rescaling time
Taking the total derivative (and chain rule) one finds
The derivatives of the first term can be excanged. Now project the equation from the left onto the zeroth Floquet mode
The last term can be rewritten by transposing the operator
Hence
There is still a lack of a formal definnition of the spike time in Eq. (1). Electrophysiologists typically envision a voltage threshold, but these would only exist if the voltage dynamics was a one dimensional equation, which it is not. In fact, it can not be if true limit cycles are to exist. In the general state-space of action potential dynamics there may exist a generalised threshold manifold. A reasonable alternative definition would take the (local) peak/maximum of the voltage
.
Again, a definition based on properties of the state variables and it also leaves all other dimensions undefined. The general problem seems to be to uniquely associate time with states. If the dynamics was restricted to only on the limit cycle this is easy. As the maximum voltage value is only reached once per spike it also implies unique values for the other state variables, wich can be called the spike boundary condition: . Most importantly it also implies a unique time indices at which such events occur, . The normalised time index of points on the limit cycle is also called the phase, s.t. . Given synaptic bombardment, intrinsic noise and other perturbations the dynamics, however, may be temporarily not exactly on the limit cycle. The simple idea is that the time phase of any point in the basin of attraction of a limit cycle can be defined as the phase it converges back to when asymptotically approaching the limit cycle. To make this precise, first define the flow:
Def. (Flow of an ODE) #def:flow
For a given initial condition , where is the basin of attraction of the limit cycle, the flow or trajectory induced by the ODE in Eq. (3) is denoted as .
Numerically, the flow can be evaluated by forward-integration of the ODE starting at a given initial value.
Def. (Assymptotic phase) #def:assphase
For point , in the basin of attraction of the limit cycle solution of Eq. (3), the asymptotic phase, , is defined as
- $\psi(\vec y)=\argmin_\psi\lim_{t\to\infty}\|\vec x(t-\psi {P};\vec y)-\vec x_{\mathrm{lc}}(t;\vec x_\mathrm{sp})\|$.
Note that the limit cycle solution was initiated at time at the spike boundary condition. Based on the asymptotic phase, the concept of isochrons can be introduced (Winfree 1974; Osinga and Moehlis 2010) as the level sets of constant phase:
Def. (Isochrons) #def:isochron
The isochron manifolds, , are the level sets of the asymptotic phase variable
.
An isochron defines a dimensional manifold of points that converge to the same asymptotic phase on the limit cycle.
Based on the isochron one can now define a spike as a crossing of the flow/trajectory with .
Note: There are several other ways to define a phase (Hilbert transform, linear interpolation between spikes, …). In principle, there should be monotonous mappings between each other.
An alternative to studying the high dimensional set of coupled nonlinear biophysical equations is to find an (approximately) input/output(I/O)-equivalent equation for the phase evolution like in Fig. #. In the figure the phase is ploted as modulo 1. However, it is also convenient to consider the freely evolving phase since this is related to the spike count process,
.
The spike time is then defined as
.
A spike-train can be written as
Def. (I/O-equivalent phase oscillator) #def:io_phase
An I/O-equivalent phase oscillator produces with its threshold crossing process the same spike times as a particular conductance-based model (Lazar 2010).
Fig. #fig:phasered
The phase equation turns the spiking neuron into a one-dimensional level-crossing process with equivalent spike times. The phase is plotted modulo 1.
The evolution of the phase of Eq. (2) is given by the chain rule as
.
To first order this is
The geometric interpretation of the phase gradient, , is that it points perpendincular to the isochrons towards the steapest change of the asymptotic phase variable .
Def. (empirical PRC) #def:empiricalPRC
Given a short -pulse of amplitude , delivered at a certain phase, , (time point) in the inter-spike interval of a tonically firing neuron the linear phase response is defined as
$\prc(\psi)=\lim\limits_{\epsilon\to0}\frac1\epsilon\frac{{P}_0-{P}_\epsilon(\psi)}{{P}_0}$.
The underlying assumption of interpreting PRC is that it is a function, $\prc(\psi):[0,1)\mapsto{I\!\!R}$, that mapps phase or time of occurence of a perturbation to a persistent measurable phase shift. (Please appreciate how idealised and “noise free” these definitions are.) The convention is that a prolongation of intervall corresponds to a phase delay, while shortinging means phase advance.
The perturbed period can be written as the original period minus the phase shift converted back to time (by multiplying with the unperturbed period)
.
Given that the time evolution of has been perturbed at some time point , The phase shift can be measured at the spike time of the unperturbed system as
.
The empirical PRC in Def. # can then be expressed as
$\prc(\psi)=\frac1\epsilon\frac{{P}_0}{{P}_0}(1-(\phi({P}_0)-1)-1)=\frac1\epsilon(\phi({P}_0)-1)$.
Apply the identify operator of functions (convolution with a -function) on the left and the fundamental theorem of calculus on the right hand side then to first order
$\epsilon{P}_0\int_0^{{P}_0}\prc(\phi(t))\delta(\psi-\phi(t)){\mathrm{d}}t=\int_0^{{P}_0}(\dot\phi-\frac1{{P}_0}){\mathrm{d}}t$
This step is not really allowed, but for well behaved integrals we can identify the integrands to see
$\dot\phi=\frac1{{P}_0}+\epsilon\prc(\phi)\delta(\psi-\phi)$.
One can bring this equation in agreement with the phase evolution in Eq. (20), if one identifies
$\nabla\phi(\psi)=\prc(\psi)$ and .
The assumption underlying the experimental proceduce described in Def. # is that the phase shift stays constant in time, say . Without perturbation there is not long-term shift of the phase. Hence it must be the perturbation off the limit cycle which causes the phase shift. The phase change must be due to the and is given by
const. From one gets
Again the evolution of can be linearised if it is small
Such that by Fredholm’s alternative
This is the same ODE as for the zeroth left Floquet mode, rendering them identical up to scaling.
Motivation and aim
In this lecture ion channels are introduced as stochastic devices floating in the membrane of a nerve cell. It should motivate why the analysis techniques and models introduced in this lecture need to deal with fluctuations and noise. (Nerve) cells produce stochastic processes on several levels:
The ion channels in their membrane stochastically jump beteen conformations, governed by Master equations.
On a more macroscipic level the their membrane voltage fluctuations show properties of coloured noise, well described by diffusion processes and stochastic differential equations.
The trains of action potentials they emmit from point processes.
In the stationary state these processes can be subjected to spectral analysis.
Note that in 1952, the first equations describing membrane voltage dynamics where the deterministic rate equations by Hodgin & Huxley (Hodgkin and Huxley 1952). Only in 1994 Fox and Lu derived these equations from the continuum limit of an ensemble of stochastic ion channels (Fox and Lu 1994). Essentially, this was a chived by doing a diffusion approximation in the thermodynamic limit of many ion channels.
Finally, in 1998 the structure of a K+ ion channel from Streptomyces lividans, which is nearly identical to Drosophila’s shaker channel was discovered using X-ray analysis (Doyle 1998).
Since then the nuences of appllying diffusion approximations to neurons have been investigated (Linaro, Storace, and Giugliano 2011,Orio and Soudry (2012)) and reviewed (Goldwyn et al. 2011,Goldwyn and Shea-Brown (2011),Pezo, Soudry, and Orio (2014)).
Observation (Conserning the ion channels of excitable membranes) #obs:chan
Ion pores show different degrees of selectivity to ionic species. Understanding of the mechanisms of the ionic filter was achieved by structural analaysis (Doyle 1998). Potassiume channels conduct K+ 1000 fold more than Na+ and allow for conduction rate of 108 ions per second.
In single channel patch experiments ion channels how discrete levels of conductivity, which look like a Telegraph process.
Independent ion channels (no cooperativity or spatial heterogeneity)
Membrane bound proteins like many others change conformation on various triggering signals:
- a change in pH
- a change in the surrounding electric field
- mechanical preassure
- ligand binding
To relate measured voltage changes to ion flux accorss the membrane note that according to Kirchofs law and electro neutrality
The currents are assumed to follow Ohms law:
where is the unitary conductance and is the total number of open pores.
Starting with a hypothetical ion pore of which there is a fixed total concentration, [Pore]=const. The pore an open conformation, , in which, say only K+, ions can pass and a cloased states, , which blocks ion flow completely:
Def. (Two-state ion-channel) #def:twochan
Based on the law of mass-action, the corresponding first-order rate equation is
The second equality is by conservation of mass const.
has the typical strucutre of its stochastic counter parts the Master equation of a birth-death process.
Hodgkin and Huxley normalised the equation by defining .
Under a voltage-clamp experiment, for each voltage level this equation predicts an exponential charge up to a level with the time constant that could be fitted.
Hodgkin and Huxley discovered that their measured traces could not be fitted well by the above equation.
Def. (A tetrameric K+ pore) #def:fourchan
Instead of writing down heuristic macroscopic laws for channel concentrations, one can attempt to derive them from microscopic dynamics of single pores. For this the state diagram in Def. # is reinterpreted as the state of a single molecule.
Based on Obs. # such single channel dynamics can be described by a finite state Markov model which is mathematically described as a Master equation.
A Markov model represented by a graph like Def. # is a special kind of chemical reaction network studied in systems biology. The connectivity/topology can be stored in a
Def. (Stoichiometry matrix) #def:stoichmat
The stoichiometry matrix, describes reactions (transitions) involving chemical species (states). It can be decomposed into the educt matrix and the product matrix by
The stoichiometry matrix of Def. # would look like
In addition one needs the rate vector
to define the complete reaction network for the channel.
To understand the dynamical evolution of occupation propability of different conformational states, take the two-state channel as an example. Define a vector of the probability of being closed and open, , respectively.
In actuality the channels are part of the membrane dynamical system, where and depend at least on and hence are not constant during a spike. One needs an update rule of how to get from the probability of being open at time to the probability of begin open at time . Given are the rates, so the propabilities of change in an interval of time are the rates times the interval length, e.g. . For the two state channel this cann be summarised in matrix notation
or in vector form
The infinitesimal limit yields a kind of Kolmogorov forward-equation
Def. (Transition rate matrix) #def:transmat
The transition rate matrix (or infinitesimal generator) is .
In general the transition rate matrix can be obtained from the graph
$\boldsymbol{Q} = \boldsymbol{N}\diag(\vec v)\boldsymbol{E}^{\mathsf{T}}$.
Exercise (Transition rate matrix) #ex:mc1
Derive of a three state channel
Note that Eq. (21), though a consistent description of the occupation probabilities, does not give us realisations or sample paths comparable to the traces measured in single-channel patch clamp experiments.
Given a total number of ion channels , the goal is to track the actual number of channels in a particular state. Based on reaction rate theory and given a certain reaction rate , the number of reaction events in a time window is
If several reactions can occur and several molecules are in volved then (because Poisson distribution is closed under convolution) one has
.
So the rate of the occurence of any reaction is given by this linear combination of the rates of individual reactions. The waiting time distribution between reactions is exponential with the same rate:
Starting in a particular state at time the life time of staying in that state until is
The escape rate of the state is (any reaction occuring)
where are the rate of leaving state . For example in the K+-channel is
But which reaction did occur? Let be the reaction (not the state!). For example, there are 8 reactions in the K+ channel. The probabilities of occurance associated with any one of them is
In a computer algorithm one can draw the waiting time for the next reaction simply by generating a uniform random number and then
The reaction is determined with a second random number
while
r1 = rand
r2 = rand
tau = ln(1/r1) / a
With we can express one row of this equation as
or
with and
which has solution
Fourier transformation leads to
or
a Lorenzian spectrum. Inverse Fouerier transform yields the covariance function
An ionic current produced in the open state would be
and are the unitary conductance and the Nernst potential respectively. The average such current would be
where is the total number of channels in the membrane patch under consideration. But what about a particular stochastic realisation of the current, what about the fluctuations around the mean?
If we have channels than the number of of them being open is binomially distributed
Let us try to calculate the statistics of the current originating from an -state channel (just like in the two state case). Why would one do this? The idea, later on is to be able to find a continuous stochastic process that we can simulate and analyise easily.
Let be the realisation of the -state Markov channel. For example a K+-channel with four subunits
For the K+ channel with conformations states, of which one is conducting, let us further define
If there are several conducting states (with possibly different conductance level, one could have e.g., .
The single channel current at a voltage clamp is then
How does it evolve? Define and , then
With formal solution
Use the singular value decompostion, , the matrix power can be written as . Or (recall Eq. (8))
If Eq. (23) has a stationary distribution in the limit, then this must correspond to the eigenvalue (let us assume they are ordered). So for . Therefore, the solution can be written as
The average channel current of Eq. (22) is
,
which if the chain is stable () has steady state
.
The steady state covariance in this case is
Hence
If then
is a sum of exponentials. The spectral density is a superposition of Lorenzians.
The jump process discussed in the previous sections is a continuous time Markov-process on a discrete domain, with . A diffusion process is a continuous time Markov-process on a continuous domain, with .
Can are in search for a a diffusion process such that
has the same first and second order statistics (in voltage clamp) as Eq. (22)? Let us try
To solve it, use the Fourier Transform
The place holder symbold was introduced for the Fourier transform of the stochastic process . Rearanging yields
The spectrum is
By definition is the Fourier Transform of the covariance function and from Eq. ((???)) this is one. Hence,
Applying the inverse Fourier transform results in the correlation function
.
A super position of independent such OU process
leads to a correlation function with the same structure as in Eq. (26). We identify and .
The idea of matching the second order statistics can be formulated in a far more abstract way in terms of the Kramers-Moyal-van-Kampen expansion of the Master requation
from
Often small changes, say in an ion channel density, has only small quantitative effect on a neurons behaviour.
Often numerical solutions to nonlinear ordinary differential equations are found by (forward) time-integration. An interesting alternative is to track a found solution through parameter space, for which the solution must be persistent in a certain parameter. If it is not, then a bifurcation occurs and one observes a qualitative change of the solution. The study of system solutions under parameter perturbations is also the subject of sensitivity analysis.
For book on bifurcation theory consult (Izhikevich 2007) and numerical bifurcation analysis (Kielhöfer 2011).
Asume for
there is a steady state solution
In biological systems like cells, parameters are rarely stable over longer periods, because there is a continuous turnover (synthesis and degradation) of all molecules. Hence, so one should be interested in families of solutions for varying prarameters, .
Def. (solution banch) #def:bifbranch
A branch, familiy or curve of solutions with regard to an index parameter is denoted as .
The existence of the solution as a function of the parameter is governed by the implicite function theorem:
Theorem (Implicite function theorem) #thm:implicitefoo
Consider a system of equations
, with , , and .
Let and be smooth near . Then if the Jacobian is nonsingular, a unique, smooth solution family such that
.
This establishes the existence of lines in a bifurcation diagram.
The method of continuation is a predictor-corrector method. In practice, assume the fixpoint is known for one particular parameter value , then for a small change in parameter the solution is is predicted
- .
To predict the new solution, the old solution is required and the derivative of the solution w.r.t. the parameter that is changed. How to compute the latter? Take the total derivative of Eq. (28) w.r.t. the parameter
.
with formal solution
.
If is full rank one can use some efficient linear algebra library to find the vector and back insert it into Eq. (29). For too large the predicted solution will be wrong. Yet, it is a good initial guess form which to find the correct version.
Actually that Newton’s iterations are also obtained by linearisation
,
which if solved for yields Eq. (30).
Convergence analysis of Newton iterations yields that with each newton iterations the number of correct decimal places doubles. Hence often a low number of iterations (3-7) suffice to achieve sufficient numerical accuracy.
Example (Continuation of the QIF model) #ex:contQIF
with the solution branches .
From and , one gets . What happens at ?
unstable
stable
Also and thus . What happens at or ?
Or more general: What happens if we find that the condition number of explodes?
In the example above the branch switches its stability and it bends back in a fold or turning point. In general folds can occure with and without stability change.
Bifurcatin analysis is one big “case discrimination” task. Luckily, and due to topology the cases are, given certain constraints, finite.
Recall some difinitions
and .
So that the nullspace is spanned by .
This section considers . Hence, the implicite function theorem is not applicable. From the example above it is apparent that at the bifurcation. The problem can be circumvented by defining a new “arclength parameter”, . The bifurcation branch is then a parametric curve, . Without loss of generality the bifurcation is to be at .
If the Jacobian matrix is rank-deficient the Lyapunov-Schmidt reduction can be applied. Intutiefely the problem is reduced from a high-dimensional, possibly infinite-dimensional, one to one that has as many dimension as the deffect of .
The nullspace is spanned by the eigenvector, of , corresponding to the eigenvalue 0.
Assume that is twice differentiable w.r.t. , then differentiate w.r.t. and evaluate at
Let be the Hessian tensor.
At one has and hence from Eq. (31) . Projecting onto the left-eigenvector to the eigenvalue 0. at one finds
or with
.
This is a test function of wether the bifurcation is
There are several cases to be distinguised.
At a simple fold one eigenvalue is zero. Study the eigen system of near
.
With a bit of analysis
Continue the extened system (linear and nonlinear)
The same procedure than above can be applied to the continuation of periodic orbits and boundary value problems. Define the implicit function to encompass the time derivative
, with .
Then proceed as above. Note that the time derivative is a linear operator which has a matrix representation just like the Jacobian. in that sense
Def. (saddle-node/fold bifurcation) #def:bif-sn
The fixpoint defined by is a fold of a saddle and a node if
- the Jacobian has one eigenvalue with algebraic multiplicity one,
Def. (Hopf bifurcation) #def:bif-hopf
Assuming all but two eigenvalues have negative real parts, except for a conjugate pair of purely imaginary eigenvalues.
At arbitrary parameters the periodic soluiton adjoint to the first variational equation on the limit cycle yields the PRC. I can be calculated numerically with the continuation method. Near a bifurcation, however, if major parts of the dynamics happen in the centre manifold the PRC can be calculated analytically. As an example take the saddle-node on limit cycle bifurcation (SNLC). The spike in this case is a homoclinic orbit to a saddle-node, that enters and leaves via the semi-stable (centre) manifold that is associated with the eigenvalue of the Jacobian at the saddle.
Let there be a saddle-node (fold) bifurcation at some value and the saddle-node position is . In the previous lecture we noted that if the bifurcation is not transcritical (degenerated)
The change of the solution w.r.t. the pseudo-arclength parameter was
.
Given that the linear part of the dynamics is degenerated at the bifurcation, one might look at the leading nonlinear term.
Expanding the right-hand side around saddle-node fixpoint, , yields (B. Ermentrout 1996)
Establish the eigen system at the saddle-node
and with .
By assumption of a saddle-node the Jacobian has a simple zero with an associated eigenvector. All other eigenvalues are exponentially contracting , while is slow.
Write the dynamics arround the saddle-node as , and also use small changes in the bifurcation parameter . Then Eq. (32) is
.
Projecting this equation onto the left eigenvector yields the isolated dynamics along the centre manifold:
The centre manifold is only tangential to the spiking limit cycle dynamics near the saddle. Although the proportion of time spend near the saddle is large at some point the trajectories depart so it is only locally valid.
The formal solution of Eq. (33) with initial condition is
Away from the saddle-node the slow manifold accelerates again and the variable explodes at . This blowing up is like a spike. So for the SNLC centre manifold one can think of the the spike at and reset to . The time it takes from to is finite and given by
.
Note that for it is
and
.
The bifurcation parameter enter in . If , then the firing rate has the typical square-root scaling
.
Let us check this numerically in the pratical.
The last chapter argued that conductance-based neuron models, are a consistent continuum limit of Markov description of finite ion channels. Hallmarks of the resuliting ordinary differential equations include first-order kinetic equations for ion channel components, i.e. gates, and certain polynomial nonlinearities. There is a rich literature on the dynamical systems analysis of single neurons, such that it is difficult to structure it for a lecture. Typically, the starting point is the different “simple” bifurcations that can be found in neuron models. One can also start with the “simplest” of the complicated bifurcations that summarises all possible simple ones. This is asking for the simplest chimera neuron with as many of the known features as possible (Kirst 2012; Kirst et al. 2015).
Start by restating some empirical observations on the phenomenology of excitable nerve cells. For pulsed-based communication these equations need a stable resting potential and a limit cycle-like transient pulsatile solution for certain input perturbations (Pereira, Coullet, and Tirapegui 2015).
Phenomenon | Dynamics |
---|---|
resting potential | stable node |
subtreshold oscillations | stable spiral |
action potential | stable limit cycle |
thresholds | saddle, unstable limit cycle |
Note that the focus on the generation of single or tonic spikes, not bursts. Bursts require at least one additional time scale to the pulse generating ones and are not minimal in that sense.
Observation (nerve cells) #obs:neuron
- stable resting potential, for in the convex hull of reversal potentials,
- stable pulses
Remark (dimensionality of state space) #rem:dim
The minimal dimensionality of the state space to get spikes (but not bursts) is .
Def. (Class-1 excitability) #def:excite1
Class one neurons have arbirary low frequencies (long periods). This also implies they can show very long (and in the presence of noise varied) latencies to the first spike.
Lets write the conductance-based model as follows
Def. (Conductance-based neuron model) #def:cbm
where
and the components of the vector of gates evolve according to
.
For brevity, introduce and denote and , such that .
The goal is to find “the simplest” of the complicated bifurcations that allows for all known phenomena. What is meant by the simplest? Bifurcations can be ordered according to the codimension, i.e. how many parameters are required to tune into them. Codimension one bifurcations a points in one-dimensional parameter spaces, hey are lines in two-dimensional spaces and sheets in three-dimensional spaces, and so on. Which codimension should one allow for? It does not make sense to use more parameters than are biophysically present in all the neuron models from the model class of interest. These are three for conductance-based neuron models:
Def. (fundamental parameter) #def:params
Conductance-based neuron models may have a very diverse potpourri of ion channel with different parameters, but all share the following fundamental parameters: . it the time-scale separation between kinetics and current balance equation. and are the constant and linear term of the total membrane current. All other “kinetic parameters” depend on which exact channels are present in the nerve cell model and may vary substantially.
The simplest known bifurcation the can be broad into agreement with the phonomenology above is
Def. (Bogdanov-Takens cusp bifurcation) #def:bif-bt
A zero eigenvalue of algebraic multiplicity two. And a degenerecy assumption to be able to get rid of the fold via a cusp, which reads that the quadratic normal form coeficient must be zero.
Assertion (BT point) #assert:btc
A BT point implies the emergence of a homoclinic orbit to a saddle. Importantly, due to the coincidence with the planar limit cycle of the Hopf bifurcation, the homoclinic orbit is planar and not as three dimensional as a homoclinic orbit to a saddle-focus (Shilnikov). It may turn into that further away from the unfolding.
Assumptions (Electrophysiology) #ass:cbm
The following assumptions can be motivated from biophysics of excitable membranes or electrophysiological measurements:
- All conductances are positive, .
- The kinetics of the ion channel gates is of first order, with all activation curves are bounded, monotonic, twice-differentiable functions, that become sufficiently flat at . More specifically the derivative of the activation curve has to vanish at and do so faster than linearly, i.e. .
- Channels are voltage-gated only.
- Class 1 some slow positive feedback exists, i.e.,
Remark (excitable system vs limit cycle) #rem:excitesys
Before proceeding to the actual argument, what of “excitable systems”. Many neurons do not spike tonically but only if excited by a synaptic perturbation or noise. Or in more physiological terms this is tonic spiking vs Poisson like spiking (low vs high CV). In such cases they cross a “potential barrier” or some threshold manifold and return to rest after a long excursion through state space. One way of prduceing a system with such a long excursion is to have it have a stable limit-cycle for “other parameters”. It turns out, that the Bogdavnov-Takens point plays a crucial role here two because it garantees there are homoclinic orbits attached to a saddle, whcih is one way to get such excrusions.
Proposition (fixpoints of CBM) #prop:fp
From and one gets:
.
Provided solutions to the implicite definition of the voltage fixpoint exists, then for they lie within the convex hull of the reversal potentials .
Given that the fixpoints of the gates can readily be found, the steady states of the whole system are the zeros of the steady-state steady-state - curve:
Def. (steady-state - curve) #def:ivcurve
Given that all gates have time to equilibrate to their steady state values (are in kinetic steady state), the steady-state - curve is defined as
.
It is a continuous and differentiable function just like . This allows some weal statements bout the number of steady states expected in a conductance-based model.
Lemma (fixponts of class 1 neurons) #lem:fp-class1
Depending on their class 1 neurons can have at least one or three steady states.
The partial derivative of the gating equations w.r.t at the kinetic steady state is
.
Hence, the total derivative of a kinetic equation w.r.t is in that case
,
which shows that in kinetic steady state . Therefore, the total derivative of the steady-state - relation is
.
Inserting the definitions yields
.
This can be used to inpsect the asymptotic behaviour of at large positive and negative voltages:
,
where the last two terms have vanished due to Assumptions #a) and b). This means the slope of the - curve is negative at , and thus should cross zero at least once. For the class 1 excitable system also Assumption #d) stands and consequently must have at least one maximum. Consequently can be tuned such that there is at there are at least three fixpoints.
Let us call the maximums
Next we are making use of Axiom #c) to calculate the Jacobian of a conductance-based neuron model
Observation (rank-1 updated diagonal Jacobian) #obs:rank1diag
The special structure of the Jacobian in a conductance-based neuron model, a rank-1 updated diagonal matrix,
,
Proposition (Zero-eigenvectors of a CBM) #thm:cbnJacobian
Consider a conductance-based neuron model (Def. #) at a fold bifurcation, i.e., with a simple zero eigenvalue, , at the saddle-node. Then the semi-stable (centre-)manifold is tangential to the right-eigenvector corresponding to , which is given by
- ,
where and was used. The normalisation constant is given at the end of Proof #.
The associated left-eigenvector is
- ,
where denotes the r.h.s. of the current-balance equation and for the r.h.s. of the kinetic equations. Note that the right eigenvector gives the direction of the semi-stable manifold.
The determinant can be rewritten as
allows for an explicit calculation of the left and right eigenvectors corresponding to the eigenvalue .
Choosing the zero eigenvector as , one obtains from a set of equations:
, and
for .
From Eqs. (40), it follows that . This solution also solves Eq. (39), because the required zero eigenvalue implies that . Note that according to Leibniz’s formula the determinant of a matrix
is
- .
Since , the first factor of the determinant must be zero. Hence, also solves Eq. (39).
An analogue derivation for the left eigenvector, , results in the following set of equations,
, and
for ,
which yields . The normalisation constant to ensure is
. Note that is in units of volt here. Other conventions are possible.
A long calculation shows that
.
To the condition for the BTC bifurcation require the eigenvalue to be not simple, but of algebraic multiplicity two. This means there is a Jordan chain of rank two, such that for the left and right eigenvectors, and , corresponding to there are also a pair of generalised eigenvectors (KUZNETSOV 2011), which are not denoted as and .
and with and .
Note that the orthonormalisation is somewhat unintuitive, but useful. The BT case is , such that and . Numerically, this equivalent to solving an extended eigenvalue problem
Proposition (Generalised zero-eigenvectors of a CBM) #prop:gen-eig
Given the eigenvectors and and the difinition one finds
As the normal eigenvectors have already been calculated above up to normalisation, the equations for the coefficients of the generalised eigenvectors are
From one obtains
From we obtain
Once the generalised eigensystem spanning the nullspace are determined the normal form of BT bifurcation can be written as
After some longer calculation one obtains a condition for obtaining the generalised eigenvectors:
Conditions (BTC) #cond:btc
and the previous conditions
The requirement of a zero determinant as defined in Eq. (38) leads to Eq. (45). Eq. (46) follows from the degenerecy requirement of the CUSP. It was shown as a consequence of Lemma # that such a point actually exists if one starts with a class 1 neuron.
From these conditions the following statement can be derived.
Lemma (unique param) #lem:uniquepar
From the above one can conclude that from the conditions of the BTC point a unique set of parameter can be determined.
Eq. (44) can be solved for
Eq. (45) depends only on and . It can be solved for
The input current can be obtained from the fixpoint condition of the current-balance equation
,
only if the steady state is available. Luckily condition Eq. (46) is independent of the fundamental parameters and can be solved to find .
Given the biophysical nature of the parameter both the conductance and the capacitance require positive values to be realistic.
Lemma (Permissible) #lem:permissible
For a CBM as defined above the BTC point is biophysically permissable if
From . Hence if to begin with then
$\sign(g_l)=\sign(C_{\mathrm{m}})$
All SNIC neurons can be tuned into Hopf neurons (not necessariliy visa versa)
Proposition (Normal form of the BTC bifurcation) #prop:normal-BTC
The normal form of the degenerated BTC bifucation is
with .
Expand the solution in small variations along the directions spanning the centre manifold: . Then, project the flowfield onto the correpsonding left eignvectors and keep only leading order terms of the nonlinearity. The perturbation evolves according to
.
Remember and and , .
Projecting onto
Projecting onto
Comming back to the type of questions biological theory attempts to answer, see §(???), capters like §(???) describe the mechanistic builing block present in nerve tissue, theire function and where they came from is bracketed off. This is the discussion of proximate causes and the static form. What a bout function?
Note that function is not to be understood in a teleological sense, that it has a purpose given to it by something. All attempts by early German “theoretical biologist” (KE von Baer, JJ Uexkülls, J Reinke or J Schaxel) to use the nonscientific (non empirical) ideas of Hegel, Kant or Goethe as a theoretical bases for biology have proven useless today. Their theories (vitalism, semiotic, teleology) were either incorrect or had no utility, predictive power, empirical basis or falsifiability. They were romantic fancies of an introspective and ultimately authoritarian, self-serving view of the living.
But “why is it like this?”-questions are nagging us. Ultimately, evolutionary considerations, e.g. the contribution of nervous systems to fitness, may come into play. Fitness, however, is an abstract concept difficult to relate to some specific mechanism of a complex nervous system. In part this due to the general problem of fitness being a defined as an expectation value of future offspring.
Surely, one first needs to understand more about the function of nerve tissues, what it does. But where to start?
Nerve tissues can roughly be categorised into
There is no rule that nervous systems have to obay this modular design. These are to some extent arbitrary labels given by scientists to find their way arround.
Let’s briefly go through the three clases and discuss which kind of signals are expected (not an exclusive list).
A classical, ecological theory of sensory organs aims to predict the structural organisation of the peripheral sensory architecture (???; ???), and the coding scheme used to represent information early in the pathway (???), based on the properties of the adequate stimuli occurring in the natural habitat. An alternative approach is inspired by the engineering task of system identification, in which blackbox (technical) systems are probed to obtain a response which helps uncovering the exact transformation executed by them. In contrast to behavioural experiments that utilise abstractions of the actual stimuli perceived by the animal (???) (Cpt. 3), systems identification often relies on unnatural stimuli that, however, possess beneficial statistical properties to facilitate the reverse-engineering scheme. Information theory can be applied within both paradigms as long as there is sufficient data. Examples of such artificial stimuli are a short transient pulse, mathematically idealised by Dirac’s -function (Green’s function method) or ,,white’’ unstructured noise (Wiener’s kernel method). Both signals have a flat, uniform spectrum that contains all frequencies in an unbiased manner and thus the system’s response will not lack structures simply because of missing input complexity. In reality, a uniform spectrum of a physical stimulus can not extend its frequency range to infinity as this would require infinite power. Instead, there will be a reasonable upper bound to the frequency content, . In addition, artificial stimuli are at times chosen from a Gaussian process, for reasons including the invocation of a central limit theorem, arguments based on maximum entropy, or easement of analytical treatments.
One of the first observations that environmental conditions are indeed mapped into nerve pulse activity was an experiment on the toad’s optic nerve (Edgar Adrian 1928).
The external signals from the natural world relevant to animals can often be categorised into the three broad classes: signals relevant to basic body kinematics and movement; signals for predator or prey detection; and communication signals form other (often conspecific) animals. Here are some examples from each category.
Example (Botoxic blowfies and careful crickets) #eg:blowfly-circket
Large monopolar cells, a first order interneuron of the insect compound eye, ecnodes light intensities differences (contrast) into the neurons raded potential response . This was investigated in Calliphora stygia (Laughlin 1981). The cricket Acheta domesticus represents the wind direction in 4 interneurons of the cercal system, each encoding a different subinterval of the angular variable into its firing rate (Theunissen and Miller 1991).
In the above examples the input is a real valued interval , while the encoding neuronal activity is real or a discrete number, or . For example, the cercal system maps . These are seemingly simple mappings, what could possibly go wrong?
A couple of questions arise:
How exactly is this read-out by upstream parts of the nervous system? Will there be errors?
A related question is: How acute is the sense? What difference in wind direction can still be resolved?
How long does the animal have to observe a constant wind direction to be able to encode it? How constant or stationary is the signal in the environment?
What is the relation between the function and the dynamics discussed in previous chapters?
If there is no noise and the coding variable is a real number infinite precission could be possible. If the output is a firing rate, then upstream neurons have to estimate this based on finite amount of observed spikes3. At the same time it needs to be understood how these spike numbers are mechanistically generated from the neuronal dynamics.
Def (spike count) #def:spikecount
.
Note that here the statistical quantity was already connected to a dynamic phase variable for which the governing evolution equations are known in some regimes. Hence, one has related neuronal dynamics and coding (???).
Some signals may not be constant in time, but at least wide-sense stationary, meaning they have temporal correlations, but these are shift invariant.
If the signals in the environment are time dependent functions from some function space, then the neuronal signals representing them could also be functions of time, .
Again some questions:
- But how many degrees of freedom should the function have to represent the signal?
At the dispose are the pulse trains emmited by a neuron. An event time series
Def. (Spike train and empirical measure) #def:spiketrain
The event time series of action potential can be defined in terms of the dynamics, as the spike obaying
.
Further, a spike-train can be written as the empirical measure
- .
Def. (I/O-equivalent phase oscillator) #def:io_phase
An I/O-equivalent phase oscillator produces with its threshold crossing process the same spike times as a particular conductance-based model (Lazar 2010).
Def. (Instantaneous firing rate) #def:instfire
This idealised quantity does not exist in finite neuronal systems. It is the ensemble average
,
and can be interpreted as a infinitely large population of the statistically i.i.d. neurons experiencing the same stimulus or one neuron being stimulated time and again. The probability flux is defined as
. The “advection”, , is due to active transport.
A simplifying observations is that there often (but not always) is a separation of time scales between stimulus induced perturbations and those fluctuations that originate from internal biophysical noise in receptor neurons, see §(???). The three examples clearly fall into this category where the stimulation cutoff renders the stimuli orders of magnitude slower than the intrinsic noise. Therefore, the fast (white) noise, , in the phase equation can be averaged
with .
Continuity equation
A time-continuous signal does not necessary imply an infinite degree of freedom. Deterministic signals like sinusoidal is specified by only two degrees of freedom. In contrast to white noise, a band limited signal, with no power above , has only a limited number of degrees of freedom per second (???). In accordance with Whittacker-Shannon sampling it can be represented as a countable set of random numbers (finite degrees of freedom per unit time). Mutual information estimations can then be based on these instead of the continuous signal without loosing information. The physical signals that announce themselves on sensory organs can have their power distributed over many octaves. However, signals emitted by other animals, either voluntarily like the communication songs of grasshoppers or involuntary as moth falling prey to a bat, are often constrained to a frequency range that is related to the body size of the sound producing animal (see the next paragraph for three examples of relevant natural stimuli). In such occasions one can identify an upper frequency, , above which no relevant information is expected. Additionally, all primary transduction processes have a time constant restricting the signal frequencies they can follow. This is sufficient reason to consider the stimulus to fall into the class of bandlimited signals (also called baseband signals).
More precisely according to the sampling theorem a band limit signal with cutoff frequency requires at least a sampling rate of . With the help of the Whittaker-Shannon interpolation formula the signal can be written as (???)
For an observation interval one needs a minimum of samples to specify the continuous bandlimited process completely.
The coefficients can be continous Gaussian random variables such that one has a bandlimited Gaussian random processs. In order to compare information theoretic bounds on the information transmitted through continuous channels (physical systems) to histogram based methods for the estimation of mutual information the Whittaker-Shannon sampling can be augmented by discretisation of the (Gaussian) random variables . The simplest strategy is to start with a discrete set of binomially distributed random variables, $x_n\sim\binomialdist(k,p)$ with $p={\textstyle\frac 12}$, and use the fact that for large the statistics will be approximately Gaussian. With this there is a correspondence between a time-continuous process on an interval and a discrete binomial vector, so that the histogram frequency can be estimated. For example the blue time continuous process in {+(???)} between 50 and 100ms with Hz is described by the 11 samples marked as green circles. The continuous process is statistically similar to a Gaussian process.
A bandlimited and binomially distributed (, $p_{b}={\textstyle\frac 12}$) signal as an approximation to a Gaussian process (, ). The cutoff frequency was 100~Hz, implying that the entire stimulus over 250~ms is exactly defined by 50 regularly spaced samples, as according to the Nyquist theorem one needs at least two samples per period to define an oscillation. To check if the binomial amplitude distribution is an adequate approximation to a Gaussian, the Shapiro-Wilk test for normality is used. The Samples are accepted as Gaussian with an -level of 0.4 (very small -leves would mean rejection of normality).
In previous lectures the biophysical details of the action potential voltage dynamics in membrane patches were discussed, highlighting its key features: the nonlinear feedback, provided by voltage gated channels; and the inherent stochasticity. These two combined make for a complex system that shirks being amenable to a complete analytical treatment. How can one answer the question ,,Which frequencies in a stimulus is it sensitive to?’’.
A particular useful strategy has been to derive approximate filters4 from model equations that are easier to analyse (???). These models must omit many biophysical details to remain analytically tractable.
The calculation of response filters for neuronal models is a long standing exercise in theoretical biology (???).
A classic approach to investigate which stimulus characteristics are represented in the activity of sensory neurons is to estimate their receptive field or spike-triggered average . The analysis characterises the components of a complex stimulus that are relevant to elicit a spike response or increased firing rate. Hence, the concept is related to that of a visual receptive field and could ultimately cause behavioural reactions. It is, therefore, related to the antedating concept of sign stimuli in ethology (???; ???). The method has been applied to recordings from the visual (???) and auditory (???) system.
One can also turn the paradigm around by applying a simple stimulus such as a delta pulse and characterising the nature of the response, which is then called transfer functions or point spread function in the visual system. The term transfer function stems form signal processing and stresses the fact that the neuron is viewed as a linear filter. It is surprising how complete a description of an inharently nonlinear neuron, a linear response filter can give, despite the fact that neurons contain so many nonlinear elements. In a sense it parallels the success of linear response theory in complex physical problems (???). Of course, there is no one linear filter that describes the whole range of dynamical features of the neuron. But different aspects may be characterised by different filter. In particular, we will linearise around the suprathreshold periodic solution to derive a firing rate filter in . Linearising the voltage response around the resting potential yields quite a different filter that describes the subthreshold response. So essentially it depends on which output variable one analyses.
There are two different filter
This is related to the question “what causes a neuron to spike” or how does observing the neuronal spike response affect the conditional statistics of the stimulus ? After observing the spike train some external stimuli should have been more likely their cause than others. An immediate idea is to look at the first moments of the conditional ensemble. For example, the mean stimulus before a spike in one trial is
$\sta(\tau)=\frac{1}{{N_\text{sp}}}\sum_{k=1}^{N_\text{sp}}x(t_k-\tau)=\frac{1}{{N_\text{sp}}}\int_{-\infty}^\infty{\mathrm{d}}t\;y(t)x(t-\tau)$,
where we used the platonic spike train from Def. #.
To improve the signal to noise ratio, an experimenter might repeat the same stimulus, so that a trial average might smooth out the effect of intrinsic noise. Then, if the number of spikes, , per trial is large and almost constant (modest Fano factor) it may be approximated by
where is the mean firing rate. This gives rise to the term method, see Ref.~ and Ref.~.
Given that the rate causally depends on previous stimulus it can be expanded as a Volterra series in orders of the stimulus magnitude
This implies that also and , as well as the Fokker-Planck operator
, where , and .
The right eigen system of the Fokker-Planck operator, has the fourier basis as solution
with eigenvalues .
the eigenvalues have negative real part and the contribution of the corresponding eigenfunction vanishes for , only the eigen mode , i.e., survives.
Hence, the zeroth order equation (), has a stationary solution . Consequently the zeroth order flux is constant, too. Hence, .
The first order equation is
.
The simplest time dependent soution can be obtained by Laplace transforming and Fourier series expanding this equation, .
With and in the orthogonal Fourier basis
or
From the definition of the flux , the first order equation for the flux is
with
for
The low frequency limit of the transfer function is
.
Consequently, PRCs with a mean can transfer arbitrariliy low frequencies.
.
The latter identity was already established in Eq. (17).
Proposition (DC limit of transfer function) #propo:dc-transfer
In the mean driven regime, the susceptibility of the instantaneous firing rate to very low frequencies is related to the mean component of the PRC and the derivative of the - curve
.
Remark #rem:dc-transfer
As all known neuron models have above their rheobase, PRC are garantued to have a – at least small – mean component and the transfer function is susceptible to low frequencies. There are however quantitative differences between the different onset bifurcations.
Using one obtains . With this
Back to function. But where to start?
During the disucussion of action potential stability “pulse-based communication” using spike events was mentioned as a motivation in §(???). In very general terms, nervous system communication information for the senses to some place where decissions are made and further to places where motor commands are produced and executed, a reflex would be the simplest form (Braitenbergian view [@]). In the literature one finds a theory of communication called information theory (Shannon 1948).
The author of this script is uncertain about the utility of information theory in biology at large (this is always a good reason to give a lecture on it), lit looks exciting but has some short commings: This theory was originally never intended to be used to describe living systems in which information bares meaning and consequences, i.e., the content of a message actually matters for your survival. Information theory deals with optimal compression, lossless transmission of signals irrespective of weather it is relevant or a youtube video. Nontheless a look at a quantitative science of communication may be insightfull.
Information theory has long been in the discussion as an ecological theory upon which to judge the performance of sensory processing (Atick 1992). This led Joseph Atick, Horace Barlow and others to postulate to use this theory to study how nervous systems adapt to the environment. The goal is to make quantitative predictions about what the connectivity in the nervous system and the structure of receptive fields should look like, for instance. In this sense, information theory was hoped to become the basis for an ecological theory of adaptation to environmental statistics (Atick 1992,Barlow (1961)).
It may not seems obvious how the mechanistic theories of dynamical system and the functional theory of inromation should be brought to gehterh, but that is of course the gould. One would like to decide which dynamical features have what kind of function, this will be addressed after the intruduction to information theory (G. B. Ermentrout, Galán, and Urban 2007; Schleimer and Stemmler 2009).
The science of communication is concerned with at least two subtopics: the efficient representation of data (compression); and the safe transmission of data through unreliable channels (coding).
A source (the message) typically runns through the following processing sequence:
Compress | Encode | ||
Channel | |||
Decompress | Decode |
One of the formal assertions of information theory is that these two problems can be addressed separately (without loss of generality or efficiency): Meaning first one compresses by removing redundancies. Then one again adds failsafe redundancy to combat the unreliability.
However convenient for engineering, this does not mean that biological systems have to make use of this fact (source coding and channel coding could be very intertwined).
Also note that many of the mathematical results in information theory are bounds, inequalities not achievable in real systems.
(An ergodic system has an invariant phase volume, which is a necessary condition for an organism to exist, in the sense that it would otherwise transgress phase boundaries and cese to exist)
One may start by wrongly assuming that the function of nerves is to communication information, irrespective of its semanting meaning. In reality, it is more likely that the function is to filter information, i.e., reliably communication the import parts and filter out the irrelevant information (or noise). But to introduce the classical quanities, dispense with the semantics.
If the function of nervous system (in part) is to communicate information, then the repertoire and speed of biological computations are limited by thermodynamic or metabolic constraints (???-). The second will deferred to later chapters, but as an outlook. There is no free lunch. It follows from basic thermodynamics that to obtain information work needs to be performed.
Here we consider a ,,memoryless’’ channel:
The rate with which information can be transmitted over a channel without loss is measured for a discrete time channel or for a continuous time channel. Operationally, we wish all bits that are transmitted to be recovered with negligible probability of error.
A measure of information could be:
The average reduction in the number of binary-questions needed to identify before and after observing .
This would just be the difference in entropies:
${\mathcal I}(X;Y)=H(X)-H(X|Y)= -\sum\limits_{x\in X} p(x)\log_2 p(x) + \sum\limits_{x\in X\atop y\in Y}p(x,y)\log_2p(x|y)$.
This goes by the name of mutual information or transinformation. Remember marginalisation
.
So the mutual information is
${\mathcal I}=-\sum_{x\in X\atop y\in Y} p(x,y)\log_2 p(x) + \sum_{x\in X\atop y\in Y}p(x,y)\log_2\left(\frac{p(x|y)p(y)}{p(y)}\right)$
or
${\mathcal I}(X;Y)=\sum_{x\in X\atop y\in Y} p(x,y)\log\frac{p(x,y)}{p(x)p(y)}$
From the complete symmetry of this quantity we can also write it as
.
The following Figure illustrates how the mutual information is related to respective (conditional) entropies of the input and output ensemble.
We have heard that quantifies the statistical dependence of and , but how is that related to error-free communication? What makes the error is the channel. , however, also depends on the input ensemble . To focus on the properties of the channel one can simply take an ,,optimal’’ input ensemble and define the channel capacity
It will be left to the sender to actually find the optimal input statistics. Note that is a concave function () of over a convex set of probabilities (this is relevant for procedures like the Arimoto-Blahut algorithm for estimating ) and hence a local maximum is a global maximum.
p = []
while
p(
Shannon’s channel coding theorem established that this capacity indeed measures the maximum amount of error-free information that can be transmitted.
A trivial upper bound on the channel capacity is
.
This is due to the maximum entropy property of the uniform distribution in the discrete case:
Some of the signals in nature (and those applied in the laboratory) have continuous alphabets, e.g., light intensity values. Does it make sense to extend the definition of the entropy as
Maybe. Let us see how far one gets with this definition. It is called differential entropy by the way. Through quantisation this can be related back to the entropy of discrete alphabets.
If the is smooth then one associates the probability of being in with
The entropy of the quantised version is
This is problematic as the second term goes to infinity for small quantisations. Formally, if is Rieman integrable, then
Since the infinitesimal limit is taken we can also take to be the number of quantal intervals so that in the limit
$\lim_{\Delta\to0\atop n\to\infty}\ln\Delta\approx n$.
So that an -bit quantisation of a continous random variable has entropy
.
With the mutual information being the difference of entropies the quantisation term vanishes.
A first example can be taken from rate coding5. In terms of our spike trains from Eq. (52), the instantaneous firing rate can be defined as
Def. (Gaussian nonlinear channel) #def:infomax
The channel is defined by the following mapping:
where is zero mean normally distributed .
Which encoding function (e.g., the constrast-response function from above) maximises the mutual information across the channel given that we know the input statistics?
Assumptions:
The mutual information is
The noise entropy does not depend on nor on in this particular channel.
To calculate the entropy of the output random variable not that it is added from two numbers, . This means
.
The last identity used Assumption 1). From the transformation rules of random variables one has or .
Hence,
From assumption 2) we have
Maximum Mutual information given is
or .
This means that under the assumptions above, the contrast-response function should be proportional to the cummulative density function of the input.
Let us pause and ask: What is the maximum entropy distribution for a continuous alphabet?
One searches for that fulfils and , where is the variantional derivative. One finds
.
With the normalisation from we get the normal distribution (, and ).
What did we learn? Well
If we consider the whole spike train (see Eq. (52)) as the ouput, not just its mean input intensity and mean firing rate, we have a continuous time dependent stochastic processes to deal with. Note that the definition of the spike train, if integrated, is related to the empirical distribution function.
If we average infinitely many trials we get the instantaneous firing rate . We will give a mathematical definition of later. Our communication channel looks like that
input signal | neural | response | ||
pathway |
The entropy rate of an ergodic process can be defined as the the entropy of the process at a given time conditioned on its past realisations in the limit of large time
The mutual information rate measures the amount of information a neural pathway transmits about an input signal is the mutual information rate,
between the stochastic process, , and the stochastic response process, . The entropy rate measures the number of discriminable input or output states, either by themselves, or conditioned on other variables.
The mutual information rates, which is the difference between unconditional and conditional entropy rates, characterises the number of input states that can be distinguished upon observing the output. The response entropy rates , for instance, quantifies the number of typical responses per unit time, while is a measure of the decoding noise in the model. If this noise is zero, then the mutual information rate is simply , provided that this is finite.
The conditional entropy rates and , characterising the noise in the encoding and decoding model respectively, are each greater than zero. In information theory, these quantities are also called equivocation. Hence, both the stimulus and response entropy rates, and , are upper bounds for the transmitted information.
This is similar to the quantisation problem. It might be reasonable to drop the term (sometimes this is done, sometimes not). For the one dimensional case we have
or, if we drop the
Any practical consequences?
The best estimator is the mean, so the statisticians say. Therefore a lower bound to the estimation error is given by
The lasst inequality followed from Eq. (55).
Up to an additive constant the entropy of a multivariate Gaussian was67
$H=\frac12\ln|\boldsymbol{K}|=\frac12\tr\ln\boldsymbol{K}=\frac12\tr\ln\boldsymbol{\Lambda}=\frac12\sum_k\ln\lambda_k$.
First let us observe the process for ever , a bi-infinte series with countable elements. The elements of the covariance matrix . The orthogonal eigen-function for the continous covariance operator on are the Fourier bases. It can be shown that in the continuum limit
The result is due to Kolmogorov see also (???,Golshani and Pasha (2010)).
Without a complete probablilistic description of the model the mutual information can not be calculated. And even with a model the involved integrals may not be tracktable. At least two strategies to estimate it exist, though: Either, create a statistical ensemble of inputs and outputs by stimulation, followed by (histogram based) estimation techniques for the mutual information; or, find bounds on the information that can be evaluated more easily. In general, the estimation of mutual information from empirical data is difficult, as the sample size should be much larger than the size of the alphabet. Indeed, each element of the alphabet should be sampled multiple times so that the underlying statistical distribution can, in principle, be accurately estimated. But this prerequisite is often violated, so some techniques of estimating the information from data directly rely on extrapolation (???). The problem becomes particularly hairy when the alphabet is continuous or a temporal processes had to be discretised, resulting in large alphabets.
Another approach, which will allow us to perform a theoretical analysis of phase dynamics, relies on a comparison of the neuronal ``channel’’ to the continuous Gaussian channel (???,Cpt.~13) is analytically solvable (Cover and Thomas 2012). The approach can be used to estimate the information transmission of neuronal models (???). Also experimental system have ben analysed in this way, e.g.:
It was prooven that in that this method leads to a guaranteed lower bound of the actual information transmitted (???).
If one has experimental control of the stimulus ensemble it can choosen to be a Gaussian process with a flat spectrum up to a cutoff as to not introduce biases for certain frequency bands. The mutual information between stimulus and response can be bound from below as
Here, is the equivocation of a process with the same mean and covariance structure as the original decoding noise, but with Gaussian statistics. The conditional entropy of the stimulus given the response is also called reconstruction noise entropy. It reflects the uncertainty remaining about the stimulus when particular responses have been observed.
It turns out that the inequality in Eq. (59) also holds if the estimator is conditioned. Say from the output of the neuron we estimate its input
So if the process has a stationary variance
The second line uses the fact that in this case the optimal estimator is given by the conditional mean. We have the following bound on the equivocation
The deviation between stimulus and its estimate, , is treated as the noise process.
In order to obtain a tight bound the estimator should be as close to optimal as possible. For the case of additional information given by the response of the neural system to the process , the estimator should make use of it, . For simplicity one can assume it is carried out by a filtering operation, specified later (Gabbiani and Koch 1998). Like the whole system the noise process is stationary, and its power spectral density, , is
$H_\text{gauss}[x|y]\leqslant {\textstyle\frac 12}\ln\langle n^2(t)\rangle={\textstyle\frac 12}\int_{-\infty}^\infty\frac{{\mathrm{d}}\omega}{2\pi}\ln P_{nn}(\omega).$
Together
So as to render the inequality in Eq. (61) as tight a bound as possible one should use the optimal reconstruction filter from to . In other words, it is necessary to extract as much information about from the spike train as possible.
The next step should be to find an expression for the noise spectrum, , based on the idea of ideal reconstruction of the stimulus. As opposed to the forward filter, the reconstruction filter depends on the stimulus statistics (even without effects such as adaptation). We seek the filter that minimises the variance of the mean square error
Taking the variational derivative of the error w.r.t.
the filter (coefficients) and equating this to zero one obtains the orthogonality condition for the optimal Wiener filter (???)
Inserting the definition of the error, , into Eq. (64) yields
In order to obtain we need to deconvolve the equation, which amounts to a division in the Fourier domain
To compute the mutual information rate, we now calculate the full auto-correlation of the noise when the filter is given by Eq. (65). For an arbitrary filter , we have
.
When the orthogonality condition of Eq. (64) holds, the right-most term vanishes. Proceeding by expanding the first term algebraically leads to an expression for the noise correlations
.
This expression can be Fourier transformed in order to obtain the required noise spectrum
,
where the definition of the optimal filter, Eq. (65), was utilised. This result can then be inserted into Eq. (62) to obtain the following well known bound on the information rate (???,lindner2005j:mi,holden1976b,stein1972j:coherenceInfo)
This information bound involves only spectra and cross-spectra of the communication channel’s input and output processes which are experimentally measurable in macroscopic recordings . The channel, in this case the neuron, can remain a black box. But since we can bridge the divide between microscopic, biophysical models and their filtering properties, we will, in the following section, derive the mutual information rates.
- .
It quantifies the linearity of the relation between and in a way that it equals 1 if there is no noise and a linear filter transforms input to output. Both nonlinearities and noise reduce the coherence. The coherence can be estimated from data using the FFT algorithm and spectrum estimation. It is implemented in the free software packages
scipy
andmatplotlib
.
What renders the coherence a useful quanity? While the cross-spectrum informs us when stimulus and output have correlated power in a spectral band, the normlisation with the output auto-spectrum can be crucial. Say we find a particular high power in , this may not be ralted to the stimulus but could just be the intrinsic frequency of the neuron itself.
The coherence is a quantity without mention of the explicite decoding filter, in fact it is symmetric in input and output just as the mutual information. This is beneficial because one can now take the encoding view in the next chapter.
The stimulus spectral density is given by the environment or controlled by the experimental setup, while cross- and output spectra need to be measured or calculated from the model in question. In this lecture cross-spectral and spike train spectral densities are derived from phase oscillator, see Eq. ((???)), that are in turn derived from biophysical models. This means we do not treat the channel as a blackbox but assume a particular model.
The first quantity we need to calculate Eq. (67) is the cross-spactrum. On the one hand it is the Fourier of the cross-corrlation, on the other it can be written as averages of the Fourier transforms (FT and average are linear operation).
What has happened here? The cross-spectrum can be obtained by averaging the Fourier transformed quantities over trials and the stimulus ensemble. The average can be split into the conditinal average over trials , given a fixed, frozen stimulus and the average over the stimulus ensemble, (). The former is essential an average over the encoding noise (Chacron, Lindner, and Longtin 2004,Lindner, Chacron, and Longtin (2005)).
Observe that is Fourier transform of the trial averaged firing rate conditional on a frozen stimulus
.
Thus, it is sufficient to derive a filter that maps input to a firing rate, not an individual spike train.
The filter is causal, since it is implemented by a differential equation and the Laplace transform yields
where denotes the transfer function of the encoding filter.
With this definition the cross-spectrum is written as
This shows us that although we are computing the cross-spectrum of stimulus and spike train the response filter for the instantaneous firing rate suffices. This simple relation reminds us of the fact that the cross-spectrum is not really a second order quantity, but can be exactly determined by linear response theory. The spike train spectrum , on the other hand, is truly a second order quantity, viz, the Fourier transform of the auto covariance, and can not be related to the linear response filter without further approximations.
The instantaneous firing rate can be estimated via binning and trial averaging
: | 0 | 3 | 1 | 0 | 0 | 0 | 0 | 3 | 1 | 0 | 0 | 0 | 1 | 2 | 1 | 0 |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Trials: | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 |
0 | 0 | 1 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | |
0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | |
0 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 |
Two equivalent descriptions8 of Markov dynamics exist
?.
For (i) one can simulate many paths of a stochastic differential equation, with different intitial conditions and noise realisations. Histograms can provide the ensemble statistics. But it is also possible to find an evolution equation for the whole ensemble.
The relation between the two can be formally established by the
With all the lovely properties of a -function.
The probability density
where the average is over all paths, , and therefore over all realisations of the noise process .
The chain rule yields
Solving such a PDE involves time integration or other integral transformations (Fourier and Laplace’s). Since
Therefore
Averaging on both sides results in
.
The correlation between a stochastic process and a nonlinear functional of it is given by the Novikov-Furutsu-Donsker formula
All to gether we have the
This is a diffusion equation. It can be written in the form
One needs boundary conditions and initial conditions to solve this PDE.
For convenience rewrite the I/O-equivalent phase oscillator from Eq. ((???)) as
Here, as opposed to Eq. ((???)) was split into the part that results from the presented stimulus, now denoted , and the part that originated from, e.g. intrinsic noise. From Eq. (27) the perturbation vector has
.
As long as the intrinic noise is fast enough compared to the stimulus an averaging can be applied9 to obtain an effective diffusion
,
which enters Eq. (79). The benefit is that the corresponding The Fokker-Planck equation
is tractable in a perturbation expansion. But first, remember what is the goal: Identification of the forward filter in .
The equation is solved with the following perturabtion ansatz
with the normalisation requirement
The are assumed small correctin terms, given that the stimulus is also small.
at the same time one gets
In perturbation theory one solves iteratively the equations of the same order
.
This equation is homogeneous so we find a time independent steady state solution
.
One may test this by back insertion. Here both the boundary conditions and Eq. (82) are inforced. The solution can be inserted into the definition of the flux to obtain the zeroth order flux
The next order involves the time-dependent stimulus
To turn the PDE into an algebraic equation one can apply both the Fourier series expansion and the Laplace transform. For this the Laplace transform of the stimulus is denoted as and the periodic function
Solving for
For brevity define the pole
The first order flux is
and
Happily and consistently one finds
The power spectrum correponds to the imaginary axis, . The low frequency limit is
.
With , the high frequency limit is
.
Adrian, E.D., and Y. Zotterman. 1926. “The Impulses Produced by Sensory Nerve-Endings. Part 2. the Reponse of a Single End-Organ.” J Physiol 61.
Adrian, Edgar. 1928. The Basis of Sensation: The Action of the Sense Organs. Christophers.
Atick, Joseph J. 1992. “Could Information Theory Provide an Ecological Theory of Sensory Processing?” Network: Computation in Neural Systems 3 (2): 213–51. http://www.tandfonline.com/doi/abs/10.1088/0954-898X_3_2_009.
Barlow, HB. 1961. “Possible Principles Underlying the Transformations of Sensory Messages.” In Sensory Communication, edited by WA Rosenblith, 217–34. MIT Press.
Chacron, Maurice J., Benjamin Lindner, and André Longtin. 2004. “Noise Shaping by Interval Correlations Increases Information Transfer.” Physical Review Letters 92 (8). doi:10.1103/PhysRevLett.92.080601.
Chicone, Carmen. 2006. Ordinary Differential Equations with Applications. Springer Science & Business Media.
Cover, Thomas M., and Joy A. Thomas. 2012. Elements of Information Theory. John Wiley & Sons.
Doyle, D. A. 1998. “The Structure of the Potassium Channel: Molecular Basis of K+ Conduction and Selectivity.” Article. Science 280 (5360): 69–77. doi:10.1126/science.280.5360.69.
Ermentrout, Bard. 1996. “Type I Membranes, Phase Resetting Curves, and Synchrony.” Neural Computation 8 (5): 979–1001. http://www.mitpressjournals.org/doi/abs/10.1162/neco.1996.8.5.979.
Ermentrout, G. Bard, Roberto F. Galán, and Nathaniel N. Urban. 2007. “Relating Neural Dynamics to Neural Coding.” Article. Physical Review Letters 99 (24). American Physical Society (APS). doi:10.1103/physrevlett.99.248103.
Fox, Ronald F., and Yan-nan Lu. 1994. “Emergent Collective Behavior in Large Numbers of Globally Coupled Independently Stochastic Ion Channels.” Physical Review E 49 (4): 3421–31. doi:10.1103/PhysRevE.49.3421.
Gabbiani, Fabrizio, and Christof Koch. 1998. “Principles of Spike Train Analysis.” Methods in Neuronal Modeling 12 (4): 313–60. http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.383.7571&rep=rep1&type=pdf.
Goldwyn, Joshua H., and Eric Shea-Brown. 2011. “The What and Where of Adding Channel Noise to the Hodgkin-Huxley Equations.” PLoS Comput Biol 7 (11): e1002247. doi:10.1371/journal.pcbi.1002247.
Goldwyn, Joshua H., Nikita S. Imennov, Michael Famulare, and Eric Shea-Brown. 2011. “Stochastic Differential Equation Models for Ion Channel Noise in Hodgkin-Huxley Neurons.” Physical Review E 83 (4): 041908. doi:10.1103/PhysRevE.83.041908.
Golshani, Leila, and Einollah Pasha. 2010. “Rényi Entropy Rate for Gaussian Processes.” Information Sciences 180 (8): 1486–91. doi:10.1016/j.ins.2009.12.012.
Hodgkin, Alan L., and Andrew F. Huxley. 1952. “A Quantitative Description of Membrane Current and Its Application to Conduction and Excitation in Nerve.” The Journal of Physiology 117 (4): 500. http://www.ncbi.nlm.nih.gov/pmc/articles/pmc1392413/.
Izhikevich, Eugene M. 2007. Dynamical Systems in Neuroscience. MIT Press.
Kielhöfer, Hansjörg. 2011. Bifurcation Theory: An Introduction with Applications to Partial Differential Equations. Springer Science & Business Media.
Kirst, Christoph. 2012. “Synchronization, Neuronal Excitability, and Information Flow in Networks of Neuronal Oscillators.” PhD thesis, Niedersächsische Staats-und Universitätsbibliothek Göttingen. http://ediss.uni-goettingen.de/bitstream/handle/11858/00-1735-0000-000D-F08D-2/kirst.pdf?sequence=1.
Kirst, Christoph, Julian Ammer, Felix Felmy, Andreas Herz, and Martin Stemmler. 2015. “Fundamental Structure and Modulation of Neuronal Excitability: Synaptic Control of Coding, Resonance, and Network Synchronization.” biorxiv;022475v1. http://biorxiv.org/lookup/doi/10.1101/022475.
KUZNETSOV, YU. A. 2011. “PRACTICAL Computation of Normal Forms on Center Manifolds at Degenerate BogdanovTAKENS Bifurcations.” Article. International Journal of Bifurcation and Chaos 15 (11): 3535–46. doi:10.1142/S0218127405014209.
Laughlin, S. 1981. “A Simple Coding Procedure Enhances a Neuron’s Information Capacity.” Zeitschrift Fur Naturforschung. Section C, Biosciences 36 (9-10): 910–12.
Lazar, Aurel A. 2010. “Population Encoding with Hodgkin–Huxley Neurons.” IEEE Transactions on Information Theory 56 (2): 821–37. doi:10.1109/TIT.2009.2037040.
Linaro, Daniele, Marco Storace, and Michele Giugliano. 2011. “Accurate and Fast Simulation of Channel Noise in Conductance-Based Model Neurons by Diffusion Approximation.” PLoS Comput Biol 7 (3): e1001102. doi:10.1371/journal.pcbi.1001102.
Lindner, Benjamin, Maurice J. Chacron, and André Longtin. 2005. “Integrate-and-Fire Neurons with Threshold Noise: A Tractable Model of How Interspike Interval Correlations Affect Neuronal Signal Transmission.” Physical Review E 72 (2). doi:10.1103/PhysRevE.72.021911.
Orio, Patricio, and Daniel Soudry. 2012. “Simple, Fast and Accurate Implementation of the Diffusion Approximation Algorithm for Stochastic Ion Channels with Multiple States.” PLoS ONE 7 (5): e36670. doi:10.1371/journal.pone.0036670.
Osinga, Hinke M., and Jeff Moehlis. 2010. “Continuation-Based Computation of Global Isochrons.” SIAM Journal on Applied Dynamical Systems 9 (4): 1201–28. doi:10.1137/090777244.
Pereira, Ulises, Pierre Coullet, and Enrique Tirapegui. 2015. “The Bogdanov–Takens Normal Form: A Minimal Model for Single Neuron Dynamics.” Entropy 17 (12): 7859–74. doi:10.3390/e17127850.
Pezo, Danilo, Daniel Soudry, and Patricio Orio. 2014. “Diffusion Approximation-Based Simulation of Stochastic Ion Channels: Which Method to Use?” Frontiers in Computational Neuroscience 8: 139. doi:10.3389/fncom.2014.00139.
Schleimer, Jan-Hendrik, and Martin Stemmler. 2009. “Coding of Information in Limit Cycle Oscillators.” Article. Physical Review Letters 103 (24). American Physical Society (APS). doi:10.1103/physrevlett.103.248105.
Shannon, C. E. 1948. “A Mathematical Theory of Communication.” Article. Bell System Technical Journal 27 (3). Institute of Electrical; Electronics Engineers (IEEE): 379–423. doi:10.1002/j.1538-7305.1948.tb01338.x.
Theunissen, F. E., and J. P. Miller. 1991. “Representation of Sensory Information in the Cricket Cercal Sensory System. Ii. Information Theoretic Calculation of System Accuracy and Optimal Tuning-Curve Widths of Four Primary Interneurons.” Article. Journal of Neurophysiology 66 (5). American Physiological Society: 1690–1703. doi:10.1152/jn.1991.66.5.1690.
Winfree, A. T. 1974. “Patterns of Phase Compromise in Biological Cycles.” Journal of Mathematical Biology 1 (1): 73–93. http://www.springerlink.com/index/0668278661N54486.pdf.
Actually it does, if you bring energetic considerations into baring.↩
For now this means that its an invertable matrix; ↩
Estimators should be at least consistent, hopefuly unbiased, maybe efficient.↩
The biological literature often uses dynamical systems jargon to describe mechanistic explanation and resorts to engineering vocabulary in order to describe the function of nerve cells.↩
Here, the alphabet is a firing rate . It might be more reasonable to think about spike counts in a given time window, which is still a countable set.↩
Remember the determinant is . So . In terms of the matrix-logarithm and the trace the determinant can be expressed as $|\boldsymbol{K}|=\exp\tr\ln\boldsymbol{K}$.↩
Because the trace is invariant under similarity transforms $\tr\boldsymbol{K}=\sum_{i=1}^n\lambda_i$.↩
Similar to the particle and wave duality in quantum mechanics.↩
This can be done formally using Chapman-Enskog or adiabatic elimination procedures. The derivation may be included in to future.↩