Aug 13, 2011 - Springer Science+Business Media, LLC 2011. Abstract We ... and that this transition accounts for changes
J Comput Neurosci (2012) 32:327–346 DOI 10.1007/s10827-011-0355-7
Reduction of stochastic conductance-based neuron models with time-scales separation Gilles Wainrib · Michèle Thieullen · Khashayar Pakdaman
Received: 16 January 2011 / Revised: 13 June 2011 / Accepted: 19 July 2011 / Published online: 13 August 2011 © Springer Science+Business Media, LLC 2011
Abstract We introduce a method for systematically reducing the dimension of biophysically realistic neuron models with stochastic ion channels exploiting timescales separation. Based on a combination of singular perturbation methods for kinetic Markov schemes with some recent mathematical developments of the averaging method, the techniques are general and applicable to a large class of models. As an example, we derive and analyze reductions of different stochastic versions of the Hodgkin Huxley (HH) model, leading to distinct reduced models. The bifurcation analysis of one of the reduced models with the number of channels as a parameter provides new insights into some features of noisy discharge patterns, such as the bimodality of interspike
intervals distribution. Our analysis of the stochastic HH model shows that, besides being a method to reduce the number of variables of neuronal models, our reduction scheme is a powerful method for gaining understanding on the impact of fluctuations due to finite size effects on the dynamics of slow fast systems. Our analysis of the reduced model reveals that decreasing the number of sodium channels in the HH model leads to a transition in the dynamics reminiscent of the Hopf bifurcation and that this transition accounts for changes in characteristics of the spike train generated by the model. Finally, we also examine the impact of these results on neuronal coding, notably, reliability of discharge times and spike latency, showing that reducing the number of channels can enhance discharge time reliability in response to weak inputs and that this phenomenon can be accounted for through the analysis of the reduced model.
Action Editor: Alain Destexhe This work has been supported by the Agence Nationale de la Recherche through the ANR project MANDy “Mathematical Analysis of Neuronal Dynamics” ANR-09-BLAN-0008. G. Wainrib (B) Department of Mathematics, Stanford University, Building 380, Serra Mall, Stanford, CA, USA e-mail:
[email protected] M. Thieullen Laboratoire de Probabilités et Modèles Aléatoires, UMR7599, Boîte 188, Université Pierre et Marie Curie 4, Place Jussieu, 75252, Paris Cedex 05, France e-mail:
[email protected] K. Pakdaman Institut Jacques Monod, UMR7592, Bâtiment Buffon, 15 rue Hélène Brion, 75205, Paris Cedex 13, France e-mail:
[email protected]
Keywords Stochastic ion channels · Time-scales separation · Averaging method · Hodgkin-Huxley
1 Introduction Since the seminal work of Hodgkin and Huxley (1952), biophysically realistic neuron models are based on a detailed description of the gating dynamics of various ionic currents. The original four variables Hodgkin Huxley (HH) model has inspired a plethora of conductance-based neuron models with many more variables. Theoretical analysis of these high dimensional models, a prerequisite for understanding their functional response properties, can be mathematically
328
challenging, as it is already the case for the original HH model (see for instance Rubin and Wechselberger 2007). Indeed, the higher the dimension of the model and the more difficult it becomes to have an intuitive representation of the dynamics. Furthermore, numerical simulations of large networks of high-dimensional models can pose computational difficulties. One way to partly overcome both theoretical and numerical hurdles has been through model reduction, that is, the design of lower dimensional models that retain essential characteristics of the original high dimensional ones. Dimension reduction for these models is often achieved by taking advantage of the separation of gating variables time-scales (Rinzel 2944; Kepler et al. 1992). Thanks to reduction, a deeper understanding of dynamical properties of neuronal models has been gained using singular perturbation theory (e.g. Suckley and Biktashev 2003; Rubin and Wechselberger 2007) and phase plane analysis of two-dimensional models (Izhikevich 2007). All these reduction schemes are designed for deterministic neuronal models. None of them takes into account channel noise arising from the intrinsic stochasticity of ion channels. The purpose of this paper is to present a reduction scheme for stochastic neuronal models with channel noise described by kinetic Markov schemes (Hille 2001; White et al. 2000). Reductions and singular perturbation theory have been the cornerstone in the analysis of the dynamics of deterministic neuronal models. Similarly, one expects a stochastic reduction method to provide a means to explore and gain insights into the interplay between intrinsic noise and dynamical properties of neurons. For an infinite number of ion channels, the deterministic models are obtained by a law of large numbers from Pakdaman et al. (2010). However, when the number of ion channels is finite, the remaining fluctuations may be the source of many noise-induced phenomena (Chow and White 1996; Steinmetz et al. 2000; Schmid et al. 2003; Shuai and Jung 2003; Rowat 2007) which have putative functional consequences in neuronal coding and information processing. Therefore, it is of high relevance to develop a systematic theoretical approach to analyze these stochastic phenomena. As we will show in this work, reduction of conductance-based neuron models with stochastic ion channels provides a method to study noise-induced phenomena with the well developed techniques of deterministic systems. An important example in the deterministic setting is the classical reduction of HH, setting m = m∞ (V) where m is sodium activation, which is for instance, the starting stage of the analysis carried in Rubin and Wechselberger (2007). Here we examine the stochastic counterpart to this deterministic reduction: the steady-
J Comput Neurosci (2012) 32:327–346
state approximation is replaced by an averaging over the quasi-stationary distribution of the fast variable. We show that for HH, the stochastic reduction yields not only standard terms, but also modifies the deterministic model. These modifications clarify the scenario of noise-induced changes in discharge patterns. In other words, reduction transforms the analysis of the dynamics of a stochastic system into that of a deterministic dynamical systems. In this way noise induced qualitative changes of dynamics are brought within the well established field of bifurcation analysis of deterministic systems. This is all the more remarkable that reduction does not require any hypothesis on the noise intensity, that is, it is not confined to small or large noise limits as other theoretical analyses of stochastic systems often do. In this sense, reduction is a powerful tool, not only for decreasing the dimension of a model, but to gain understanding on the interplay between fluctuations and nonlinearities in slow fast systems. The analysis of the reduced model reveals the existence of a transition in the dynamics of the model as the number of channels is decreased. Noise induced transitions in excitable systems have been reported in all classical neuronal models, such as the leaky integrate and fire model (Pakdaman et al. 2001), the FitzHugh Nagumo model (Tanabe and Pakdaman 2001), the Morris–Lecar model (Tateno and Pakdaman 2004), the active rotator and the Hodgkin–Huxley model (Tanabe and Pakdaman 2001; Takahata et al. 2002). These transitions have been analyzed through derivation of deterministic bifurcation diagrams and their consequences for neuronal coding, such as noise induced enhancement of discharge time precision and reliability, have been discussed (Tanabe et al. 1999; Pakdaman et al. 2001; Tanabe and Pakdaman 2001). The present work shows that a similar phenomenon occurs when the source of the noise is channel fluctuations rather than external. However, it is important to emphasize that this does not constitute a mere extension of previous results, because, unlike external noise, channel noise is inherently implicated in the nonlinear threshold-like response of neurons so that our approach and analysis methods are altogether different from previous studies. In terms of methods, the question of stochastic reduction based on a separation of time-scales has been investigated in other fields, especially for biochemical reaction networks mainly through expansions of master equations (Mastny et al. 2007 and ref. therein). These situations are simpler than neuronal membrane models in which stochastic jump processes are coupled with differential equations. Our systematic stochastic reduction scheme is adapted for such hybrid systems that combine differential equations and jump processes. It
J Comput Neurosci (2012) 32:327–346
329
is based upon a singular perturbation approach to multiscale continuous-time Markov chains (Yin and Zhang 1998) combined with a recent development of the averaging method for piecewise-deterministic Markov processes, see Faggionato et al. (2010) and Pakdaman et al. (2010, submitted). The averaging method has been widely used in the analysis of systems perturbed by fast deterministic oscillatory motion (Arnold 1983). In stochastic systems, the situation is compounded by the random nature of the fluctuations, even though, in essence, the steps retain some similarity to the deterministic case. Before going into details in Sections 2 and 3, we provide here a schematic view of our approach. First, multiscale analysis of kinetic Markov schemes enables the introduction of fast sub-processes and of an aggregated slow process accounting for the jumps between the fast sub-processes. The fast evolution within each sub-process is replaced by its quasi-stationary behavior. Then, averaging the slow dynamics with respect to these quasi-stationary distributions leads to an effective reduced stochastic hybrid system at the limit of a perfect separation of time-scales. However, averaging over the quasi-stationary measure leads to an error. A key step in our reduction approach is to estimate this error deriving a second-order term in the form of a diffusion approximation (Pakdaman et al. 2010, submitted). This article is organized as follows. We first treat the reduction of stochastic versions of the HH model in order to illustrate in a specific example the three following points. (1) To describe the two types of new terms, namely a deterministic and stochastic one, that can appear in reduction which have no counterpart in deterministic reduction. (2) To present in a step by step manner the stages leading to the computations of these terms. (3) To illustrate the type of information regarding neuronal dynamics that may be obtained from the reduced model and that is not readily accessible in the full model. The general reduction scheme for arbitrary stochastic conductance based neuronal models that requires the manipulation of numerous notations, is described in full detail after treating this example. We hope that this particular layout, which starts from the analysis of a specific example rather than the depiction of the general scheme is more helpful to the reader and for potential applications of stochastic reduction in other models.
An overview of the full models and their reductions is given in Fig. 2.
2 Reduction of HH models
Stochastic two-state model TS N The gating variables m and h are treated separately. Gates m and h are assumed to follow open-closed schemes as depicted in Fig. 1 with opening and closing rates αm (V), βm (V)
In this section we present the result of our reduction approach on two stochastic versions of the HH model.
2.1 Full models Deterministic model HH We recall here the classical deterministic HH model (Hodgkin and Huxley 1952), defined by the following system of differential equations: dV = F(V, m, n, h) dt dm = τm−1 (V)(m∞ (V) − m) dt = (1 − m)αm (V) − mβm (V) dn = τn−1 (V)(n∞ (V) − n) = (1 − n)αn (V) − nβn (V) dt dh = τh−1 (V)(h∞ (V) − h) = (1 − h)αh (V) − hβh (V) dt with F(V, m, n, h) := C−1 [I −g L (V −V L )−g Na m3 h(V− V Na ) − g K (V − V K )] where g L , g Na , g K , V L , V Na , V K are respectively the maximal conductances and reversal potential associated with leakage, sodium and potassium ionic cu rrents, and I is the input current. The functions αx , βx , τx and x∞ for x = m, n, h are given in Appendix A. We now introduce stochastic versions of HH with a finite number of ion channels: as metastable molecular devices subject to thermal noise, their stochastic conformational changes are usually (Hille 2001) modeled by kinetic Markov schemes (KMS). For the first version, called the “two-state” model TS N , the limit of infinite number of channels N → ∞ (by increasing the area of the membrane patch and keeping the densities constant) gives exactly the deterministic system HH (Pakdaman et al. 2010). However, this is not the case for the other version, called here the “multistate” model MS N , introduced in Skaugen and Walloe (1979) and frequently used in the literature. Indeed, its deterministic limit MS∞ does not match with HH, unless a specific binomial relationship between the initial conditions is satisfied (see Section 4.2 for more details). This difference, which may seem innocuous for the deterministic reduction, has a substantial impact in the stochastic case as will become clear in the following, and we think that understanding this difference is important from a modeling perspective.
330
J Comput Neurosci (2012) 32:327–346
each one following an open-closed scheme with rates αn (V), βn (V) for x = n. Denoting x N the proportion of open gates x, for x = m, n, h, the equation for V is: V˙ = F(V, m N , h N , n N )
Fig. 1 Kinetic scheme for the gate m
(1)
Here N stands for the triple (Nm , Nh , Nn ). for x = m and αh (V), βh (V) for x = h. If the number of sodium channel is N Na then the number of gates m is Nm = 3N Na and of gates h is Nh = N Na . Similarly, the potassium channel is composed of four gates,
Fig. 2 Overview of the full models and their reductions. To account for the time-scale separation (the time-constant of sodium activation “m” is small compared to the other time-constants), we have introduced a small parameter . On the left column are stated the full models and on the right their reduction when → 0. More precisely, on the first line, the deterministic H H model is reduced to a 3D model. On the second and third lines, two stochastic versions of HH with a finite number of ion channels are considered. On the second line, the stochastic two-state model T S N, is reduced to a model T S N,0 where an additional deterministic term has appeared. Is also included a corrective stochastic term for small . On the third line, the stochastic multistate model MS N, (sodium and potassium channels follow multistate kinetic schemes) is reduced to a model MS N,0 where no additional term appear. A corrective stochastic term is also included for small . See text for more details
Stochastic multistate model MS N In this model, introduced in Skaugen and Walloe (1979), each of the N Na sodium channels follows an eight-state Markov scheme SNa as shown in Fig. 2.
Full models
Reduced models
J Comput Neurosci (2012) 32:327–346
331
In the equation for V the product m3 h corresponding to the sodium conductance is replaced by the proportion of channels in the state m3 h1 : u Na (t) =
N Na 1 1m3 h1 cnNa (t) N Na n=1
(2)
where each cnNa (t) follows independently the scheme SNa for n ∈ {1, .., N Na }. Similarly, each of the Nk potassium channel variables cnK (t) is assumed to follow independently a five-state Markov scheme SK shown in Fig. 2. The term n4 in the equation for V is replaced by the proportion of channels in the state n4 :
present the results and then summarize the method and explain how to obtain each term in Section 3. The general framework is exposed in Section 6. Stochastic reduction of the two-state gating model Applying the general theory (cf. Section 6), we derive the reduced model TS N,0 when = 0: ¯ n) ¯ h, ¯ V˙¯ = F¯ N (V,
(7)
where F¯ N is defined by: F¯ N (V, h, n) := C−1 (F(V, m∞ (V), h, n) − g Na h(V − V Na )K N (V)) and
u K (t) =
NK
1 1n4 cnK (t) N K n=1
(3)
K N (V) :=
3 m∞ (V)2 (1 − m∞ (V)) Nm +
Eventually, the equation for V reads: (4)
:= F MS (V, u Na , u K )
(6)
−g K u K (V − V K ) − g Na u Na (V − V Na ) (5)
Here N stands for the pair N = (N Na , N K ). 2.2 Reduced models Time-scale separation As noticed in Suckley and Biktashev (2003) about the deterministic model HH, the time-scale of the variable m is faster than the others and we introduce a modified model HH (with a small positive parameter) where τm (V) is replaced by τm (V) (equivalently αm (V) and βm (V) are replaced by −1 αm (V) and −1 βm (V)). Similarly, we define TS N, and MS N, the associated modified stochastic models. We assume the reader is familiar with deterministic reduction techniques, leading to the following reduced deterministic system HH0 when → 0, substituting m by its steady-state m∞ (V) in the equation for V: dV = F(V, m∞ (V), n, h) dt dn = τn−1 (V)(n∞ (V) − n) dt dh = τh−1 (V)(h∞ (V) − h) dt We now proceed to the reduction of the two stochastic models TS N, and MS N, introduced above. We first
As before, the variables h¯ and n¯ follow their usual kinetic Markov scheme with transition rates function ¯ of V. The expression of F¯ N (V) contains an additional deterministic term, namely K N (V), that depends explicitly on the number of channels. This term comes from the average of m3 against the quasi-stationary binomial distribution (cf. formula (9)). Thus, considering the case of a finite number Nm of stochastic sodium gates leads to a new term in the effective description, as a result of the cubic non-linearity. In Fig. 3, the corrective term Icor = −g Na h(V − V Na )K N (V) is displayed as a function of V for a given value of h and for different values of Nm . When V is very negative (resp. very positive), most of sodium channels are closed (resp.
20 Corrective term (µA/cm2 )
dV = C−1 I − g L (V − V L ) dt
1 m∞ (V)(1 − 3m∞ (V) + 2m∞ (V)2 ) 2 Nm
15
10
5
0 -100
-50
0
50
100
V (mV)
Fig. 3 Corrective term Icor = −g Na h(V − V Na )K N (V) appearing in the reduced two-state model TS N,0 . Bold Nm = 200; Regular Nm = 500; Thin Nm = 1,000. Parameter h = 0.6
332
J Comput Neurosci (2012) 32:327–346
open), so that fluctuations due to sodium channel noise are very small, which is consistent with the bell shape of the corrective term. This term is always positive in the region V < V Na = 115 mV, and acts as an additional positive voltage-dependent current which depolarizes the membrane. This suggests that for Nm low enough, this corrective term can induce regular firing, through a bifurcation path similar to the classical bifurcation of the HH model with I as a parameter. We will investigate further this phenomenon in Section 4.1. For small , a correction term can be included and a switching diffusion approximation is obtained in the form: ¯ ¯ ¯ ¯ ¯ ¯ h)dW ¯ dVt = F N (V, h, n)dt + BT S (V, t Here BT S depends on Nm and its computation is explained in Section 3, paragraph 5 with more details in Appendix B. Stochastic reduction of the multistate gating model To operate the reduction when → 0 of the multistate model, the SNa scheme is reduced into a two-state open-closed scheme for h, whose transition rates, αh (V) and βh (V), are unchanged because independent of the state of m. However, when averaging the vector field of V, it is necessary to replace the term m3 h1 by the quasi stationary distribution at m = m3 multiplied by h, which gives m∞ (V)3 h. Thus, taking the limit → 0 for the multistate model leads to a reduced system with m = m∞ , as the usual deterministic reduction, with stochastic h following a two-state scheme (and n following the SK scheme). Moreover, when is small it is of interest to incorporate a corrective term in the form of switching √ diffusion correction of order on the equation for the variable V. The computation of variance of this corrective term (see Sections 3 and 6) gives the following approximation: ¯ n¯ dt ¯ h, ¯ m∞ (V), dV¯ = F V,
¯ h¯ dWt + B MS V, N Na
(8)
with 2 1 3 4 ¯ h¯ = C−1 g Na h¯ V¯ − V Na B MS V, m V¯ τm V¯ 3 ∞ 2 2 3 × 18αm βm + 9αm βm + 2βm ¯ The other variwhere αm and βm are evaluated at V. ¯ ables n¯ and h are the empirical measures associated
with two-state kinetic Markov schemes, respectively ¯ with transition rates αn , βn and αh , βh , functions of V. The obtained expression is consistent with the approximation obtained in Chow and White (1996). Such an approximation may have many applications, both theoretical (to estimate the variability in exit problems) and numerical (the simulation of jump process is very heavy when is small, compared to this switching diffusion approximation). These comparisons of the forms of the deterministic and stochastic reduced models highlight the key following point. In the deterministic case, we need only to replace the activation variable of the sodium current by its stationary value. No other changes are brought to the equation. In the stochastic case, there are two types of changes. One is the possible addition of a deterministic term at the limit = 0 and the other is √ the correction term for the fluctuations, of order for small . This latter term replaces fluctuations due to the jump process (channel noise) with a diffusion term. Furthermore, the terms obtained depend on the KMS model of channel noise.
3 The reduction algorithm Here we summarize the reduction steps, with comments referring to the reduction of the stochastic HH models presented above. The general theory behind this reduction method is given in Section 6. 1. Identify the time-scale separation and partition the state space In order to identify the fast transitions, one has to compare the magnitudes of the kinetic rates present in the system (higher rates correspond to faster transitions). Two related issues are expected to be encountered: first, the transition rates are function of the membrane potential so it is not straightforward to compare them, second the membrane potential dynamics has its own time constant. A solution suggested in Suckley and Biktashev (2003) is to compute τx := | ∂∂x dx | for dt each variable x. Once the fast transitions have been identified in a given scheme (such as in the S Na scheme) one has to partition the state space into fast subsets. In the case of the model TS N, , the partition is trivial since there is only one transition in the two-state scheme for m, which is fast. 2. Find the quasi-stationary distributions within fast subsets For a single channel following the twostate model with rates proportional to −1 , the stationary distribution characterizes the limiting behavior when → 0 and is given (in the case of
J Comput Neurosci (2012) 32:327–346
m) by ρT S = (1 − m∞ (V), m∞ (V)). In the multistate model, within a fast subset the stationary distribution is given by: ρ MS = ((1 − m∞ (V))3 , 3(1 − m∞ (V))2 m∞ (V), 3(1 − m∞ (V))m∞ (V)2 , m∞ (V)3 ). 3. Average the transition rates to find the slow reduced kinetic scheme This step concerns only the multiscale model: one has to construct a reduced slow process which aggregates the slow transitions between states of the fast subsets (for instance transitions between mi h0 and mi h1 , 1 ≤ i ≤ 3, in the S Na scheme). Each fast subset is identified with a state of this slow process. In a general setting, the transition rates for this reduced slow process are given by a specific average with respect to the quasistationary distribution (cf. coarse-graining formula in Section 6.2). In the case of the S Na scheme, the reduced slow process obtained by this approach is exactly the classical two-state process h, since all the vertical transitions are identical. 4. Average the vector field for the reduced dynamic of V In this second averaging step one is looking for a reduced equation for V. Let us describe the example of the two-state model TS N, . From the quasi-stationary distribution for a single twostate process (cf. 2 above), one deduces the binomial quasi-stationary distribution for the fraction of open gates m N : (Nm ) ρopen
k Nm
333
for the multiscale scheme, one has to solve for a fixed value of V (if possible analytically) the system ˙ = Am (V)M for various initial conditions with M ⎛ ⎞ βm 0 0 −3αm ⎜ 3αm −2αm − βm 2βm 0 ⎟ ⎟ Am = ⎜ ⎝ 0 2αm −2βm − αm 3βm ⎠ 0 0 αm −3βm and then compute integrals of the form ∞ R= (M(t) − ρ MS ) dt 0
(see formula (20) in Section 6. for more details) which characterize the convergence towards the stationary distribution. We have done this computation using a formal calculus software to obtain ¯ in Eq. (8). As far as the two-state ¯ h) the term B(V, model is concerned, the vector field is non-linear with respect to m and the computations should be done directly on the law of the empirical measure. The R integrals are difficult to compute analytically and we give in Appendix B an expression in terms of special functions. Then, the variance can be obtained by formula (21).
4 Numerical analysis of the reduced models = CkNm m∞ (V)k (1 − m∞ (V)) Nm −k (9)
Then, averaging the term m3N with respect to this distribution one obtains the classical term m∞ (V)3 together with the additional term K N (V). In the multiscale model, the dependence of the vector field F MS upon u Na is linear: in this case no additional term appears and one recovers only the value of the quasistationary distribution ρ MS corresponding to the state m3 h1 , which is precisely m∞ (V)3 . In the case of a non-linear dependence with a multiscale scheme, one has to work carefully to derive the equivalent of the binomial quasi-stationary distribution (cf. the formula for m distribution in Section 6.2). 5. Compute the variance of the corrective term This step is the more computationally involved. In the multistate model, the linearity of F MS with respect to u Na reduces the complexity of this computation (cf. Section 6.3 for more details). One has to apply formula (21) of Section 6.3, which requires to solve the full evolution of the law for each fast subset, and not only the stationary distribution. Namely,
In this section, we analyze the reduced HH models obtained in Section 2, showing how an analysis of the two-state reduced model can be useful to understand the stochastic bifurcations of the original model (Section 4.1). Moreover, as the reduction method gives two different results for seemingly similar stochastic HH models, we compare the results obtained for both stochastic models (Section 4.2). 4.1 Bifurcation analysis and ISI bimodality We have seen that additional terms appear when reducing the stochastic two-state TS N, model. To clarify their impact, we analyze a system where the variables n and h evolve according to their deterministic limits: dn = (1 − n)αn (V) − nβn (V) dt dh = (1 − h)αh (V) − hβh (V) dt while m evolves according to a two-state kinetic scheme (cf. Fig. 1. Indeed, we have chosen to introduce this semi-deterministic model in order to isolate the
334
consequences of the additional terms. First, there is a slight change in the value of the equilibrium point. Then, these terms have consequences in terms of bifurcations. The reduced HH system with m = m∞ (V), corresponding to the case N = ∞, has a stable equilibrium point for I < I H , a Hopf bifurcation occurs at I = I H and a double cycle bifurcation occurs at I = Idc < I H . In the region (Idc , I H ) the system is bistable: a stable equilibrium point and a stable limit cycle coexist. For I > I H an unstable equilibrium point coexists with a stable limit cycle. The additional terms change the values Idc and I H . Incorporating the additional term K N (V) into the model modifies this bifurcation diagram. In this term, the number of channels N = 1/η appears as a parameter. We can analyze systematically its impact on the dynamics of the model through the computation of bifurcation diagrams as is done for any parameter in a deterministic model. The first notable change that is observed is that, even in the absence of any external current (i.e. I = 0), reducing the number of channels (i.e. increasing η = 1/N) leads to a bifurcation diagram similar to the one obtained when varying I in the reduced deterministic model. More precisely, as η is increased, the stable equilibrium loses is stability through a sub-critical Hopf
(a)
Fig. 4 Bifurcation diagrams for system TS N with deterministic n and h variable. (a) Bifurcation diagram with η as parameter for I = 0. The line crossing the figure left to right represents the equilibrium potential. The point is stable for low values of η (thick solid line) and becomes unstable (dashed line) after undergoing a subcritical Hopf bifurcation at η H ≈ 0.01944. The maximal and minimal voltage values of a branch of unstable periodic orbits initiated at the Hopf bifurcation are represented by the dotted line. This branch goes through a double cycle bifurcation as it folds and gives place to a branch of stable periodic orbits (maximal and minimal voltages represented by a solid line).
J Comput Neurosci (2012) 32:327–346
bifurcation an we observe the same three qualitatively distinct asymptotic dynamics. For I = 0, there is a Hopf bifurcation when varying the parameter η = N1 at a value η H ≈ 0.01944 corresponding to N H = 51 (see Fig. 4(a)). A double cycle bifurcation seems to occur for η1 ≈ 0.0151, that is N1 = 66. Thus, for N ≤ N1 = 66, the reduced two-state HH model Eq. (7) exhibits repetitive firing, with a bistable region for N H < N ≤ N1 . We can consider the influence of I on this diagram. This is done through the computation of the twoparameter bifurcation diagram. It reveals that the loci of the Hopf and double cycle bifurcations mode to lower channel numbers (i.e. higher η) as more current is injected. Ultimately these loci meet with fold bifurcation, probably at a Takens–Bogdanov (TB) point. Even though such bifurcations have been encountered in the standard HH model (Holden et al. 1985), here we consider that the putative TB bifurcation occurs at very low channel numbers that may be of little interest. The bifurcation analysis of the deterministic system, with the noise level η = 1/N as a parameter, helps understanding the behavior of the stochastic model TS N, (with deterministic n and h).
(b)
(b) Two-parameter bifurcation diagram with I and η as parameters. The solid and dashed lines represent, respectively, the loci of the Hopf and double cycle bifurcations. The lines in the upper left part of the figure represent loci of two folds merging into a cusp. In the region (1) to the left of the double cycle, there exists a single stable equilibrium point. In the region (3) to the right of the Hopf, an unstable stable equilibrium coexists with a stable limit cycle. In the region (2) delineated by the double cycle and Hopf curves, the system is bistable (coexistence of a stable equilibrium and stable limit cycle)
J Comput Neurosci (2012) 32:327–346
335
Indeed, if is small enough, the dynamics of the stochastic HH model with a finite number of channels are expected to resemble those of the deterministic reduced model perturbed by a small stochastic switching diffusion. In this way, the stochastic HH model should produce different forms of dynamics depending on the number of channels, and these dynamics and the changes between them should be according to the bifurcation scenario described above. More precisely, in the regime where the reduced model displays a single stable equilibrium point that is globally stable, the stochastic system displays fluctuations around the equilibrium with only rare noise induced excursions away from this state (Fig. 5(c)). In contrast, when the reduced model stabilizes at a limit cycle, the stochastic model generates a nearly periodic train of action potentials (Fig. 5(a)). The transition between these two dynamics corresponds to the bistable regime. In this case, the stochastic model generates bursts of nearly periodic spikes separated by phases of quiescence, the transitions between the two corresponding to noise induced transitions from the neighborhood of the limit cycle to that of the equilibrium and vice versa (Fig. 5(b)). The interspike interval histograms of these regimes (Fig. 6) confirm the description given above illustrated through representative waveforms of the membrane potential V and is consistent with the bifurcation analysis of the reduced model: 1. Below the double cycle curve is a region with a unique stable equilibrium point: ISI distribution is approximately exponential, since a spike corresponds to a threshold crossing. 2. Between the double cycle and the Hopf curves is a bistable region: ISI distribution is bimodal, one peak corresponding to the escape from the stable equilibrium, and the other peak to the fluctuations around the limit cycle. 3. Above the Hopf curve is a region with a stable limit cycle and an unstable equilibrium point: ISI distribution is centered around the period of the limit cycle. We have checked this prediction with extensive and systematic numerical simulations. We summarize the results below with representative numerical illustrations, where we have set the number of gates n and h arbitrarily at 104 in order to reduce the impact of the other sources of noise and to focus on the effect of m. Equivalent results are obtained with deterministic gates n and h. Configurations with small Nm and large Nh are not completely realistic since one should keep Nm = 3Nh to account for the structure of the sodium channel. However, it provides a interesting setting to check the
Fig. 5 Sample trajectories of the stochastic two-state gating model TS N, with N gate m, and 104 gates n and h with I = 0 and = 0.1. (a) With N = 30 (zone 3), the sample path looks close to a noisy periodic trajectory, as expected. (b) With N = 70 (zone 2), the expected bimodality of ISI’s appears clearly on this sample path. (c) With N = 120, ISI statistics are closer to a poissonian behavior; however spikes may occur by pair as in this particular sample, which corresponds the peak in the ISI distribution (cf. Fig 6, N = 100)
336
J Comput Neurosci (2012) 32:327–346
Fig. 6 Interspike Interval (ISI) distribution for the stochastic two-state gating model TS N, with N gate m, and 104 gates n and h with I = 0 and = 0.1. These figures show the apparition of bimodal ISI distribution as N is increased. For N = 30, the distribution is unimodal and corresponds to a noisy limit cycle behavior. For higher values of N, bimodality appears and can be interpreted as the signature of bistability for the underlying deterministic system. For N ≥ 100 the bimodality tends to disap-
pear towards an exponential behavior, corresponding the escape from a stable equilibrium point. The transition from a noisy limit cycle to an exponential behavior through a bimodal distribution is in line with the bifurcation diagram in Fig. 4. ISI histograms obtained from stochastic simulations of TS N, with a modified Gillespie algorithm Gillespie (1977); White et al. (2000). Note the different scales for the time-axis
theoretical results obtained so far with the averaging principle. We have also checked the predictions 1, 2 and 3 with higher values of I still below I H , for which the averaging-induced double cycle and Hopf bifurcations occur at lower values of η (higher Nm ). In this case, one can keep the ratio Nm = 3Nh and still see the spontaneous quasi-periodic firing, since noise coming from h gates is small enough. We have also studied the impact of relaxing the assumption 0, the solution XtMS of MS∞, satisf ies (Xt ) = XtMS . As an example, consider the multistate scheme of Fig. 7 described by the deterministic kinetic equations: dm2 = −2βm2 + αm1 dt dm1 = 2βm2 − (α + β)m1 + 2αm0 dt dm0 = βm1 − 2αm0 dt and the two-state scheme described by
(10) (11) (12)
dm = (1 − m)α − mβ (13) dt With (m) = m2 , 2m(1 − m), (1 − m)2 , if (m(0)) = (m2 (0), m1 (0), m0 (0)), then for all t ≥ 0, (m(t)) = (m2 (t), m1 (t), m0 (t)). This result means that there exists an invariant manifold on which both deterministic limits coincide. In this sense, both stochastic HH models can be considered as compatible with the original HH model as long as one is interested in the dynamics for theses special initial conditions. In Keener (2009), another result is obtained in a slightly different setting: assuming the transition coefficients to be only function of time, and forgetting V, it is shown that the corresponding invariant manifold is globally attracting under eigenvalues conditions. We now investigate how above coincidence property of the deterministic equations translates into the stochastic models TS N, and MS N, . Although their deterministic limit when N → ∞, fixed coincide, we know from Section 2. that their limit when → 0, N
Fig. 8 Four-state kinetic scheme for the m gates
fixed are different. One may expect the behavior of MS N, to be similar to TS N, . Even if the reduction of MS N, for → 0, N fixed is actually independent of N (Eq. (8)), one may expect the analysis carried out in Section 4.1 to be still relevant, because of the coincidence property of the limits at N → ∞. In particular, it is natural to ask whether the ISI bimodality persists in the MS N, model. In order to compare only the effects associated with the variable m, we consider an intermediate multistate model IMS N, , where h and n follow an open-closed scheme as in the two-state model, and where the gates m follow a four-state kinetic scheme, as depicted in Fig. 8. (A number N of these gates corresponds to 3N two-state gates m). Obtained from stochastic simulation of IMS N, for different values of N and , ISI distributions for I = 3 are displayed in Fig. 9. Bimodality is not present for = 0.1, in contrast with the two-state model (Fig. 6), but appears for a lower value of = 0.01. This phenomenon may seem unexpected: when → 0 the model IMS N, is closer to MS N,0 (actually independent of N), which does not exhibit bistability for I = 3. However, there may be a competition between the speed of convergence to the invariant manifold and the speed of convergence to the reduced model, so that one can suggest the following hypothesis. Smaller may increase the attractivity of the invariant manifold corresponding to the two-state model: bimodality for IMS N, when is small may thus be explained by the bistability of TS N,0 even if MS N,0 is not in a bistable regime. As has been shown above, reduction of the two stochastic variants of the HH model, namely , the TS and MS models leads to different results. These reduced models differ both in their deterministic and stochastic components. This result is all the more interesting that whether at the level of a single channel or at the level of a large membrane, the two variants produce similar dynamics. More precisely, at the level of a single channel, the relevant information is the jump process between the open and closed states. Whether this channel dynamics is modeled according to the TS or the MS formalism, has no observable impact as the two jump processes have exactly the same law. In other words, the jump statistics cannot be used to differentiate one model from the other. In this sense the two models are equivalent.
338
J Comput Neurosci (2012) 32:327–346
Fig. 9 Interspike Interval (ISI) distribution for the intermediate multistate model (HH,N I MS ) with N = 10, 20, 30, 40 gates m, and 104 gates n and h, with I = 3. Bold line: = 0.001, dashed line: = 0.01, thin line: = 0.1. As in Fig. 4, these figures show the apparition of bimodal ISI distribution as N is increased for small enough = 0.01 and = 0.001 (solid lines). For higher = 0.1, the distributions are unimodal
Similarly, given a large membrane patch, the system is in the deterministic limit. In this regime, even though the MS model has a higher dimension than the TS, the two models have similar dynamics. Indeed, the trajectories of the MS approach the invariant manifold upon which the dynamics are those of the TS. Differences between MS and TS dynamics are only transient and tend to vanish with time. In other words, after discarding such transients, it is not possible to distinguish the two models. Based upon the above considerations, one can argue that the two variants are equivalent at the two extreme cases of a clamped single channel or a membrane patch with infinitely many channels. Reduction reveals that in the intermediate stage where there are finitely many channels, the situation is different. The dynamics of the TS and MS are not necessarily equivalent. It results from the interplay between channel noise and the non linear dynamics of the system and depends explicitly on the Markovian model selected for the description of channel noise. In this sense, reduction appears as a method to expose these differences and provide a means to investigate their consequences on neuronal dynamics.
5 Effects of channel noise on neuronal coding After deriving the reduced stochastic HH models and analyzing their dynamics, in this section, we discuss the relevance of the results in terms of neuronal coding. Neurons encode incoming signals into trains of stereotyped action potentials. One of the issues in neuronal coding has been to analyze the extent in which the timing of the individual spikes is altered by the presence of internal noise due to the stochastic fluctuations of ionic channels. In this section, we use the reduction scheme to account for one of the counterintuitive effects of channel noise on discharge time reliability of neurons, namely that for certain classes of inputs, reliability can go through a maximum by increasing channel noise (i.e. by decreasing the number of channels). Previous studies have investigated the impact of channel noise on discharge time reliability. For instance, based upon numerical simulations, it has been reported in Schneidmann et al. (1998) that the reliability is higher for rapidly varying input signals than for slow ones. In Jung and Shuai (2001) and Schmid et al. (2001), a phenomenon of intrinsic stochastic resonance has been observed by numerical simulations, suggesting an
J Comput Neurosci (2012) 32:327–346
optimal number of channels for information transmission and reliability of discharge patterns. Here, our aim is to illustrate how thanks to reduction one can account for a counterintuitive effect of channel noise on discharge time reliability. We take on the simplest setting that is used in previous studies of neuronal reliability, namely the response of neurons to constant DC currents, and that allows us to highlight the contribution of reduction. Clearly, this particular selection is not limiting. More precisely, we consider the response of the stochastic TS N, model to a step current I, for various numbers Nm . To highlight the impact of the number of sodium channels and the ensuring transition on neuronal coding, we have also represented in Fig. 10 the raster plots of simulations of stochastic HH models with various sodium channel numbers. All simulations were started with the potential at the resting state and the unit receiving a step current of amplitude I = 1 μA/cm2 at time zero. This current intensity is subthreshold in the sense that in the deterministic limit (large number of sodium and potassium channels), it does not evoke any discharges. For stochastic models, the situation is different because, due to channel fluctuations, such units display spontaneous firing. The panels of Fig. 10 show that for large numbers of channels, when the models are close to the deterministic regime, firing reliability is low in the sense that the discharge times in the raster plot are very poorly vertically aligned. However, as the number of channels is decreased, the reliability improves and the first discharge times of the units become strongly aligned. At very low channel numbers, the first spikes are well aligned however, reliability of later discharges deteriorates. In summary, we see that firing reliability goes through a maximum as channel number is decreased. We have used a constant step current to highlight this phenomenon, first because of its relevant to neuronal coding and the fact that it has been used to analyze and discuss neuronal reliability (Cecchi et al. 2000) and it is also simple enough to fully illustrate our purpose, that is, the transition discovered thanks to reduction can have substantial impact on neuronal reliability. The above phenomenon is also observed when one keeps the ratio between the number of h gates and of m gates to on third as discussed below in the case of latency coding. The stochastic transition described in Section 4.1 can have a significant impact in terms of neuronal coding. To illustrate this proposition, we have focused on latency coding, where information is
339
Fig. 10 Each panel represents the raster plot for a given number of sodium m gates N = Nm of the response of 50 units to a constant DC current of amplitude I = 1 μA/cm2 applied from time t = 0. Each line on each panel represents the discharge times obtained from a single simulation run of the two-state model TS N, . All runs, whether within each panel or from one panel to another, were carried out independently. The initial condition is the equilibrium of the deterministic model without input. Parameters Nh = 103 , Nn = 104 and = 0.1
encoded in the latency between stimulus onset and the first action potential. In Wainrib et al. (2010), the impact of channel noise on latency coding has been studied, focusing on large numbers of ion channels. We take on the simplest setting that is used in previous
340
J Comput Neurosci (2012) 32:327–346
1.6
analyzed in Shuai and Jung (2005), where an explanation is provided in terms of elementary arithmetic properties of the number of channels. This example illustrates how qualitative changes in the dynamics of the reduced model account for the differences in reliability observed in the full model, shedding in particular a new light on a phenomenon of intrinsic stochastic resonance in small patches of excitable membrane.
CV of the latency to first-spike
1.5 1.4 1.3 1.2 1.1 1 0.9 0.8 0.7 0
20
40
60
80 100 120 140 Number of m gates (Nm)
160
180
200
6 General theory 6.1 Models
Fig. 11 Coefficient of variation (CV) for the latency to first spike as a function of the number of sodium channels N Na , in the subthreshold case (Iext = 1 μA/cm2 ) for the two-state model TS N, with = 0.1. Number of gates: Nm = 3Nh = 3N Na and Nn = 104 . The initial condition is the equilibrium of the deterministic model without input
As a stochastic interpretation of HH like models, we consider a general class of conductance-based neuron models with stochastic ion channels as follows: –
studies of neuronal reliability, namely the response of neurons to constant DC currents, and that allows us to highlight the contribution of reduction. Clearly, this particular selection is not limiting. More precisely, we consider the response of the stochastic TS N, model to a step current I, for various numbers of sodium channels N Na = Nh with Nm = 3Nh . We show here how the reduction method helps understanding how relatively low number of channels can be beneficial in terms of latency variability. Indeed, numerical simulations displayed in Fig. 11 reveal that, contrary to what one may expect, for low values of Nm (that is higher channel noise) the coefficient of variation (CV) of latency to first spike can be lower. Thus, we see that reduction helps in understanding some of the counterintuitive effects of noise. In this example, reduction reveals that for large Nm , the input current is subthreshold, and the distribution of latency is expected to be close to exponential (CV close to 1), while it becomes suprathreshold at lower Nm . More precisely, for low values of Nm , the analysis of Section 4.1 shows that the reduced system has a unique stable limit cycle, which induces a low value of the CV, while for intermediate values of Nm the system is bistable, and the resulting bimodality of latencies distribution accounts for the increase of the CV above 1. At very low values of Nm , CV is increased due to an increase in channel noise, leading to a optimal number of channels in terms of latency variability. Notice that another phenomenon is visible in Fig. 11, namely the sawtooth-like behavior of the CV as a function of Nm . Such behavior has been previously
–
–
For each ion type, labeled by k ∈ {1, .., d}, we consider a population of Nk associated ion channels. Within a given population k, each ion channel, constituted of several gates, is modeled by one or several KMSs with voltage-dependent transition rates. A KMS is defined as a jump Markov process by a finite state space E and transition rates αi, j(V) between two states i, j ∈ E. We further make an assumption of time-scale separation which is stated below. The conductance induced by the population k is a function of the empirical measures of the underlying KMSs. Typically, the empirical measures represent the proportion of open channels. If u denotes the vector of all empirical measures, then we assume that there is a function hk and a constant g¯ k such that the conductance associated with population k is given by gk = g¯ k hk (u). The membrane potential V(t) evolves dynamically according to: dV = I(t) − g¯ L (V − V L ) dt −
d
g¯ k hk (u)(V − Vk )
(14)
k=1
:= f (V, u)
(15)
where I(t) is the input current, gi and Vi the conductance and resting potentials corresponding to the leak current for i = L and to different ionic currents for each value of k ∈ {1, .., d}. The theoretical results are still valid for any smooth function f .
J Comput Neurosci (2012) 32:327–346
341
Our time-scale separation assumption (TSA) is the following. We assume that one KMS, have the following property: There exist n subsets of fast transitions. In other words, the state space may be partitioned as:
one recovers (Pakdaman et al. 2010) the deterministic conductance-based neuron models. In the present paper, we do not take this limit and the analysis is valid for any number of channels.
E = E1 ∪ E2 ∪ .. ∪ En
6.2 Reduction
where the sets {Ei }i=1,..,n are def ined such as:
We are interested in the limit of perfect time-scale separation, namely → 0. Let us first explain how to handle a single KMS under TSA, and then we will see next how to apply this to reduce a full neuron model with stochastic ion channels. We denote X the considered process under TSA, and we introduce the quasi-stationary distributions (ρ ij) j∈Ei within fast sub-sets Ei , for i ∈ {1, .., n}. We construct an ¯ on the state space E¯ = {1, .., n} aggregated process ( X) with transition rates: α¯ k,l = ρxk αx,y (coarse-graining) (16)
– –
Fast transitions within each Ei : if i, j ∈ Ek then αi, j is of order O( −1 ), Slow transitions between fast sub-sets: otherwise, if i ∈ Ek and j ∈ El , with k = l then αi, j is of order O(1).
The elements of E will be denoted by ei,k where i ∈ {1, ..., N} labels the fast sub-set and k ∈ Ei the position within the fast sub-set Ei . Thus, if for a given KMS satisfying the TSA, we denote by A (V) the transition matrix defined by Ai, j(V) = αi, j(V) if i = j and Ai,i (V) = − αi, j(V), j =i
then we can decompose this matrix as: A (V) = A S (V) + −1 A F (V) where A S accounts for the slow transitions and A F is a block matrix accounting for the fast transitions (each block corresponding to a subset Ei ). A typical example of KMS which satisfies the TSA is the sodium kinetic scheme (see Fig. 2) commonly used to study the impact of channel noise in the HH model. The other important assumptions concern the ergodicity of the fast sub-processes, which is ensured by an irreducibility assumption since we are in the case of finite state spaces.
x∈Ek y∈El
¯ for k, l ∈ E. This aggregated process behaves on a slow timescale, and since we have averaged out the transition rates against the quasi-stationary distribution, the fast time-scale has now disappeared. With such an aggre¯ one can approximate time-averages gated process X, T T ¯ ¯ X(s))ds where of the form 0 ψ(X (s))ds by 0 ψ( i ¯ ψ(i) := k∈Ei ψ(ei,k )ρk . Now, a consequence is that if the process X is coupled with a differential equation dV /dt = F(V , X ) which is on a O(1) time-scale, then one obtains a reduced differential equation: dV¯ ¯ X ¯ (averaging) = F¯ V, dt ¯ ¯ with F(v, i) := F(v, k)ρki for v ∈ R and i ∈ E. k∈Ei
Notation We denote the vector of empirical measures by u = (u1 , ..., ud ). Each uk , corresponding to population k, may gather several empirical measures as well: ui = (u11 , ..., uk1 1 ) (cf. m and h for the sodium channel in j the TS model). Amongst one of the uk we assume that j one of them, say uk00 = uTSA is the empirical measure for a population of KMSs under TSA. We denote:
˜ uTSA u = u,
where u˜ gathers all the other empirical measures, behaving on the O(1) time-scale. Remark If one considers that the number of channels Nk are all proportional to the area A of the patch and that the densities are fixed, then when A is large
Remark As recently shown in Faggionato et al. (2010), this convergence result when → 0 is valid even if the transition rates for the jump process are also voltagedependent (this setting is called piecewise-deterministic Markov process). The remaining step is to obtain the quasi-stationary distribution for the empirical measure associated with the population of jump processes under TSA. Indeed, in general the coarse-graining and the averaging steps should be done directly on the empirical measure and not on each single KMS. Thus, we explain next how to find these quasi-stationary distributions. To clarify the presentation, we first assume that the KMS satisfying the TSA has only fast transitions. In other words, with the notations above, n = 1 and
342
J Comput Neurosci (2012) 32:327–346
E ≡ E1 . Namely, if one considers a population of N KMSs satisfying the TSA, the quasi-stationary distribution for the proportion of KMSs in a state i ∈ E is given by a binomial formula:
with
ρi(N) (c) = CcN (ρi )c (1 − ρi ) N−c
and
n E¯N := (ξk )nk=1 ∈ I N /
where (ρi )i∈E is the stationary distribution for the single KMS. We have thus expressed the stationary distribution for the empirical measure in terms of the one for the single KMS. We now proceed to the averaging of the vector field with respect to this binomial distribution. For the vector field f given by Eq. (15), this leads to the following reduced equation when → 0:
E N (ξ ) :=
d
(18)
k=1
with h¯ (N) := k
N
– ˜ c/N)ρi(N) (c) hk (u,
and u˜ the vector with
c=1
all the empirical measures except the one on the fast time scale which has been considered for the averaging. A key remark here is that the number of channels N has appeared in the reduced equation (18) for V. As illustrated in Section 4, this enables a bifurcation analysis with N as a parameter. We now turn to the case of a population of multiscale processes. The aim is to apply directly the general coarse graining result, summarized by formula (16), on the empirical measure, and not on each KMS of the population. So, we consider a population of multiscale processes Xi , where each Xi satisfies the TSA, and we define: N U N (t) := 1 Xi (t)=ei,k (19) i=1
⎧ ⎨
d (xi,k )1≤i≤n, k∈Ei ∈ I N /
⎩
n ! i=1
E = E1 ∪ .. ∪ En , |E| = d , |Ei | = di –
i,k
This process also has a multiscale structure:
ξ ∈E¯N
1≤k≤di
xi,k = ξi
⎫ ⎬ ⎭
i i between states x and x + δk,l where δk,l is a vector d in {−1, 0, 1} with a “−1” at position (i, k) and a “+1” at position (i, l) and 0 elsewhere. Such a transition means that one of the processes in the population has undergone a jump between the state ei,k ∈ Ei and the state ei,l ∈ Ei . The rate of such a transition depends upon the value of ξi which counts the number of processes in Ei . An important quantity associated with each fast sub-process θξ is its stationary distribution, which can be obtained, after a slight combinatorial work, as a product of multinomial distributions, called m distribution:
πξ (x) :=
Partitioning the state space: One can make a partition of the state space E N as EN = E N (ξ )
μik,l := −1 xi,k αei,k ,ei,l
where we recall that
–
ξk = N
Here, an element ξk counts the number of processes in the population that are in the fast subset Ek . Then, given a constraint ξ ∈ E¯N , we denote by E N (ξ ) the set of all possible configurations. The key point here is that ξ is invariant when one of the underlying processes in the population undergoes a fast transition: ξ changes only when a slow transition happens. Fast sub-processes: For each ξ ∈ E¯N we define a fast sub-process θξ which is characterized by its state space E N (ξ ) and its transition rates
ei,k ∈E
and that ei,k are the elements of E, with 1 ≤ i ≤ n x and 1 ≤ k ≤ di . For x ∈ N∗ , we denote I N := {0, ..., N}x . The process U N is a jump Markov process on the state space ⎧ ⎫ ⎨ ⎬ d E N := xi,k 1≤i≤n, k∈Ei ∈ I N / xi,k = N ⎩ ⎭
n k=1
(17)
dV¯ g¯ k h¯ (N) = I(t) − g¯ L (V − V L ) − k (V − Vk ) dt
i xi,1 i xi,di ξi ! ρ ... ρdi xi,1 !..xi,di ! 1
for x ∈ E N . Slow transitions between aggregated states: Following the coarse-graining formula (16), we are now able construct a slow process U¯ N on the reduced state space E¯N . The transition between states ξ and ξ = ξ + δk,l (where δk,l ∈ {−1, 0, 1}n has a “−1” at position k, a “+1” at position l and 0 elsewhere) occurs with rate α¯ ξ,ξ := πξ (x)μ(x, y) x∈E N (ξ ) y∈E N (ξ )
where μ(x, y) :=
xi,k αei,k ,e j,l
i∈Ek , j∈El
–
The averaging of vector field F is made with respect to the m quasi-stationary distributions πξ (x).
J Comput Neurosci (2012) 32:327–346
Special case In the case where the vector field F depends linearly on the multiscale empirical measure of interest (for instance in the stochastic multistate model MS N, ), the result is more straightforward. The essence of reduction is to approximate, for small , time averages of the form: 1 I f = f (u(s/))ds 0
by ergodic averages I f = f (u)ρ(du) where u denotes a stationary ergodic process with invariant measure ρ and satisfying appropriate mixing conditions. Thus, in the linear case, reduction only involves timeaverage of the type 1 1 U N (t)dt N 0 which can be written as a sum of elementary averages of the type 1 1 X j (t)=ei,k dt
343
term is included as a switching diffusion of order that is added on the equation for V: ˜ u)dt ¯ dV = f¯N (V, u, +
√ ˜ u)dW ¯ σ f (V, u, t
√
(20)
with u¯ the reduced process with rates given by the coarse-graining formula (16). The mathematical aspects of this approximation are studied in a separate paper (Pakdaman et al. 2010, submitted), which extends results of Faggionato et al. (2010) and Yin and Zhang (1998). To give the expression of σ f , some further notations are required. In order to keep them as handy as possible, we consider that the population of processes under TSA will be denoted by the generic notation X . This means that we pretend the population of N processes is just a single process. We have shown below that this is justified since we have exhibited a multiscale structure for the empirical measure as well. So instead of the complicated notations designed for the state space of empirical measures E N , we have chosen to work with the generic notations. We introduce the associated sub-processes xi defined on each subset Ei . Let us rescale this sub-processes by putting = 1 and consider the laws of xi1 : 1 1 P(i) V (x, y, t) := P xi (t) = y | xi (0) = x forx, y ∈ Ei
0
Each of these terms can be approximated (using the general coarse-graining theory) by: 1 ρki 1 X¯ j (t)=i dt 0
In other words, the time spent by X j in state ei,k can be approximated by the product of the time spent by X j in the subspace Ei times the stationary distribution ρki of the fast-process evolving within Ei . For instance, in the eight-state scheme S Na , the time spent in the state m3 h1 can be approximated when is small by the time spent in any of the state mx h1 for x = 0, 1, 2, 3 multiplied by the quasi-stationary probability associated with the state m3 , that is m∞ (V)3 . As in the multiscale model MS N, , F MS depends linearly on m3 h1 , the averaged vector field is F MS ¯ which is precisely the drift term evaluated at m∞ (V)3 h, obtained in Eq. (8). 6.3 Fluctuations As the time-scale separation is not perfect in biological models, it is interesting to be able to include a corrective term to account for the error made when considering the reduced model at the limit = 0. This
We then define a quantity that accounts for the discrepancy between the law PV (x, y, t) and the quasistationary distribution ρV(i) (y): R(i) V (x, y) :=
∞ 0
(i) P(i) (x, y, t) − ρ (y) dt V V
The diffusion coefficient σ 2f is then given by: ˜ u) ¯ 2= σ f (V, u,
¯ ˜ x) f (V, u, ˜ y)L(Vu) f (V, u, (x, y) (21)
x,y∈Eu¯
where ¯ ¯ ¯ ¯ ¯ L(Vu) (x, y) := ρV(u) (x)R(Vu) (x, y) + ρV(u) (y)R(Vu) (y, x)
Computation of R(i) V (x, y) involves not only the quasistationary distribution, but also the full time-course of the law of the empirical measure. In the linear case one only needs to do this work for a single process under TSA and not for the empirical measure. Indeed, with the linearity, one obtains an average of N diffusion terms, one for each member of the population, and gathering them gives a single diffusion scaled as √1N . This is how we have obtained the diffusion correction in Eq. (8).
344
7 Discussion In the deterministic case, reduction has a geometric interpretation in the phase space of the dynamical systems corresponding to the neuronal model. The aim of reduction is to derive a system of differential equations with a lower number of variables that is quantitatively close to the original system. The essence of this reduction is that in some sense there are redundant variables in the original model. Geometrically, in the phase space of the system, this translates into the fact that there exists a manifold that is invariant under the flow of the dynamical system and that attracts trajectories. Describing the dynamics on this manifold requires less variables than there are in the original model, yet this lower dimensional dynamical system is totally accurate. In such a case one obtains a perfect reduction of the original model. In practice, it is not always possible to obtain either the invariant manifold or the dynamics on it perfectly. Instead, one constructs an approximate reduced system that is quantitatively close (in some mathematical sense) to the perfect reduced model. The procedures of reduction of HH models offer methods to obtain such approximations. In the stochastic case, interpretation of reduction is more complicated. On the one hand, one may suggest a sample path interpretation: the trajectories of the original model stay close with high probability to the trajectories of the reduced averaged system. On the other hand, since the law obeys a deterministic equation—possibly infinite-dimensional—such as the master equation, a singular perturbation approach for the law would provide a geometric interpretation in the law space. However, it remains unclear how to relate this geometry with the properties of the trajectories. In this paper, we have provided a method to reduce stochastic conductance-based neuron models based on a time-scale separation. Based on a stochastic averaging method, the techniques presented here are general, widely applicable and all the computations can be carried out referring to a step-by-step reduction algorithm. We have applied this reduction method to various stochastic HH models. These applications illustrate in detail the computational aspects of the scheme and the type of information that can be gained from it. In particular, we have derived an analytical expression for the diffusion approximation of the reduced 3D version of the stochastic HH model with multistate Markov scheme. More interestingly, for the two-state HH model, unexpected correction terms have appeared in the effective vector field due to the non-linearity m3 , modifying the bifurcations and providing an in-
J Comput Neurosci (2012) 32:327–346
sight into the emergence of bimodal ISI distributions due to channel noise. Thanks to reduction, our work has revealed that existence of a dramatic transition in neuronal dynamics and consequently spike generation properties when channel numbers are decreased. Finally, we have shown that this transition strongly affects neuronal coding in ways that had not been described previously in relation with channel noise. A mathematical analysis of the link between the reduced system and the original one, of the other timescales in the system and of the scaling between the population size and the time-scale separation clearly deserves to be investigated. In biological terms, such an analysis suggests that it would be interesting to study whether there exists a link between ion channels densities and the time-scales on which they operate. Moreover, the stochastic reduction method is expected to be useful both for the stochastic simulation of neuron models with channel noise and for further analysis of channel noise induced phenomena. From a theoretical perspective, there exists several methods in the theory of noisy dynamical systems, (for instance: stochastic bifurcations,large deviations theory) to investigate the impact of a parameter on the behavior of the system. Our reduction approach provides a new method in this direction, and the singular perturbation approach provides a limiting system where the stochastic finite-size effects can be quantified by a deterministic analysis. Such an approach is expected to provide insights into other stochastic multiscale systems, especially deriving from the kinetic formalism, which are ubiquitous in neuron modeling.
Appendix A: Auxiliary functions The transition rates for the HH model are given by: αn (V) = (0.1 − 0.01V)/[exp(1 − 0.1V) − 1], βn (V) = 0.125 exp(−V/80), αm (V) = (2.5 − 0.1V)/[exp(2.5 − 0.1V) − 1], βm (V) = 4 exp(−V/18), αh (V) = 0.07 exp(V/20), βh (V) = 1/[exp(3 − 0.1V) + 1]. Accordingly, τx (V) and x∞ (V), for x = m, n, h, are given by: τx (V) = 1/(αx (V) + βx (V)) and x∞ (V) = αx (V)/(αx (V) + βx (V)).
J Comput Neurosci (2012) 32:327–346
345
Appendix B: Diffusion term in the two-state model
does not depend on j0 and is given by: j
The purpose of this appendix is to give more details about the computation of the diffusion term in the twostate model. The computation is more complicated than in the multi-state model because of the non-linearity m3 . Indeed, in the multi-state case, the vector field F MS is linear with respect to u Na and one only needs to compute the corrective √ diffusion for a single channel, and then divide it by N Na , since the average of N Na independent Brownian motions is √ equal in law to a single Brownian motion divided by N Na . In the nonlinear case (two-state model), one needs to work out the computation directly on the process describing the empirical measure: 1. First, one has to compute the law at time t of the empirical measure, with V fixed, that is the probability PVN ( j, j0 , t) of having j open channels at time t starting from a population with j0 open channels. Denote N = Nm . Starting from a proportion of open m gates u N (0) = Nj0 at t = 0, with 0 ≤ j0 ≤ N, one can show that the proportion of open gates u N (t) at time t follows a bi-binomial distribution:
V ( j) = μ j,N (m∞ ) = C N m∞ (V) j(1 − m∞ (V)) N− j ρN
2. Then, integrating with respect to time the V ( j), we can difference between PVN ( j, i, t) − ρ N express the quantity R(i, j) required to compute the diffusion term. Introducing K = (1 − m∞ (V))/ (m∞ (V)), and making the change of variable z = e−t/τm (V) , the computation boils down to integrals of the form: 1 (1 − z)a (1 + Kz)b (1 + z/K)c dz 0
Using a formal computation software, R(i, j) can be expressed as a sum involving Appell F1 functions, defined by F1 (a, b 1 , b 2 , c; x, y) =
∞ (a)m+n (b 1 )m (b 2 )n m n x y (c)m+n m! n! m,n=0
with (q)n = q (q + 1) · · · (q + n − 1). Then we can write:
min(i, j)
R(i, j) = τm
(N) Cix C N−i Y∞ (x, i, j) j−x
x=max(0, j−N+i)
PVN ( j, j0 , t) "
:= P u N (t) =
j j0 |u N (0) = N N
(N) (x, i, j)F1(N) (x, i, j) − 1 × HK
# with:
(N) x j−x Y∞ (x, i, j) := m∞ m∞ (1 − m∞ )i−x
min( j, j0 )
=
μx, j0 ( p0 (t)) μ j−x,N− j0 ( p1 (t))
× (1 − m∞ ) N−i− j+x
x=max(0, j−N+ j0 )
with μi, j( p) = Cij pi (1 − p) j−i and where p0 (t) and p1 (t) are the solution of y˙ = (1 − y)αm (V) − yβm (V)
(22)
with respective initial conditions p0 (0) = 0 and p1 (0) = 1. Defining τm := αm (V) + βm (V) and m∞ (V) :=
α(V) τ y (V)
(23)
the variation of constant gives: p0 (t) = m∞ (V) 1 − e−t/τm (V) and p1 (t) = e−t/τm (V) + p0 (t). If j0 = 0, then one retrieves the classic binomial distribution with parameters (N, p0 (t)): PVN ( j, 0, t) = j C N p0 (t) j(1 − p0 (t)) N− j. When t → ∞, one checks that the quasi-stationary distribution V ρN ( j) = lim PVN ( j, j0 , t) t→∞
(N) HK (x, i, j) :=
K x−i (1 + K)i+ j−2x 1 + 2x − N − i − j
F1(N) (x, i, j) := F1 (w, x − i, x − j, w + 1, K 1 , 1+K 1+K w := 1 + 2x + N − i − j 3. The complete expression for the variance is then given by plugging the above expression for R(i, j) into formula (20).
References Arnold, V. I. (1983). Geometric methods for ordinary dif ferential equations. New York: Springer. Cecchi, G. A., Sigman, M., Alonso, J. M., Martinez, L., Chialvo, D. R., & Magnasco, M. O. (2000). Noise in neurons is message dependent. Proceedings of the National Academy of Sciences, 97(10), 5557–5561.
346 Chow, C. C., & White, J. A. (1996). Spontaneous action potentials due to channel fluctuations. Biophysical Journal, 71(6), 3013–3021. Faggionato, A., Gabrielli, D., & Crivellari M. R. (2010). Averaging and large deviation principles for fully-coupled piecewise deterministic Markov processes and applications to molecular motors. Markov Processes and Related Fields, 16(3), 497–548. Gillespie, D. T. (1977). Exact stochastic simulation of coupled chemical reactions. The Journal of Physical Chemistry, 81(25), 2340–2361. Hille, B. (2001). Ion channels of excitable membranes. Massachusettes: Sinauer Sunderland. Hodgkin, A. L., & Huxley, A. F. (1952). A quantitative description of membrane current and its application to conduction and excitation in nerve. Journal of Physiology, 117(4), 500– 544. Holden, A. V., Muhamad, M. A., & Schierwagen, A. K. (1985). Repolarizing currents and periodic activity in nerve membrane. Journal of Theoretical Neurobiology, 4, 61–71. Izhikevich, E. M. (2007). Dynamical systems in neuroscience: The geometry of excitability and bursting. Cambridge: MIT Press. Jung, P., & Shuai, J. W. (2001). Optimal sizes of ion channel clusters. Europhysics Letters, 56, 29–35. Keener, J. (2009). Invariant manifold reductions for Markovian ion channel dynamics. Journal of Mathematical Biology, 58(3), 447–57. Kepler, T. B., Abbott, L. F., & Marder, E. (1992). Reduction of conductance-based neuron models. Biological Cybernetics, 66, 381–387. Mastny, E. A., Haseltine, E. L., & Rawlings, J. B. (2007). Two classes of quasi-steady-state model reductions for stochastic kinetics. The Journal of Chemical Physics, 127, 094106. Pakdaman, K., Tanabe, S., & Shimokawa, T. (2001). Coherence resonance and discharge time reliability in neurons and neuronal models. Neural Networks, 14(6–7), 895–90. Pakdaman, K., Thieullen, M., & Wainrib, G. (2010). Fluid limit theorems for stochastic hybrid systems with application to neuron models. Advances in Applied Probability, 42(3), 761–794. Rinzel, J. (1985). Excitation dynamics: Insights from simplified membrane models. Feredation Proceedings, 44, 2944– 2946. Rowat, P. (2007). Interspike interval statistics in the stochastic Hodgkin–Huxley model: Coexistence of gamma frequency bursts and highly irregular firing. Neural Computation, 19(5), 1215. Rubin, J., & Wechselberger, M. (2007). Giant squid—hidden canard: The 3d geometry of the Hodgkin Huxley model. Biological Cybernetics, 97, 5–32.
J Comput Neurosci (2012) 32:327–346 Schmid, G., Goychuk, I., & Hänggi, P. (2001). Stochastic resonance as a collective property of ion channel assemblies. Europhysics Letters, 56, 22–28. Schmid, G., Goychuk, I., & Hänggi, P. (2003). Channel noise and synchronization in excitable membranes. Physica A: Statistical Mechanics and its Applications, 325(1–2), 165–175. Schneidmann, E., Freedman, B., & Segev, I. (1998). Ion channel stochasticity may be critical in determining the reliability and precision of spike timing. Neural Computation, 10, 1679– 1703. Shuai, J. W., & Jung, P. (2003). Optimal ion channel clustering for intracellular calcium signaling. Proceedings of the National Academy of Sciences, 100(2), 506–512. Shuai, J. W., & Jung, P. (2005). Entropically enhanced excitability in small systems. Physical Review Letters, 95(11), 114501. Skaugen, E., & Walloe, L. (1979). Firing behavior in a stochastic nerve membrane model based upon the Hodgkin– Huxley equations. Acta Physiologica Scandinavica, 107(4), 343–63. Steinmetz, P. N., Manwani, A., Koch, C., London, M., & Segev, I. (2000). Subthreshold voltage noise due to channel fluctuations in active neuronal membranes. Journal of Computational Neuroscience, 9(16), 133–148. Suckley, R., & Biktashev, V. N. (2003). Comparison of asymptotics of heart and nerve excitability. Physical Review E, 68, 011902, 1–15. Takahata, T., Tanabe, S., & Pakdaman, K. (2002). White noise stimulation of the Hodgkin–Huxley model. Biological Cybernetics, 86, 403–417. Tanabe, S., & Pakdaman, K. (2001). Noise-induced transition in excitable neuron models. Biological Cybernetics 85, 269–280. Tanabe, S., & Pakdaman, K. (2001). Noise-enhanced neuronal reliability. Physical Review E, 64, 041904. Tanabe, S., & Pakdaman, K. (2001). Dynamics of moments of Fitz–Hugh–Nagumo neuronal models and stochastic bifurcations. Physical Review E, 63, 031911. Tanabe, S., Sato, S., & Pakdaman, K. (1999). Response of an ensemble of noisy neuron models to a single input. Physical Review E, 60, 7235–7238. Tateno, T., & Pakdaman, K. (2004). Random dynamics of the Morris–Lecar neural model. Chaos, 14, 511–530. Wainrib, G., Thieullen, M., & Pakdaman, K. (2010). Intrinsic variability of latency to first spike. Biological Cybernetics, 103(1), 43–56. White, J. A., Rubinstein, J. T., & Kay, A. R. (2000). Channel noise in neurons. Trends in Neurosciences, 23(3), 131–137. Yin, G. G., & Zhang, Q. (1998). Continuous-time Markov chains and applications: A singular perturbation approach. New York: Springer.