Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Blog
    • Collections
    • Podcast
  • TOPICS
    • Cognition and Behavior
    • Development
    • Disorders of the Nervous System
    • History, Teaching and Public Awareness
    • Integrative Systems
    • Neuronal Excitability
    • Novel Tools and Methods
    • Sensory and Motor Systems
  • ALERTS
  • FOR AUTHORS
  • ABOUT
    • Overview
    • Editorial Board
    • For the Media
    • Privacy Policy
    • Contact Us
    • Feedback
  • SUBMIT

User menu

Search

  • Advanced search
eNeuro

eNeuro

Advanced Search

 

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Blog
    • Collections
    • Podcast
  • TOPICS
    • Cognition and Behavior
    • Development
    • Disorders of the Nervous System
    • History, Teaching and Public Awareness
    • Integrative Systems
    • Neuronal Excitability
    • Novel Tools and Methods
    • Sensory and Motor Systems
  • ALERTS
  • FOR AUTHORS
  • ABOUT
    • Overview
    • Editorial Board
    • For the Media
    • Privacy Policy
    • Contact Us
    • Feedback
  • SUBMIT
PreviousNext
Research ArticleMethods/New Tools, Novel Tools and Methods

An Efficient Population Density Method for Modeling Neural Networks with Synaptic Dynamics Manifesting Finite Relaxation Time and Short-Term Plasticity

Chih-Hsu Huang and Chou-Ching K. Lin
eNeuro 6 December 2018, 5 (6) ENEURO.0002-18.2018; DOI: https://doi.org/10.1523/ENEURO.0002-18.2018
Chih-Hsu Huang
1Department of Neurology, National Cheng Kung University Hospital, College of Medicine, National Cheng Kung University, Tainan, Taiwan 70403
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Chou-Ching K. Lin
1Department of Neurology, National Cheng Kung University Hospital, College of Medicine, National Cheng Kung University, Tainan, Taiwan 70403
2Innovation Centre of Medical Devices and Technology, National Cheng Kung University Hospital, College of Medicine, National Cheng Kung University, Tainan, Taiwan 70403
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Chou-Ching K. Lin
  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF
Loading

Abstract

When incorporating more realistic synaptic dynamics, the computational efficiency of population density methods (PDMs) declines sharply due to the increase in the dimension of master equations. To avoid such a decline, we develop an efficient PDM, termed colored-synapse PDM (csPDM), in which the dimension of the master equations does not depend on the number of synapse-associated state variables in the underlying network model. Our goal is to allow the PDM to incorporate realistic synaptic dynamics that possesses not only finite relaxation time but also short-term plasticity (STP). The model equations of csPDM are derived based on the diffusion approximation on synaptic dynamics and probability density function methods for Langevin equations with colored noise. Numerical examples, given by simulations of the population dynamics of uncoupled exponential integrate-and-fire (EIF) neurons, show good agreement between the results of csPDM and Monte Carlo simulations (MCSs). Compared to the original full-dimensional PDM (fdPDM), the csPDM reveals more excellent computational efficiency because of the lower dimension of the master equations. In addition, it permits network dynamics to possess the short-term plastic characteristics inherited from plastic synapses. The novel csPDM has potential applicability to any spiking neuron models because of no assumptions on neuronal dynamics, and, more importantly, this is the first report of PDM to successfully encompass short-term facilitation/depression properties.

  • colored noise
  • diffusion approximation
  • exponential integrate-and-fire model
  • Langevin equations

Significance Statement

Our study successfully solve an outstanding problem, how to reduce the dimension of population density equations when realistic synaptic dynamics is incorporated. With the newly proposed Fokker–Planck formalism, population density method (PDM) is conferred short-term plasticity (STP) properties and therefore becomes more widely applicable. As such, our method offers an opportunity to use the PDM to gain new insights into neural mechanisms of brain functions that are strongly dependent on STP synapses. This is the first step toward macroscopic description of large-scale neural network activities, reflected in some commonly used neurophysiological measurements, e.g., EEG, MEG, fMRI, and voltage-sensitive dye imaging data.

Introduction

The population density method (PDM) appears as a time-saving alternative to direct numerical simulations [i.e., Monte Carlo simulation (MCS)] of large-scale neural networks of spiking neurons (Knight et al., 1996; Brunel and Sergi, 1998; Brunel, 2000; Nykamp and Tranchina, 2000; Omurtag et al., 2000; Casti et al., 2002; Huertas and Smith, 2006; Augustin et al., 2013). In this method, biophysically similar neurons are grouped into a population, and a probability density function for each population is defined to describe the variability of state variables of such neurons. The evolution of this density function is governed by a master equation (generally a partial differential equation such as a Fokker–Planck equation). A prominent advantage of PDM is that if the master equation can be easily solved in analytical or numerical ways, precise estimates of the quantities of interest can be obtained much faster than MCS.

Because of the significant impact of synaptic dynamics on network dynamics, PDMs must ultimately include accurate descriptions of synaptic kinetics to be widely applicable. Unfortunately, incorporating more synaptic dynamics increases the dimension of density functions, leading to a catastrophic decline in computational efficiency because of the increasing difficulty of solving high-dimensional master equations. How to incorporate realistic synaptic dynamics in a computationally efficient manner therefore becomes an outstanding problem in the field of PDMs (Apfaltrer et al., 2006; Tranchina, 2009).

Although many previous studies addressed this issue (Haskell et al., 2001; Nykamp and Tranchina, 2001; Apfaltrer et al., 2006; Rangan and Cai, 2006; Ly and Tranchina, 2007; Rangan et al., 2008), they left out some unsolved problems (section 7.10 in Tranchina, 2009). Recently, Ly (2013) developed a principled dimension-reduction method, called modified mean-field method (MMFM), that could deal with those unsolved problems. Nevertheless, we found that the MMFM might not be a robust method and not suitable for nonlinear spiking neuron models. The previous studies concentrated, to the best of our knowledge, only on synaptic dynamics with finite decay and neglected another important synaptic attribute, the short-term plasticity (STP; Markram and Tsodyks, 1996; Markram et al., 1998; Zucker and Regehr, 2002) that also played an important role in certain specific network dynamics, such as reverberatory bursting activity in in vitro and in vivo neuronal networks (Volman et al., 2007; Gritsun et al., 2011), self-organized criticality (Levina et al., 2007), anticipative neural responses (Fung et al., 2012), and sustained population activity related to working memory (Barak and Tsodyks, 2007; Mongillo et al., 2008). Therefore, our goal in the present study is to develop a new solution that allows the PDM to include realistic synaptic dynamics of both finite decay and STP in a computationally efficient manner.

In this study, we present a novel and efficient PDM, called colored-synapse PDM (csPDM), to achieve our goal. The strategy behind the csPDM is also to take advantage of dimension-reduction methods, relying on the following scheme: (1) synaptic dynamics are assumed to operate in the diffusive limit so that the dynamic equations of neuronal dynamics are converted into Langevin equations with colored driving noise and all synapse-associated state variables are transformed into input terms (Moreno-Bote and Parga, 2004; Richardson and Gerstner, 2005; Destexhe and Rudolph-Lilith, 2012), and (2) the probability density function method for this dynamic system is then employed (Wang et al., 2013; Barajas-Solano and Tartakovsky, 2016). As such, the synapse-associated state variables are not viewed as the system state variables of the resulting Langevin equations, leading to the reduction of the dimension of density functions. As a consequence, incorporating realistic synaptic dynamics does not decrease the computational efficiency when adopting csPDM, because it does not increase the dimension of master equations.

We compare the simulated network activities from the csPDM with those from MCSs, original full-dimensional PDMs (fdPDMs), and MMFMs. The numerical results show that the results of the csPDM reveal good agreement with those of the MCS and fdPDM in both the steady-state and dynamic regimes. The csPDM provides more accurate simulations than the MMFM and remains more computationally efficient than the fdPDM. We hope that the csPDM will be valuable to computational neuroscientists who seek accurate and efficient network modeling tools to explore effects of synaptic dynamics with/without STP properties on macroscopic behaviors of neural networks.

Materials and Methods

Here, we describe the equations and parameters used for neuronal, synaptic and network modeling. We present the model (master) equations of the fdPDM, csPDM, and MMFM. The corresponding closed-form solutions for evaluating firing rate responses (network activities) are also described. At the end of this section, we present the numerical method used to solve the model equations of the csPDM as well as the methods used to quantify simulation errors. A summary of the notation for all the main dynamical variables and physiological parameters is given in Table 1.

Network models

The large-scale network considered here is a neuronal population of N uncoupled exponential integrate-and-fire (EIF) neurons (Fourcaud-Trocmé et al., 2003) receiving only synaptic inputs from the outside. We choose the EIF model mainly because it can offer better prediction of spiking times for a given current input than the leaky integrate-and-fire models and quadratic ones (Fourcaud-Trocmé et al., 2003). In addition, its adaption version, i.e., the adaptive EIF model (Brette and Gerstner, 2005), is capable of reproducing various firing patterns observed in in vivo recordings of cortical cells (Naud et al., 2008; Touboul and Brette, 2008). In fact, the csPDM derived in the present study can be directly applied to simulate neural networks of adaptive EIF neurons.

In this network, the external synaptic inputs to each neuron are assumed to be mediated by m different types of neurotransmitter receptors. For the sth type of receptors, there are presumably cs synaptic connections. In other words, each neuron has totally Embedded Image external synaptic connections. The synaptic input on each connection is a set of unitary spiking events whose random arrival times are assumed to be governed by an inhomogeneous Poisson process. νs(t) denotes the mean rate of the Poisson process for the input through the sth-type receptor. to satisfy basic premises of the PDM, two arbitrary synaptic inputs are assumed to be statistically independent if they are mediated by different types of receptors and to be statistically identical if mediated by the same type.

Single neuron models

For each neuron, the dynamics of its membrane voltage, V (t), is described by Embedded Image (1)where Embedded Image is the total synaptic current generated by the external synaptic inputs, C is the membrane capacitance, gl is the leak conductance, El is the leak reversal potential, κ is the sharpness factor, and VT is the threshold voltage. The spiking mechanism is the following: a spike is triggered at ts when V (ts) reaches a cutting voltage, Vc (in general, Embedded Image ). Afterward, V is immediately reset to a resetting voltage, Vr, and, then, clamped for a refractory period, τref.

Synapse models

Specifically, Embedded Image is the sum of all postsynaptic currents mediated by various types of neurotransmitter receptors:Embedded Image (2) where gs(t) is the (collective) synaptic conductance for the input mediated by the sth type of receptors, and Es is the corresponding synaptic reversal potential. Equation 2 refers to conductance-based synaptic models.

The synaptic dynamics means the dynamics of the gs(t). In our treatment, the synaptic conductance follows first order kinetics, where the rise in conductance at the arrival of a unitary synaptic event is instantaneous and the decline is exponential. Therefore, the dynamic equation for the synaptic conductance has the following form:Embedded Image (3) where τs is the decay or synaptic time constant and Γs is the total variation of the synaptic conductance induced by a single event. The size of the instantaneous jump in conductance equals to Embedded Image . Embedded Image is the arrival time of the jth synaptic event, and Embedded Image refers to the spike train consisting of synaptic events summed from all synaptic connections through the sth-type receptor.

In our numerical examples, the value of Γs is set according to the following equation:Embedded Image (4) where ΔVs is a free parameter defined as the instant voltage modulation provoked by a single synaptic event when the membrane voltage is El just before the event arrives in cases where κ (in Eq. 1) is in the limit of Embedded Image and the synaptic dynamics has a zero-order kinetics, i.e., Embedded Image .

Short-term plastic synapses

What stated above does not consider STP property on synaptic connections. In the following, we concentrate on how to add it into the synaptic dynamics. A typical dynamic model of STP was the one proposed by Barak and Tsodyks (2007). They introduced two running parameters to describe the change of the connection weight of a synaptic coupling in a phenomenological way. Following the same idea, the Γs in Equation 3 is replaced byEmbedded Image (5) where the running parameters, us and xs, are governed byEmbedded Image (6) Embedded Image (7) at the level of a single synapse. The jump size of the synaptic conductance caused by a spike now depends on us and xs. us is the running utilization parameter and xs is the running fraction of available neurotransmitters. The dynamic value ranges from Us to 1 for us and from 0 to 1 for xs. Us refers to the base level of us. us = 1 means a presynaptic spike is allowed to use all available neurotransmitters, and xs = 1 means all neurotransmitters are available.

Briefly, the Equations 6, 7 depict a phenomenon, where when a synaptic spike arrives, us increases by an amount of Embedded Image , and, afterward, decays to its baseline level, Us, with the facilitation time constant, τs,f, and, meanwhile, xs decreases by an amount of xsus and recovers to its baseline value of 1 with the recovery time constant, τs,r, afterward. In general, τs,f and τs,r are short, ranging from hundreds of milliseconds to seconds (Markram and Tsodyks, 1996; Markram et al., 1998). Besides τs, Es, and Γs, each synaptic connection with the STP property is also characterized by three additional parameters: Us (initial utilization parameter) and τs,f as well as τs,r (facilitation and recovery time constants, respectively), which control the type of the STP property from strong depression (Embedded Image and relatively large values of Us) to strong facilitation (Embedded Image and small values of Us; Barak and Tsodyks, 2007).

Note that we keep the subscript s in the Equations 5–7 above, because it is likely that the synaptic connections through different types of transmitter receptors have different STP types.

MCS

We are interested in the network activity, i.e., the output population firing rate, of such a neuronal population in response to a given set of Embedded Image (Embedded Image ) as a whole. This quantity can be computed by the direct use of the MCS, where states of all neurons and synapses are explicitly traced according to Equations 1–3 or Equations 1–7 if STP properties are involved. For all our simulation examples, MSC is performed with the brian2 simulator (Stimberg et al., 2014). In all simulations, the network composes of N = 10,000 EIF neurons, and cs = 200 for all receptor types.

In the MCS, the output population firing rate is calculated byEmbedded Image (8) where Embedded Image is the total number of spikes produced by all neurons in the population within the time period of Embedded Image , and Embedded Image is a small time window. Δ t is always set to 1 ms in our simulation examples.

fdPDM

The assumptions and conditions, under which the fdPDM provides an exact description of network activity, have been previously discussed (Nykamp and Tranchina, 2000; Omurtag et al., 2000).

Specifically, the fdPDM is a density-based modeling approach to directly estimate r(t) with a known distribution of states of neurons, which is described by a joint probability density function, Embedded Image , defined byEmbedded Image (9) where Embedded Image is the conductance state vector composed of m synaptic conductance variables. Ω is a subdomain within the state space bounded by the definition domains of all state variables. In general, the definition domain of Embedded Image ranges from 0 to ∞. About the membrane voltage V, its upper bound is Vc, and its lower bound is given by Embedded Image . Equation 9 states that the integration of the density function over Ω is the probability of finding neurons whose states are within that subdomain in a large neuronal population. Note that, here, we do not consider STP properties on synaptic connections to illustrate our main ideas in the present study in a simple way. We will consider these properties only for the csPDM later.

Unlike in the MCS, where one has to track every neuron individually, in the fdPDM, one just track one density function for each population. As explained in details in previous papers (Nykamp and Tranchina, 2000; Haskell et al., 2001; Apfaltrer et al., 2006; Ly, 2013), the master equation governing the evolution of Embedded Image forms a (1+ m)-dimensional (in space) partial-integral differential equation:Embedded Image (10) where the voltage probability flux, i.e., the probability flux along the voltage-direction, JV, is given by Embedded Image Embedded Image (11) and Embedded Image is the conductance probability flux vector composed of m conductance probability fluxes. The conductance probability flux, Embedded Image , is given byEmbedded Image (12)

The population firing rate, r(t), is obtained by integrating the voltage probability flux across Vc over conductance, yieldingEmbedded Image (13)

Embedded Image denotes the subdomain where Embedded Image , meaning that neurons at Vc intend to cross the boundary from below to generate spikes in this domain. As seen, for a given set of inputs, νs(t) (Embedded Image ), we need to solve the Equation 10 to obtain the output response of the neuronal population, i.e., the output population firing rate, r(t). For solving that equation, the boundary conditions are:Embedded Image Embedded Image Embedded Image Embedded Image where Vlb denotes the lower bound of the membrane voltage. The former boundary condition results from the resetting mechanism of spiking, meaning that the voltage probability flux across Vc, accounting for the generation of action potentials, re-enters the state space on Vr after the refractory period. The latter ones mean no probability fluxes flow outward the domain through other boundaries. As stated, the boundary conditions result in the conservation of the number of neurons. That is to say, the integration of the density function over the domain always equals to one in the absence of refractory period.

The computational time required for solving the master equation depends on the number of its dimensions. Generally, solving a high-dimensional master equation is highly time consuming. Previous studies (Apfaltrer et al., 2006; Tranchina, 2009) have suggested that the fdPDM cannot be considered as a time-saving alternative to the MCS if the dimension of the master equation exceeds three. In the following, two other methods used to tackle this issue are described. MMFM is a documented method, and csPDM is a new method proposed in this paper.

MMFM

MMFM is proposed by Ly (2013) for speeding up the estimation of the population firing rate. In this method, Embedded Image in the Equation 10 is viewed as a deterministic parameter so that the master equation becomes a one-dimensional partial differential equation:Embedded Image (14)

In this case, a steady-state firing rate for a given Embedded Image is obtained by using the mean-field model, given byEmbedded Image (15)

Recall that the synaptic inputs for different types of neurotransmitter receptors are assumed to be independent mutually; therefore, synaptic conductances for different receptor types are also independent with each other. Furthermore, they only depend on their individual inputs. So, it is realized that the value of gs evolves isolatedly. For a large neuronal population, one can define a marginal conductance density function, Embedded Image , for gs by analogy with Embedded Image . Also, the integration of Embedded Image over a certain space on gs is the probability of finding neurons whose values of gs are located in that space. According to what derived by Ly (2013), the master equation governing the time evolution of Embedded Image is yielded asEmbedded Image (16) whereEmbedded Image (17)

The boundary conditions for solving Equation 16 are Embedded Image . Assuming that the values of all elements of Embedded Image are drawn from the marginal conductance density functions individually, MMFM estimates the time-varying firing rate, Embedded Image , with expected value of Embedded Image conditioned on Embedded Image , that is,Embedded Image (18)

As such, instead of solving a (1 + m)-dimensional partial-integral differential equation in the fdPDM, one only needs to solve m one-dimensional partial differential equations simultaneously for estimating r(t) in the MMFM. Ly suggested that MMFM had higher computational efficiency than fdPDM, especially when the values of Embedded Image at all possible points of Embedded Image could be computed before simulations.

Colored-synapse population density method (csPDM)

This section presents the derivation of the csPDM. csPDM is inspired by the probability density method for Langevin equations with colored noise (Wang et al., 2013; Barajas-Solano and Tartakovsky, 2016). So, we start by presenting how to transform the Equation 1 into a stochastic Langevin equation through the diffusion approximation of gs.

In the diffusion limit, under the condition that the mean rate of synaptic input received by each neuron, Embedded Image , is sufficiently high and the jump size Embedded Image for each synaptic spike is small enough, the random synaptic conductance Embedded Image can be treated as the Ornstein–Uhlenbeck process (Uhlenbeck and Ornstein, 1930; Risken, 1996), whose dynamics is given by (Richardson, 2004)Embedded Image (19) where Embedded Image is the mean value of Embedded Image and Embedded Image is the standard deviation of Embedded Image . Embedded Image is a δ-correlated white-noise process of unit variance. Introducing a new variable Embedded Image and substituting for gs in the equation above, we obtain a new Ornstein–Uhlenbeck process for Embedded Image :Embedded Image (20)

Embedded Image has zero mean and unit variance (Risken, 1996); and its autocorrelation function isEmbedded Image (21)

As a result, Embedded Image is a colored noise due to the exponential form of Embedded Image and finite synaptic time constant, i.e., Embedded Image . Replacing Embedded Image in the Equation 1 with Embedded Image , we obtainEmbedded Image (22)as the new dynamic equation for the membrane voltage, where Embedded Image , Embedded Image , and Embedded Image . So, as shown in Equation 22, the fluctuation of Embedded Image now is characterized by a Langevin equation with m independent colored noises Embedded Image resulting from synaptic dynamics with finite synaptic time constants (Destexhe and Rudolph-Lilith, 2012). That is why we use “colored-synapse” as the prefix of this method.

Barajas-Solano and Tartakovsky (2016) proposed a closed-form quasi-Fokker–Planck equation as the master equation for Langevin equations driven by colored noise. They considered a dynamic system characterized by the stochastic Langevin equation in n dimensionsEmbedded Image (23) for Embedded Image , where Embedded Image . Each Embedded Image is decomposed into a deterministic function or “mean-field velocity” Embedded Image and a stochastic fluctuation term Embedded Image . The authors proposed a quasi-Fokker–Planck equation:Embedded Image (24) as the master equation governing the temporal evolution of the joint probability density function of the system states, Embedded Image . The stochastic diffusion tensor, D, is calculated byEmbedded Image (25) if fluctuation velocity components are exponentially autocorrelated and mutually uncorrelated, i.e.,Embedded Image Embedded Image

Embedded Image is the Jacobian of the mean-field velocity with components Embedded Image . Considering V, Embedded Image , and Embedded Image in Equation 22 as x, Embedded Image , and Embedded Image in Equation 23, respectively, we can yield the master equation for the marginal voltage density function, Embedded Image , by using Equation 24, as the following:Embedded Image (26) whereEmbedded Image (27)

By using Equation 25, we obtainEmbedded Image Embedded Image (28) Embedded Image Embedded Image (29) whereEmbedded Image (30) and Embedded Image denotes the average membrane voltage across the population.

What remains in Equation 26 are equations for evaluating Embedded Image and Embedded Image . To derive the equation for evaluating Embedded Image , we multiply Equation 16 by gs and integrate gs from 0 to ∞:Embedded Image Embedded Image Embedded Image Embedded Image Embedded Image Embedded Image (31)

Similarly, one can derive the equation for evaluating Embedded Image , yieldingEmbedded Image (32)

For Embedded Image , we use the identity, Embedded Image .

In the csPDM, the population firing rate is directly estimated by the probability flux across Vc:Embedded Image (33)

The average membrane voltage, Embedded Image , is computed by using its definition:Embedded Image (34)

The boundary conditions for solving Equation 26 are assigned as:Embedded Image Embedded Image Embedded Image

They are similar to what used in the fdPDM except for the last boundary condition, which means that no neuron locates at Vc because the neurons whose voltages reach Vc are reset to Vr immediately.

When the STP property is included, we replace Γs and Embedded Image withEmbedded Image (35) Embedded Image (36) in Equation 31, Equation 32 for evaluating Embedded Image and Embedded Image , where Embedded Image and Embedded Image are mean values of us and xs across the neuronal population, respectively. Following mean-field equations proposed by Barak and Tsodyks (2007), we use the same equations for tracking Embedded Image and Embedded Image (i.e., Eq. 1 in Barak and Tsodyks, 2007):Embedded Image (37) Embedded Image (38)

To sum up, unlike in the fdPDM, the network dynamics in the csPDM is described by a system consisting of a one-dimensional quasi-Fokker–Planck equation for tracking the marginal density, Embedded Image , and 2m ordinary differential equations for tracking the first two statistical moments of all synapse-associated variables, Embedded Image , or 4m equations for statistical moments of Embedded Image if STP is included. csPDM is expected to be more computationally efficient than fdPDM because solving one-dimensional quasi-Fokker–Planck equation undoubtedly takes less time than high-dimensional partial differential equations. Basically, they are solved numerically. In the next section, we present the numerical method used for solving the quasi-Fokker–Planck equation.

Numerical method for solving quasi-Fokker–Planck equation

To solve Equation 26, we use local Galerkin method (LGM; Cockburn and Shu, 1998; Xu and Shu, 2010). This method belongs to the discontinuous Galerkin methods and focuses on the solutions of partial differential equations with high order derivatives. We select it as the proposed numerical method because it has superior ability to handle discontinuous solutions (Huang et al., 2015) and parallelizability in computations (Biswas et al., 1994). First, in the LGM, an auxiliary variable, Embedded Image , is introduced to rewrite the Equation 26 as followings:Embedded Image (39) Embedded Image (40)

As a consequence, the Equation 26 is transformed to a conservative hyperbolic equation.

Discretization of space domain and basic notations

For a given bounded V-domain Embedded Image , we divide it into M meshes with an identical length as follows:Embedded Image (41) so that Vr and Es are exactly certain grid points. We denote the subspace Embedded Image (Embedded Image ) and its length Embedded Image . Using first-order polynomials as shape functions, the approximated values of Embedded Image and Embedded Image within Ik are defined by:Embedded Image (42) Embedded Image (43) Embedded Image in which ϕ1 and ϕ2 are shape functions. Embedded Image and Embedded Image refers to the value of Embedded Image at Embedded Image from the left mesh Ik and from right mesh Ik+1, respectively. Due to the discontinuity at the interface of adjacent meshes, Embedded Image is thus possible. Embedded Image and Embedded Image are defined in the same way.

Element equations and numerical fluxes

Substituting Embedded Image and Embedded Image with approximated values Embedded Image and Embedded Image , respectively, multiplying shape functions and integrating with respect to v over the mesh Ik, we then obtain element equations for cell Ik, given byEmbedded Image (44) Embedded Image (45) in which the notation “^” means numerical fluxes at the interfaces between cells. We set Embedded Image and Embedded Image , and employ upwind fluxes for Embedded Image , yieldingEmbedded Image (46)

According to the boundary conditions, we set Embedded Image , corresponding to Embedded Image , Embedded Image , respectively, and enforcedly assign Embedded Image as Embedded Image if Embedded Image . The firing rate r(t) is given by Embedded Image . We use the backward Euler method (Cheng and Shu, 2007) to solve element equations as well as Equations 31–38 to ensure numerical stability. Note that one needs to re-calculate element equations at each time step because Embedded Image and D are functions of time-dependent parameters, Embedded Image and Embedded Image , respectively.

Slope limiters

A slope limiter is employed to guarantee the positivity of the density function (Huang et al., 2015). After progressing one time step with the backward Euler method, the Embedded Image across all meshes go through a slope limiter Embedded Image , which is defined byEmbedded Image (47) where Embedded Image is set as Embedded Image . The ψ is the “minmod function” defined as:Embedded Image (48)

Generally, Equations 10, 16 are also solved numerically. We adopt the so-called discontinuous Galerkin method to solve them. The details of this method are not stated here. Please refer to Huang’s paper (Huang et al., 2015) for details about the numerical method.

Quantification of simulation errors

MCSs are considered as the ground-truth of all simulation examples in this study; thereby, the differences of the fdPDM, MMFM, and csPDM in simulation results with the MSC are used to explore their performances. The following two quantities are used to quantify the simulation errors of the csPDM on the estimations of marginal conductance and voltage density functions, given byEmbedded Image (49) Embedded Image (50) to check the validity of diffusion approximation on synaptic dynamics. Embedded Image describes the average error ratio on the marginal conductance density function, which is estimated only over the interval Embedded Image . It is noted that the marginal conductance density function in the csPDM is characterized by a Gaussian distribution, here denoted by Embedded Image , whose mean and standard deviation are Embedded Image and Embedded Image . Embedded Image is the average error ratio on the marginal voltage density function.

The third quantity isEmbedded Image (51) where Embedded Image . It is used for comparing performances of the fdPDM, MMFM, and csPDM. Embedded Image means the average error ratio on the population firing rate obtained from fdPDM, MMFM, or csPDM. Embedded Image is calculated from Equation 8. Embedded Image is calculated from Equation 13 for fdPDM, Equation 18 for MMFM, and Equation 33 for csPDM.

Results

We will present four simulation examples: (1) steady-state analyses, (2) dynamic population responses to excitatory input only, (3) dynamic population responses to multiple inputs, and (4) STP population responses. Using the former two examples, we aim to compare the performances of fdPDM, MMFM, and csPDM and focus on the accuracy and efficiency of csPDM. We show the applicability of csPDM to more complicated neuronal activities by the latter two examples. Thus, in both the examples, only the simulation results of MCS and cdPDM are presented. The parameters required for simulations are listed inTable 2. The values of neuronal properties are assigned according to previous studies (Fourcaud-Trocmé et al., 2003; Destexhe, 2009; Hertäg et al., 2014), and all of them are within in the physiologic ranges. The parameters relevant to synaptic dynamics and STP are regarded as free parameters which are freely tuned to let τs correspond to the time scales of common transmitter receptors or to make synaptic interactions exhibit short-term facilitation/depression if necessary. The input rate, νs, shown in this section is normalized by the minimum constant input rate that drives EIF neurons to produce spikes under fluctuation-free conditions (i.e., mean-field models), except for the third and fourth simulation examples. Such a minimum input rate is Embedded Image . Next, we start to show simulation results by first checking the validity condition of diffusion approximation of synaptic dynamics.

View this table:
  • View inline
  • View popup
Table 1

List of notation and symbols

View this table:
  • View inline
  • View popup
Table 2

List of values of model parameters

Validation of diffusion approximation of synaptic dynamics

For the correct use of the csPDM, to explore the validity condition of diffusion approximation of synaptic dynamics is necessary. To do so, we check steady-state values of Embedded Image and Embedded Image of a single neuronal population receiving only constant excitatory inputs with respect to different νs and Embedded Image . The synaptic time constant τs is set as 5 or 100 ms to match the time scale of AMPA-receptors or NMDA-receptors (Dayan and Abbott, 2001), and Es is set as 0 mV for both cases. Basically, small Embedded Image reflects the validity of diffusion approximation. As shown in Figure 1A, larger νs or smaller Embedded Image results in smaller Embedded Image for both τs = 5 and τs = 100 ms, in agreement with the fact that the diffusion approximation is valid when the input rate is large and the jump size is small. As seen, increasing τs produces small Embedded Image because of the decreased variation of gs (Embedded Image at steady states). Three distributions of gs under different sets of (νs, Embedded Image ) are shown in Figure 2A–C. As shown, the distribution of gs, Embedded Image , approaches a Gaussian distribution when Embedded Image is <0.2 (Fig. 2C). That is, the diffusion approximation of synaptic dynamics is valid if Embedded Image . Figure 1B shows Embedded Image . The amplitude of Embedded Image is proportional to that of Embedded Image , meaning that the error of csPDM in Embedded Image in large part comes from the failure of the diffusion approximation. But, surprisingly, the value of Embedded Image is much lower than that of Embedded Image under the same set of (νs, ΔVs). As seen in Figure 2A,D, under the case where Embedded Image mV, and τs = 5 ms, Embedded Image is only 0.174 whereas Embedded Image is as high as 0.653. However, the reason why Embedded Image is smaller than Embedded Image is unclear. Through the observation of Embedded Image (Fig. 2D–F), it is found that csPDM gives an adequately accurate estimation of Embedded Image comparable to MCS when Embedded Image is <0.2. To offer a quantitative and universal validity condition for diffusion approximation of synaptic dynamics, we also check the value of Embedded Image (i.e., the coefficient of variation of gs; Fig. 1C). In fact, it has been argued that the diffusion approximation is valid only when Embedded Image , meaning that the conductance mean Embedded Image should be much larger than the standard deviation Embedded Image (Richardson and Gerstner, 2005; Cai et al., 2012). It is however an impractical condition. Based on our observations of Embedded Image and Embedded Image , it can be conclusively said that the validity condition of diffusion approximation for the application of csPDM is Embedded Image , under which Embedded Image and Embedded Image are approximately <0.2. As a result, ΔVs should be <1.5 mV for the case of τs = 5 ms such that this criterion can be satisfied in the range of Embedded Image . In the following simulations, we choose ΔVs as 1 mV for the excitatory inputs.

Figure 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 1.

Steady-state values of Embedded Image , Embedded Image and Embedded Image of a single population with respect to different νs and Embedded Image A, Embedded Image B, Embedded Image C, Embedded Image . Black dashed lines indicate the sets of (νs, ΔVs) where Embedded Image . Blank areas in A, B indicate the error ratio of >0.2.

Figure 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 2.

Examples of steady-state marginal conductance density functions Embedded Image and marginal voltage density functions Embedded Image . A, D, Embedded Image and Embedded Image under Embedded Image and Embedded Image mV. B, E, Embedded Image and Embedded Image under Embedded Image and Embedded Image mV. C, F, Embedded Image and Embedded Image under Embedded Image and Embedded Image mV. As shown, the estimations of Embedded Image and Embedded Image by csPDM are similar with those by MCS if the input rate νs is increased. τs is set as 5 ms in all examples. Shaded area: MCS. Black thick line: csPDM.

Steady-state analyses

The performances of fdPDM, MMFM, and csPDM are first examined via steady-state analyses of population firing rates in response to fixed excitatory inputs. Here, we also consider two cases, τs = 5 and τs = 100 ms, and set ΔVs = 1 mV and Es = 0 mV. Figure 3 displays input-output curves computed by MSC, fdPDM, csPDM, and MMFM. Results show that csPDM gives an estimation of input-output curves that are close to those from MSC and fdPDM with errors <1 Hz for both τs = 5 and τs = 100 ms. These results indicate that csPDM can accurately estimate steady-state output population firing rates although it just tracks the marginal voltage density function. In contrast, MMFM gives accurate estimations of steady-state output firing rates only under the condition of τe = 100 ms. It overestimates the actual firing rate in the fluctuation-driven regime (i.e., νs < 1) and underestimates the actual firing rate in the mean-driven regime (i.e., νs > 1) in the case of τs = 5 ms (Fig. 3A, bottom panel, pink dashed line with triangle markers). The overestimation or underestimation of the MMFM in the case of τe = 5 ms was also observed by Ly (2013). However, the demonstration that such overestimation or underestimation disappears when the synaptic time constant increases has not been reported.

Figure 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 3.

Input-output curve comparison. The top panels in A, B show the input-output curves computed in four ways: MCS (black solid line with diamond markers), fdPDM (blue solid line with square markers), csPDM (orange dashed line with circle markers), and MMFM (pink dashed line with triangle markers). Synaptic time constant (τs) is 5 ms in A and is 100 ms in B. Embedded Image is the steady-state population firing rate as a function of a fixed input rate νs. The bottom panels in A, B show the errors of fdPDM, csPDM, and MMFM in Embedded Image from MCS. ΔVs = 1 and Es = 0 mV in this example.

Dynamic population responses to excitatory input only

In addition to the steady-state analyses described in the above, the comparison between csPDM and MMFM is also made by investigating their ability to capture dynamic population firing rates in response to time-varying inputs. This is the most essential test to reveal whether they are adequate as dimension-reduction methods. Similar to the previous cases, excitatory inputs with τs = 5 and τs = 100 ms are considered here. The time-varying input rate is shown in Figure 4A, which varies in time and takes values that guarantee the validity of diffusion approximation, i.e., Embedded Image . Figure 4B shows the output population firing rates r(t) computed by MCS, fdPDM, csPDM, and MMFM and the corresponding errors in r(t) for τs = 5 ms. Figure 4C is similar to Figure 4B except τs = 100 ms. As shown in these two panels, whatever synaptic time constant is, csPDM accurately captures all the qualitative features of the output firing rates, leading to small average error ratios, which are 0.048 and 0.043 for τs = 5 and τs = 100 ms. It also gives comparable simulation results to those of fdPDM (similar average error ratios between fdPDM and csPDM). However, like in the steady-state analyses, MMFM gives accurate output firing rates only when τs = 100 ms. It overestimates the low output firing rates (at about t = 820 ms) and underestimates the high output firing rates (at about t = 730 or t = 850 ms) when τs = 5 ms, leading to a large error ratio Embedded Image of 0.179.

To show how robust the csPDM is, we explore whether the accuracy of csPDM depends on the values of model parameters. Four parameters, ΔVs, gl, κ and Vr, are chosen for this test. The same input rate as shown in the Figure 4A is considered as the input. As shown in Figure 5A, augmenting the jump size ΔVs increases Embedded Image , meaning that the accuracy of csPDM is decreased. This is due to the fact that diffusion approximation becomes invalid when ΔVs is too large (as shown in Fig. 1). As expected, Embedded Image is less sensitive to the change of ΔVs than Embedded Image because ΔVs is arbitrary in the fdPDM. Surprisingly, although MMFM does not have limitations on ΔVs, in the case of τs = 5 ms, Embedded Image severely fluctuates over a range of 0.2 when changing ΔVs. Figure 5B–Dunravel that changing gl, κ and Vr almost does not affect the accuracy of csPDM (Embedded Image is below 0.1 in most parameter sets), implying that csPDM is a robust method. Importantly, simulation results from csPDM are comparable to those from fdPDM. MMFM can give comparable results to csPDM only when τs = 100 ms. However, in the case of τs = 5 ms, reducing gl enlarges the error ratio Embedded Image with an increasing magnitude of 0.3 (from 0.07 to 0.38). So does increasing κ (from 0.1 to 0.4). In other words, MMFM is not a robust method for EIF models. Figure 6 shows the computational time spent by fdPDM and csPDM subject to different numbers of meshes employed in numerical methods for a simulation of 1 s. The numerical simulation used in this test is the same as shown in Figure 4. Remarkably, the computational efficiency of csPDM is ∼1000 times better than fdPDM. Low computational efficiency of fdPDM certainly comes from the existence of the gs-dimension in the master equation because, except for the requirement of more grid meshes along this dimension (We set 120 meshes along this dimension), its existence leads to the necessity of extremely small time steps for satisfying the Courant–Friedrichs–Lewy condition to ensure numerical stability of discontinuous Galerkin methods (Huang et al., 2015). In this case, the time step is 0.02 ms for fdPDM but 0.2 ms for csPDM.

Figure 4.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 4.

Dynamic population responses to a time-varying excitatory input. A, Time-varying input rate νs. B, top panel, Population responses r(t) in response to νs computed by MCS (shaded area), fdPDM (blue solid line with square markers), csPDM (orange dashed line with circle markers), and MMFM (pink dotted line with traiangle markers) under τs = 5 ms. Bottom panel, Corresponding errors of fdPDM, csPDM, and MMFM in r(t) from MCS. C, Similar to B except for τs = 100 ms. Average error ratios Embedded Image were computed over the interval from 600 and 1000 ms. Embedded Image , and Embedded Image are 0.058, 0.048, and 0.179 under τs = 5 ms, respectively. They are 0.042, 0.043, and 0.056 for τs = 100 ms. ΔVs = 1 and Es = 0 mV in this example.

Figure 5.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 5.

Average error ratios of fdPDM, csPDM, and MMFM in population responses under different parameter sets. A, Subject to varying ΔVs. B, Subject to varying gl. C, Subject to varying κ. D, Subject to varying Vr. Blue solid line with square markers: fdPDM. Orange dashed line with circle markers: csPDM. Pink dotted line with traiangle markers: MMFM. Embedded Image means average error ratios corresponding to fdPDM, csPDM, and MMFM.

Figure 6.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 6.

Comparison of computational time between fdPDM and csPDM. As shown, the computational speed of csPDM is almost 1000 times faster than fdPDM.

In the next two examples, we only use csPDM to estimate the population dynamics and compare the results with those of MCS to evaluate its performance because, via the examples above, it has been demonstrated that, compared to fdPDM and MMFM, only csPDM can provide simulation results efficiently and accurately.

Dynamic population responses with excitatory and inhibitory inputs

To highlight the outstanding ability of csPDM to achieve dimension reduction, here, we consider a real case where a population of cortical pyramidal neurons receives excitatory inputs from the neighboring pyramidal neuronal population mediated by AMPA receptors and, meanwhile, inhibitory inputs from the interneuron population mediated by two types of receptors, GABAA and GABAB (Suffczynski et al., 2004; Cona et al., 2014). To mimic this situation, we consider a single uncoupled population of EIF neurons, representing pyramidal neurons, which receive external excitatory inputs mediated by AMPA-type receptors, which have τs = 5 ms and ΔVs = 1 mV, and inhibitory inputs mediated by GABAA-type receptors, which have τs = 10 ms and ΔVs = 0.25 mV, as well as GABAB-type receptors, which have τs = 100 ms and ΔVs = 0.25 mV. Each neuron is assumed to have 200 excitatory and 200 inhibitory synaptic connections.

Figure 7A, top panel, shows the time-varying input rates on excitatory and inhibitory synaptic connections. It should be noted that the inputs gated by GABAA and GABAB receptors share the common input rate in statistical sense. The Figure 7A, second panel from top, is the raster plot of output spikes from 200 neurons in MCS. With respect to the output population firing rate, as seen in Figure 7A, third panel, csPDM is able to capture the salient features of the population firing rate. The average error ratio Embedded Image is ∼ 0.067 (below 0.1). It is ∼ 4 Hz of error in the high output firing rate (such as at t = 1900 ms; Fig. 7A, forth panel) on average. csPDM also can accurately estimate the average membrane voltage across the population (Fig. 7A, fifth panel). Figure 7B shows the snapshots of the marginal density function Embedded Image at t = 1700, 1800, and 1900 ms. As seen, the density functions computed by MCS and csPDM match well in all three snapshots. The corresponding values of Embedded Image are 0.065, 0.07, and 0.122, respectively. These results indicate that Embedded Image can be correctly captured by csPDM.

Figure 7.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 7.

MCS and csPDM for a single population receiving excitatory and inhibitory inputs gated by three types of neurotransmitter receptors. A, The panels from top to bottom show (1) time-varying excitatory and inhibitory input rates (νe denotes the excitatory input rate, and νi denotes the inhibitory one); (2) raster plot of output spikes from 200 neurons in MCS; (3) the population firing rates r(t) computed by MCS (gray solid line) and csPDM (black dashed line); (4) the error in r(t) between MCS and csPMD; and (5) average membrane voltages across the population r(t) computed by MCS (gray thick solid line) and csPDM (black thick dashed line). Gray thin solid lines: voltage traces from 10 neurons in MSC. B, Three snapshots of the marginal density function Embedded Image at t = 1700 , 1800, and 1900 ms from left to right. Shaded area: MCS. Black thick line: csPDM.

STP population responses

In this part, we test the effects of the synaptic STP on population responses to see if csPDM can capture these effects. We consider a single population of EIF neurons receiving only excitatory inputs through plastic synapses with short-term facilitation or depression. For the facilitative synapses, we adopt τs = 5 ms, Embedded Image mv Es = 0 mV, τs,f = 700 ms, τs,r = 100, and Us = 0.05. They are the same except τs,f = 50 ms, τs,r = 300, and Us = 0.2 for the depressive synapses. Figure 8A shows the time-varying input rate, which consists of a fixed baseline of 500 ms and, afterward, a sinusoidal fluctuation of 1 s. The population responses and ensemble averages of us and xs obtained from MSC and csPDM for the case of short-term facilitation are plotted in Figure 8B. As seen, the population response gradually increases in the latter cycles in the sinusoidal fluctuation (800–1500 ms). The facilitative response actually stems from the sharp increase in Embedded Image from 0.05 to ∼0.29. Meanwhile, the value of Embedded Image remains at a high level (Embedded Image , eventually). Figure 8C is similar to Figure 8B except that depressive synapses are used. As shown, the population response is depressed along the sinusoidal input, resulting from the large decrease in Embedded Image roughly by 0.4, during which Embedded Image remains almost fixed. Note that, different to the above example, where csPDM can give quantitatively accurate estimates of population responses, in this case, csPDM just captures STP property of population responses qualitatively. As seen in Figure 8C, the output firing rate obtained from csPDM drops but does not vanishes, while it vanishes in MCS. This difference is caused by the mismatch of Embedded Image between MSC and csPDM. In summary, the population response can inherit the STP property from synapses, and csPDM captures such property qualitatively.

Figure 8.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 8.

STP population responses. A, Time-varying excitatory input rate. B, Population responses under short-term facilitation. The population response is facilitated during the sinusoidal input, resulting from the sharp increase in the running utilization parameter, Embedded Image As shown, the csPDM can capture the STP property of the population responses not only qualitatively but also quantitatively. C, Population responses under short-term depression. The population response is depressed, resulting from the sharp decrease in the running fraction of available neurotransmitters, Embedded Image . In this case, csPDM only captures the STP property of the population responses qualitatively. The larger error in population responses is due to the overestimation of Embedded Image

Discussion

In this study, we present a principled and straightforward dimension-reduction method for PDM to handle realistic synaptic dynamics and compare it with another dimension-reduction method, called MMFM proposed by Ly (2013). We name the newly proposed reduction method csPDM. The csPDM does not assume specific limits on synaptic time constants so that we can consider synaptic dynamics mediated by all kinds of common receptors, including AMPA, GABAA, GABAB, and even NMDA. Through our examples, it is demonstrated that csPDM can accurately capture the firing rate responses in both the steady-state and dynamic regimes over a large range of synaptic time constants from milliseconds to hundreds of milliseconds.

As seen in Equation 24, the resulting quasi-Fokker–Planck equation leaves out all synapse-associated dimensions and just tracks the marginal density function of membrane voltages across the population. As a result, csPDM is extremely computationally efficient even when many types of receptors are included in the network model. The computational speed of csPDM is much faster than that of the original fdPDM by an order of three under the case where one type of receptors is incorporated (Fig. 6). If three types of receptors are incorporated, like in the third example in this study, fdPDM can never be considered as a computationally efficient modeling tool to simulate neural networks because of the inherent difficulty of solving a four-dimensional master equation. Actually, previous studies have argued that solving a master equation with more than three dimensions may spend much more time than MCS (Apfaltrer et al., 2006; Ly and Tranchina, 2007). However, csPDM can efficiently and correctly provides simulations of network activities in this case (Fig. 7). On the other hand, increasing the number of receptor types in the network model does not reduce the accuracy of csPDM considerably. Embedded Image is 0.048 when one type of receptors is included (Fig. 4) and is 0.067 when three types are included (Fig. 7).

When csPDM is compared with MMFM, some interesting findings are observed. First, it is found that MMFM gives quantitatively accurate simulation results only for long synaptic time constants in the steady-state and dynamic regimes (Figs. 3, 4). A possible explanation is that the membrane time constant (Embedded Image ) is longer than the synaptic time constant (τs) so that the use of the expected value of the steady-state firing rates for estimating the output firing rate (Eq. 18) becomes unreasonable because the neuronal dynamics cannot reach a steady state before synaptic dynamics does. This idea is also evidenced by the decline in Embedded Image when gl is increased, i.e., reducing the membrane time constant (Fig. 5B).

By robustness analysis, we found that the accuracy of MMFM is sensitively influenced by the value of κ (Fig. 5C), which controls the nonlinearity of EIF neuronal dynamics by adjusting the strength of exponential currents (Eq. 1). Increasing κ, i.e., increasing the nonlinearity, decreases the accuracy of MMFM. Actually, MMFM relies on the assumption that the membrane voltage inherits randomness entirely from the stochastic conductance variables, and the conductance variables evolve independently of the membrane voltage (Ly, 2013). However, the author claimed that such an assumption might only be reasonable for linear models such as leaky integrate-and-fire neurons. So, MMFM may not be suitable to simulate network dynamics of nonlinear spiking neurons, like the EIF neurons used in this study. On the contrary, csPDM only assumes that the conductance variables evolve isolatedly, which is true no matter what neuron models are used. Accordingly, it is more general for the type of neuronal models. Furthermore, in the csPDM, the membrane voltage is allowed to be influenced by other random sources, for example, the random release of neurotransmitters (Faisal et al., 2008; Destexhe and Rudolph-Lilith, 2012), which may just add extra noise terms into the stochastic diffusion tensor in the quasi-Fokker–Planck equation without increasing its dimension.csPDM relies on the diffusion approximation of conductance variables. Physiologically, the assumption of diffusion approximation is quite reasonable in that the number of synaptic contacts covering a typical cortical pyramidal cell is indeed very high, in the range of 103–104 (Braitenberg and Schüz, 1991). Each spike at a synaptic contact only alters a post-synaptic potential of <1 mV on average (Markram and Tsodyks, 1996; Markram et al., 1998). So, in our cases, each excitatory or inhibitory spike is designed to cause a maximum change of ≈1 mV in the membrane potential.

Our study is similar to the work of Rudolph and Destexhe (2003), in which the authors employed Ito^ calculus to formulate the Fokker–Planck equation for Langevin equations and, then, to obtain an analytic expression for the membrane potential distribution at the steady state. They found that steady-state solutions are only available under a limited range of parameter sets. To correct this problem, they used an extended analytic expression for the membrane potential distribution (Rudolph and Destexhe, 2005), by which new solutions became available for a wider range of parameter sets, although the parameters resided in the nonphysiological extreme limits. The major difference of our approach with theirs is the method used to derive the Fokker–Planck equation. In our study, it is based on a modified large-eddy-diffusivity closure method proposed by Barajas-Solano and Tartakovsky (2016), and a quasi-Fokker–Planck equation, i.e., Equation 24, governing the temporal evolution of the membrane potential distribution, is then derived. Although we do not test whether Equation 24 is valid for any parameter sets, by means of our examples in this study, it is found that it at least provides a good solution for parameter sets in physiologic ranges. However, surprisingly, when we apply the same derivation procedures to the neuronal model adopted by them (i.e., Eq. 4.1 in Rudolph and Destexhe, 2005), we obtain the same Fokker–Planck equation as what they derived, in which the diffusion coefficient is also rescaled by Embedded Image (τm is the membrane time constant at rest). This fact implies that Equation 24 might be adequate for an even wider range of parameter sets.

It is theoretically possible to extend our method to any neuron models because no premises are put on the neuronal dynamics. We have tested it on adaptive EIF models (Brette and Gerstner, 2005) and leaky integrate-and-fire-or-burst models (Smith et al., 2000) and find that csPDM works well (data not shown). Such an extension to adaptive EIF neurons is important for the development of PDM in computational neuroscience because the adaptive EIF model has been shown to reproduce a variety of spiking patterns observed in in vivo thalamic or cortical neurons (Naud et al., 2008; Touboul and Brette, 2008) and well predict spike times in response to external inputs (Jolivet et al., 2008). csPDM with adaptive EIF models therefore will be a critical research subject in computational neuroscience. We suggest that the synaptic dynamics should be restricted to the first-order kinetics because the calculation of the diffusion tensor, D or Embedded Image , is based on an assumption that fluctuation velocity components are exponentially autocorrelated. The calculation becomes more difficult if higher-order synaptic dynamics is used. csPDM only captures STP property of population responses in a qualitative way due to incorrect estimation of Embedded Image . This poor estimation is caused by a wrong assumption that us and xs are statistically independent for deriving mean-field equations for Embedded Image and Embedded Image (Tsodyks et al., 1998). In fact, us is independent of xs because xs does not appear in the Equation 6, but, by contrast, xs has a negative correlation with us. As a consequence, the mean-field equation for Embedded Image should include higher-order statistics of both us and xs. However, such an equation is unclosed. Before a closure method is available, the assumption of independence between us and xs is a compromised solution (Tsodyks et al., 1998). Note that it is possible to replace the phenomenological model of STP used here with another elaborating model where state variables are independent with each other. If so, the STP property of population responses can be quantitatively captured by csPDM. This is one of the directions of our future work for improving csPDM.

The derivations of csPDM and numerical examples used for testing are based on a single uncoupled population. Of course, what we derived in this study can be extended to the applications of simulating larger neural networks consisting of multiple coupled populations. However, for the correct use of csPDM, synaptic connections must be sparse, because a large number of synaptic connections can result in the violation of basic assumptions of PDM–Poisson and statistically uncorrelated inputs received by each neuron (Brunel, 2000; Nykamp and Tranchina, 2000; Huertas and Smith, 2006; Augustin et al., 2013). These basic premises can be violated when two neurons share many common pre-synaptic neurons, that is, when there are a huge number of connections between presynaptic and postsynaptic populations. We found that the number of synaptic connections depends on the size of ΔVs. If ΔVs is large, the number of connections has to be small because postsynaptic neurons have to receive statistically independent inputs. Postsynaptic neurons are more likely to receive exactly the same inputs if ΔVs is large and the number of synaptic connections is huge.

Conclusions

When including many state variables, such as synaptic conductance variables for different types of receptors and synaptic facilitation and depression state variables, the original PDM becomes impractical because of the incredible increase in the computational load. In this study, this critical issue is solved by the probability density function method for Langevin equations with colored driving noise. The newly proposed method, termed csPDM, gives quantitatively accurate firing rate responses in both the steady-state and nonequilibrium regimes and highly qualified estimations of time-varying marginal density functions of membrane voltages. Importantly, csPDM is able to qualitatively manifest STP in an easy way almost without increasing computation demands. It appears that csPDM is generally applicable as a time-saving tool for modeling large-scale neural networks. We hope that our work will inspire further progress in the development of PDM and benefit computational and theoretical studies of synaptic dynamics in network dynamics.

Footnotes

  • The authors declare no competing financial interests.

  • This work was supported by the National Cheng Kung University Hospital and funded by a grant (MOST 105-2314-B-006-080-MY3) from the Ministry of Science and Technology, Taiwan.

This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International license, which permits unrestricted use, distribution and reproduction in any medium provided that the original work is properly attributed.

References

  1. ↵
    Apfaltrer F, Ly C, Tranchina D (2006) Population density methods for stochastic neurons with realistic synaptic kinetics: firing rate dynamics and fast computational methods. Network 17:373–418. doi:10.1080/09548980601069787 pmid:17162461
    OpenUrlCrossRefPubMed
  2. ↵
    Augustin M, Ladenbauer J, Obermayer K (2013) How adaptation shapes spike rate oscillations in recurrent neuronal networks. Front Comput Neurosci 7:9. doi:10.3389/fncom.2013.00009 pmid:23450654
    OpenUrlCrossRefPubMed
  3. ↵
    Barajas-Solano DA, Tartakovsky AM (2016) Probabilistic density function method for nonlinear dynamical systems driven by colored noise. Phys Rev E 93:52121. doi:10.1103/PhysRevE.93.052121
    OpenUrlCrossRef
  4. ↵
    Barak O, Tsodyks M (2007) Persistent activity in neural networks with dynamic synapses. PLoS Comput Biol 3:e35. doi:10.1371/journal.pcbi.0030035 pmid:17319739
    OpenUrlCrossRefPubMed
  5. ↵
    Biswas R, Devine KD, Flaherty JE (1994) Parallel, adaptive finite element methods for conservation laws. Appl Numer Math 14:255–283. doi:10.1016/0168-9274(94)90029-9
    OpenUrlCrossRef
  6. ↵
    Braitenberg V, Schüz A (1991) Anatomy of the cortex: statistics and geometry. New York: Springer.
  7. ↵
    Brette R, Gerstner W (2005) Adaptive exponential integrate-and-fire model as an effective description of neuronal activity. J Neurophysiol 94:3637–3642. doi:10.1152/jn.00686.2005 pmid:16014787
    OpenUrlCrossRefPubMed
  8. ↵
    Brunel N (2000) Dynamics of sparsely connected networks of excitatory and inhibitory spiking neurons. J Comput Neurosci 8:183–208. pmid:10809012
    OpenUrlCrossRefPubMed
  9. ↵
    Brunel N, Sergi S (1998) Firing frequency of leaky intergrate-and-fire neurons with synaptic current dynamics. J Theor Biol 195:87–95. doi:10.1006/jtbi.1998.0782 pmid:9802952
    OpenUrlCrossRefPubMed
  10. ↵
    Cai D, Tao L, Shkarayev MS, Rangan AV, McLaughlin DW, Kovačič G (2012) The role of fluctuations in coarse-grained descriptions of neuronal networks. Commun Math Sci 10:307–354. doi:10.4310/CMS.2012.v10.n1.a14
    OpenUrlCrossRef
  11. ↵
    Casti AR, Omurtag A, Sornborger A, Kaplan E, Knight B, Victor J, Sirovich L (2002) A population study of integrate-and-fire-or-burst neurons. Neural Comput 14:957–986. doi:10.1162/089976602753633349 pmid:11972903
    OpenUrlCrossRefPubMed
  12. ↵
    Cheng YD, Shu CW (2007) A discontinuous Galerkin finite element method for time dependent partial differential equations with higher order derivatives. Math Comput 77:699–730. doi:10.1090/S0025-5718-07-02045-5
    OpenUrlCrossRef
  13. ↵
    Cockburn B, Shu CW (1998) The local discontinuous Galerkin method for time-dependent convection-diffusion systems. SIAM J Numer Anal 35:2440–2463. doi:10.1137/S0036142997316712
    OpenUrlCrossRef
  14. ↵
    Cona F, Lacanna M, Ursino M (2014) A thalamo-cortical neural mass model for the simulation of brain rhythms during sleep. J Comput Neurosci 37:125–148. doi:10.1007/s10827-013-0493-1 pmid:24402459
    OpenUrlCrossRefPubMed
  15. ↵
    Destexhe A (2009) Self-sustained asynchronous irregular states and Up–Down states in thalamic, cortical and thalamocortical networks of nonlinear integrate-and-fire neurons. J Comput Neurosci 27:493–506. doi:10.1007/s10827-009-0164-4 pmid:19499317
    OpenUrlCrossRefPubMed
  16. ↵
    Destexhe A, Rudolph-Lilith M (2012) Neuronal noise. New York: Springer.
  17. ↵
    Dayan P, Abbott LF (2001) Theoretical neuroscience. Cambridge, MA: MIT Press.
  18. ↵
    Faisal AA, Selen LPJ, Wolpert DM (2008) Noise in the nervous system. Nat Rev Neurosci 9:292–303. doi:10.1038/nrn2258 pmid:18319728
    OpenUrlCrossRefPubMed
  19. ↵
    Fourcaud-Trocmé N, Hansel D, van Vreeswijk C, Brunel N (2003) How spike generation mechanisms determine the neuronal response to fluctuating inputs. J Neurosci 23:11628–11640. pmid:14684865
    OpenUrlAbstract/FREE Full Text
  20. ↵
    Fung CCA, Wong KYM, Wang H, Wu S (2012) Dynamical synapses enhance neural information processing: gracefulness, accuracy, and mobility. Neural Comput 24:1147–1185. doi:10.1162/NECO_a_00269
    OpenUrlCrossRef
  21. ↵
    Gritsun T, Le Feber J, Stegenga J, Rutten WLC (2011) Experimental analysis and computational modeling of interburst intervals in spontaneous activity of cortical neuronal culture. Biol Cybern 105:197–210. doi:10.1007/s00422-011-0457-3
    OpenUrlCrossRefPubMed
  22. ↵
    Haskell E, Nykamp DQ, Tranchina D (2001) Population density methods for large-scale modelling of neuronal networks with realistic synaptic kinetics: cutting the dimension down to size. Network 12:141–174. doi:10.1080/net.12.2.141.174
    OpenUrlCrossRefPubMed
  23. ↵
    Hertäg L, Durstewitz D, Brunel N (2014) Analytical approximations of the firing rate of an adaptive exponential integrate-and-fire neuron in the presence of synaptic noise. Front Comput Neurosci 8:116. doi:10.3389/fncom.2014.00116 pmid:25278872
    OpenUrlCrossRefPubMed
  24. ↵
    Huang CH, Lin CCK, Ju MS (2015) Discontinuous Galerkin finite element method for solving population density functions of cortical pyramidal and thalamic neuronal populations. Comput Biol Med 57:150–158. doi:10.1016/j.compbiomed.2014.12.011 pmid:25557200
    OpenUrlCrossRefPubMed
  25. ↵
    Huertas MA, Smith GD (2006) A multivariate population density model of the dLGN/PGN relay. J Comput Neurosci 21:171–189. doi:10.1007/s10827-006-7753-2 pmid:16788765
    OpenUrlCrossRefPubMed
  26. ↵
    Jolivet R, Schürmann F, Berger TK, Naud R, Gerstner W, Roth A (2008) The quantitative single-neuron modeling competition. Biol Cybern 99:417–426. doi:10.1007/s00422-008-0261-x pmid:19011928
    OpenUrlCrossRefPubMed
  27. ↵
    Knight BW, Manin D, Sirovich L (1996) Dynamical models of interacting neuron populations in visual cortex. Symposium on robotics and cybernetics: Computational engineering in systems applications (Gerf EC, ed). Lille, France: Cite Scientifique.
  28. ↵
    Levina A, Herrmann JM, Geisel T (2007) Dynamical synapses causing self-organized criticality in neural networks. Nat Phys 3:857–860. doi:10.1038/nphys758
    OpenUrlCrossRef
  29. ↵
    Ly C (2013) A principled dimension-reduction method for the population density approach to modeling networks of neurons with synaptic dynamics. Neural Comput 25:2682–2708. doi:10.1162/NECO_a_00489 pmid:23777517
    OpenUrlCrossRefPubMed
  30. ↵
    Ly C, Tranchina D (2007) Critical analysis of dimension reduction by a moment closure method in a population density approach to neural network modeling. Neural Comput 19:2032–2092. doi:10.1162/neco.2007.19.8.2032
    OpenUrlCrossRefPubMed
  31. ↵
    Markram H, Tsodyks M (1996) Redistribution of synaptic efficacy between neocortical pyramidal neurons. Nature 382:807. doi:10.1038/382807a0 pmid:8752273
    OpenUrlCrossRefPubMed
  32. ↵
    Markram H, Wang Y, Tsodyks M (1998) Differential signaling via the same axon of neocortical pyramidal neurons. Proc Natl Acad Sci USA 95:5323–5328. pmid:9560274
    OpenUrlAbstract/FREE Full Text
  33. ↵
    Mongillo G, Barak O, Tsodyks M (2008) Synaptic theory of working memory. Science 319:1543–1546. doi:10.1126/science.1150769 pmid:18339943
    OpenUrlAbstract/FREE Full Text
  34. ↵
    Moreno-Bote R, Parga N (2004) Role of synaptic filtering on the firing response of simple model neurons. Phys Rev Lett 92:28102. doi:10.1103/PhysRevLett.92.028102
    OpenUrlCrossRef
  35. ↵
    Naud R, Marcille N, Clopath C, Gerstner W (2008) Firing patterns in the adaptive exponential integrate-and-fire model. Biol Cybern 99:335–337. doi:10.1007/s00422-008-0264-7 pmid:19011922
    OpenUrlCrossRefPubMed
  36. ↵
    Nykamp DQ, Tranchina D (2000) A population density approach that facilitates large-scale modeling of neural networks: analysis and an application to orientation tuning. J Comput Neurosci 8:19–50.
    OpenUrlCrossRefPubMed
  37. ↵
    Nykamp DQ, Tranchina D (2001) A population density approach that facilitates large-scale modeling of neural networks: extension to slow inhibitory synapses. Neural Comput 13:511–546. pmid:11244554
    OpenUrlCrossRefPubMed
  38. ↵
    Omurtag A, Knight BW, Sirovich L (2000) On the simulation of large populations of neurons. J Comput Neurosci 8:51–63. pmid:10798499
    OpenUrlCrossRefPubMed
  39. ↵
    Rangan AV, Cai D (2006) Maximum-entropy closures for kinetic theories of neuronal network dynamics. Phys Rev Lett 96:178101. doi:10.1103/PhysRevLett.96.178101 pmid:16712338
    OpenUrlCrossRefPubMed
  40. ↵
    Rangan AV, Kovačič G, Cai D (2008) Kinetic theory for neuronal networks with fast and slow excitatory conductances driven by the same spike train. Phys Rev E 77:41915.
    OpenUrl
  41. ↵
    Richardson MJE (2004) Effects of synaptic conductance on the voltage distribution and firing rate of spiking neurons. Phys Rev E Stat Nonlin Soft Matter Phys 69:51918. doi:10.1103/PhysRevE.69.051918 pmid:15244858
    OpenUrlCrossRefPubMed
  42. ↵
    Richardson MJE, Gerstner W (2005) Synaptic shot noise and conductance fluctuations affect the membrane voltage with equal significance. Neural Comput 17:923–947. doi:10.1162/0899766053429444
    OpenUrlCrossRefPubMed
  43. ↵
    Risken H (1996) The Fokker–Planck equation. Berlin: Springer.
  44. ↵
    Rudolph M, Destexhe A (2003) Characterization of subthreshold voltage fluctuations in neuronal membranes. Neural Comput 15:2577–2618. doi:10.1162/089976603322385081 pmid:14577855
    OpenUrlCrossRefPubMed
  45. ↵
    Rudolph M, Destexhe A (2005) An extended analytic expression for the membrane potential distribution of conductance-based synaptic noise. Neural Comput 17:2301–2315. doi:10.1162/0899766054796932 pmid:16156930
    OpenUrlCrossRefPubMed
  46. ↵
    Smith GD, Cox CL, Sherman SM, Rinzel J (2000) Fourier analysis of sinusoidally driven thalamocortical relay neurons and a minimal integrate-and-fire-or-burst model. J Neurophysiol 83:588–610. doi:10.1152/jn.2000.83.1.588
    OpenUrlCrossRefPubMed
  47. ↵
    Stimberg M, Goodman D, Benichoux V, Brette R (2014) Equation-oriented specification of neural models for simulations. Front Neuroinform 8:6. doi:10.3389/fninf.2014.00006 pmid:24550820
    OpenUrlCrossRefPubMed
  48. ↵
    Suffczynski P, Kalitzin S, da Silva LFH (2004) Dynamics of non-convulsive epileptic phenomena modeled by a bistable neuronal network. Neuroscience 126:467–484. doi:10.1016/j.neuroscience.2004.03.014
    OpenUrlCrossRefPubMed
  49. ↵
    Touboul J, Brette R (2008) Dynamics and bifurcations of the adaptive exponential integrate-and-fire model. Biol Cybern 99:319–334. doi:10.1007/s00422-008-0267-4 pmid:19011921
    OpenUrlCrossRefPubMed
  50. ↵
    Tranchina D (2009) Population density methods in large-scale neural network modelling. In: Stochastic methods in neuroscience ( Laing C, Lord GJ , eds), pp 281–316. Oxford, UK: Oxford University Press.
  51. ↵
    Tsodyks M, Pawelzik K, Markram H (1998) Neural networks with dynamic synapses. Neural Comput 10:821–835. pmid:9573407
    OpenUrlCrossRefPubMed
  52. ↵
    Uhlenbeck GE, Ornstein LS (1930) On the theory of the Brownian motion. Phys Rev 36:823–841. doi:10.1103/PhysRev.36.823
    OpenUrlCrossRef
  53. ↵
    Volman V, Gerkin RC, Lau PM, Ben-Jacob E, Bi GQ (2007) Calcium and synaptic dynamics underlying reverberatory activity in neuronal networks. Phys Biol 4:91–103. doi:10.1088/1478-3975/4/2/003 pmid:17664654
    OpenUrlCrossRefPubMed
  54. ↵
    Wang P, Tartakovsky AM, Tartakovsky DM (2013) Probability density function method for Langevin equations with colored noise. Phys Rev Lett 110:140602. doi:10.1103/PhysRevLett.110.140602 pmid:25166972
    OpenUrlCrossRefPubMed
  55. ↵
    Xu Y, Shu CW (2010) Local discontinuous Galerkin methods for high-order time-dependent partial differential equations. Commun Comput Phys 7:1–46.
    OpenUrl
  56. ↵
    Zucker RS, Regehr WG (2002) Short-term synaptic plasticity. Annu Rev Physiol 64:355–405. doi:10.1146/annurev.physiol.64.092501.114547 pmid:11826273
    OpenUrlCrossRefPubMed

Synthesis

Reviewing Editor: Gustavo Deco, Universitat Pompeu Fabra

Decisions are customarily a result of the Reviewing Editor and the peer reviewers coming together and discussing their recommendations until a consensus is reached. When revisions are invited, a fact-based synthesis statement explaining their decision and outlining what is needed to prepare a revision will be listed below.

Satisfied with all improvements.

Back to top

In this issue

eneuro: 5 (6)
eNeuro
Vol. 5, Issue 6
November/December 2018
  • Table of Contents
  • Index by author
Email

Thank you for sharing this eNeuro article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
An Efficient Population Density Method for Modeling Neural Networks with Synaptic Dynamics Manifesting Finite Relaxation Time and Short-Term Plasticity
(Your Name) has forwarded a page to you from eNeuro
(Your Name) thought you would be interested in this article in eNeuro.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Print
View Full Page PDF
Citation Tools
An Efficient Population Density Method for Modeling Neural Networks with Synaptic Dynamics Manifesting Finite Relaxation Time and Short-Term Plasticity
Chih-Hsu Huang, Chou-Ching K. Lin
eNeuro 6 December 2018, 5 (6) ENEURO.0002-18.2018; DOI: 10.1523/ENEURO.0002-18.2018

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Share
An Efficient Population Density Method for Modeling Neural Networks with Synaptic Dynamics Manifesting Finite Relaxation Time and Short-Term Plasticity
Chih-Hsu Huang, Chou-Ching K. Lin
eNeuro 6 December 2018, 5 (6) ENEURO.0002-18.2018; DOI: 10.1523/ENEURO.0002-18.2018
Reddit logo Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Abstract
    • Significance Statement
    • Introduction
    • Materials and Methods
    • Results
    • Discussion
    • Conclusions
    • Footnotes
    • References
    • Synthesis
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF

Keywords

  • Colored noise
  • Diffusion approximation
  • Exponential integrate-and-fire model
  • Langevin equations

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

Methods/New Tools

  • Bicistronic expression of a high-performance calcium indicator and opsin for all-optical stimulation and imaging at cellular resolution
  • A Toolbox of Criteria for Distinguishing Cajal–Retzius Cells from Other Neuronal Types in the Postnatal Mouse Hippocampus
  • Superficial Bound of the Depth Limit of Two-Photon Imaging in Mouse Brain
Show more Methods/New Tools

Novel Tools and Methods

  • Behavioral and Functional Brain Activity Alterations Induced by TMS Coils with Different Spatial Distributions
  • Bicistronic expression of a high-performance calcium indicator and opsin for all-optical stimulation and imaging at cellular resolution
  • Synthetic Data Resource and Benchmarks for Time Cell Analysis and Detection Algorithms
Show more Novel Tools and Methods

Subjects

  • Novel Tools and Methods

  • Home
  • Alerts
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Latest Articles
  • Issue Archive
  • Blog
  • Browse by Topic

Information

  • For Authors
  • For the Media

About

  • About the Journal
  • Editorial Board
  • Privacy Policy
  • Contact
  • Feedback
(eNeuro logo)
(SfN logo)

Copyright © 2023 by the Society for Neuroscience.
eNeuro eISSN: 2373-2822

The ideas and opinions expressed in eNeuro do not necessarily reflect those of SfN or the eNeuro Editorial Board. Publication of an advertisement or other product mention in eNeuro should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in eNeuro.