Abstract
Memory retrieval is of central importance to a wide variety of brain functions. To understand the dynamic nature of memory retrieval and its underlying neurophysiological mechanisms, we develop a biologically plausible spiking neural circuit model, and demonstrate that free memory retrieval of sequences of events naturally arises from the model under the condition of excitation-inhibition (E/I) balance. Using the mean-field model of the spiking circuit, we gain further theoretical insights into how such memory retrieval emerges. We show that the spiking neural circuit model quantitatively reproduces several salient features of free memory retrieval, including its semantic proximity effect and log-normal distributions of inter-retrieval intervals. In addition, we demonstrate that our model can serve as a platform to examine memory retrieval deficits observed in neuropsychiatric diseases such as Parkinson’s and Alzheimer’s diseases. Furthermore, our model allows us to make novel and experimentally testable predictions, such as the prediction that there are long-range correlations in the sequences of retrieved items.
Similar content being viewed by others
References
Alvarez-Lacalle, E., Dorow, B., Eckmann, J.-P., & Moses, E. (2006). Hierarchical structures induce long-range dynamical correlations in written texts. Proceedings of the National Academy of Sciences, 103(21), 7956–7961.
Amit, D.J., & Brunel, N. (1997). Model of global spontaneous activity and local structured activity during delay periods in the cerebral cortex. Cerebral Cortex, 7(3), 237–252.
Austerweil, J.L., Abbott, J.T., & Griffiths, T.L. (2012). Human memory search as a random walk in a network. In Advances in neural information processing systems (pp. 3041–3049).
Baddeley, A.D. (1997). Human Memory: Theory and Practice: Psychology Press.
Begeer, S., Wierda, M., Scheeren, A.M., Teunisse, J.-P., Koot, H.M., & Geurts, H.M. (2014). Verbal fluency in children with autism spectrum disorders: clustering and switching strategies. Autism, 18(8), 1014–1018.
Botvinick, M.M. (2008). Hierarchical models of behavior and prefrontal function. Trends in Cognitive Sciences, 12(5), 201–208.
Braitenberg, V., & Schüz, A. (1991). Anatomy of the cortex: statistics and geometry. Springer-Verlag Publishing.
Brunel, N., & Latham, P.E. (2003). Firing rate of the noisy quadratic integrate-and-fire neuron. Neural Computation, 15(10), 2281–2306.
Brunel, N., & Sergi, S (1998). Firing frequency of leaky intergrate-and-fire neurons with synaptic current dynamics. Journal of Theoretical Biology, 195(1), 87–95.
Catani, M.H., & ffytche, D. (2005). The rises and falls of disconnection syndromes. Brain, 128(10), 2224–2239.
Chandrasekhar, S. (1943). Stochastic problems in physics and astronomy. Reviews of Modern Physics, 15(1), 1.
Cline, H. (2005). Synaptogenesis: a balancing act between excitation and inhibition. Current Biology, 15(6), R203—R205.
Collins, A.M., & Quillian, M.R. (1969). Retrieval time from memory. Journal of Verbal Learning and Verbal Behavior, 8(2), 240–247.
Cronin-Golomb, A. (2010). Parkinson’s disease as a disconnection syndrome. Neuropsychology Review, 20(2), 191–208.
Curti, E., Mongillo, G., La Camera, G., & Amit, D.J. (2004). Mean field and capacity in realistic networks of spiking neurons storing sparsely coded random memories. Neural Computation, 16(12), 2597–2637.
Dayan, P., & Abbott, L.F. (2001). Theoretical neuroscience. MIT Press.
Deco, G., & Hugues, E. (2012). Neural network mechanisms underlying stimulus driven variability reduction. PLoS Computational Biology, 8(3), e1002395–e1002395.
Delbeuck, X., Van der Linden, M., & Collette, F. (2003). Alzheimer’s disease as a disconnection syndrome? Neuropsychology Review, 13(2), 79–92.
Destexhe, A., Mainen, Z.F., & Sejnowski, T.J. (1994). An efficient method for computing synaptic conductances based on a kinetic model of receptor binding. Neural Computation, 6(1), 14–18.
Destexhe, A., Mainen, Z.F., & Sejnowski, T.J. (1998). kinetic models of synaptic transmission. In Koch, C., & Segev, I. (Eds.), Methods in Neuronal Modeling. 2nd edn. (pp. 1–25). Cambridge: MIT Press.
Destexhe, A., Rudolph, M., & Paré, D (2003). The high-conductance state of neocortical neurons in vivo. Nature Reviews Neuroscience, 4(9), 739–751.
Deuker, L., Olligs, J., Fell, J., Kranz, T. A., Mormann, F., Montag, C., & et al. (2013). Memory consolidation by replay of stimulus-specific neural activity. The Journal of Neuroscience, 33(49), 19373–19383.
Edwards, A.M., Phillips, R.A., Watkins, N.W., Freeman, M.P., Murphy, E.J., Afanasyev, V., & et al. (2007). Revisiting lévy flight search patterns of wandering albatrosses, bumblebees and deer. Nature, 449 (7165), 1044–1048.
Felleman, D.J., & Van Essen, D.C. (1991). Distributed hierarchical processing in the primate cerebral cortex. Cerebral Cortex, 1(1), 1–47.
Fino, E., & Yuste, R. (2011). Dense inhibitory connectivity in neocortex. Neuron, 69(6), 1188–1203.
Foster, D.J., & Wilson, M.A. (2006). Reverse replay of behavioural sequences in hippocampal place cells during the awake state. Nature, 440(7084), 680–683.
Fourcaud, N., & Brunel, N. (2002). Dynamics of the firing probability of noisy integrate-and-fire neurons. Neural Computation, 14(9), 2057–2110.
Friston, K.J. (1998). The disconnection hypothesis. Schizophrenia Research, 30(2), 115–125.
Gerstner, W., Kistler, W.M., Naud, R., & Paninski, L. (2014). Neuronal dynamics: from single neurons to networks and models of cognition. Cambridge University Press.
Gillespie, D.T. (1991). Markov processes: an introduction for physical scientists. Elsevier.
Goñi, J., Arrondo, G., Sepulcre, J., Martincorena, I.N., Vélez De Mendizábal, N., Corominas-Murtra, B., & et al. (2011). The organization of the animal category: evidence from verbal fluency and network theory. Cognitive Processing, 12(2), 183–196.
Griffiths, T.L., Steyvers, M., & Firl, A. (2007). Google and the mind predicting fluency with pagerank. Psychological Science, 18(12), 1069–1076.
Haider, B., Duque, A., Hasenstaub, A.R., & McCormick, D.A (2006). Neocortical network activity in vivo is generated through a dynamic balance of excitation and inhibition. The Journal of Neuroscience, 26(17), 4535–4545.
Hänggi, P., Talkner, P., & Borkovec, M. (1990). Reaction-rate theory: fifty years after kramers. Reviews of Modern Physics, 62(2), 251.
Haugrud, N., Crossley, M., & Vrbancic, M. (2011). Clustering and switching strategies during verbal fluency performance differentiate alzheimer’s disease and healthy aging. Journal of the International Neuropsychological Society, 17(6), 1153–1157.
Hills, T.T., Jones, M.N., & Todd, P.M. (2012). Optimal foraging in semantic memory. Psychological Review, 119(2), 431.
Hills, T.T., Todd, P.M., Lazer, D., Redish, A.D., Couzin, I.D., & Group, C.S.R (2015). Exploration versus exploitation in space, mind, and society. Trends in Cognitive Sciences, 19(1), 46–54.
Hooke, R., & Jeeves, T.A. (1961). Direct search solution of numerical and statistical problems. Journal of the ACM, 8(2), 212–229.
Hopfield, J.J. (1982). Neural networks and physical systems with emergent collective computational abilities. Proceedings of the National Academy of Sciences, 79(8), 2554–2558.
Johnson, A., & Redish, A.D. (2007). Neural ensembles in ca3 transiently encode paths forward of the animal at a decision point. The Journal of Neuroscience, 27(45), 12176–12189.
Kass, R.E., & Raftery, A.E. (1995). Bayes factors. Journal of the American Statistical Association, 90(430), 773–795.
Keane, A., & Gong, P. (2015). Propagating waves can explain irregular neural dynamics. The Journal of Neuroscience, 35(4), 1591– 1605.
Kleinfeld, D. (1986). Sequential state generation by model neural networks. Proceedings of the National Academy of Sciences, 83(24), 9469–9473.
Koch, C. (1998). Biophysics of computation: Information processing in single neurons. Oxford University Press.
Kuhl, B.A., & Chun, M.M. (2014). Successful remembering elicits event-specific activity patterns in lateral parietal cortex. The Journal of Neuroscience, 34(23), 8051–8060.
Landauer, T.K., Foltz, P.W., & Laham, D. (1998). An introduction to latent analysis. Discourse processes, 25(2-3), 259–284.
Litwin-Kumar, A., & Doiron, B. (2012). Slow dynamics and high variability in balanced cortical networks with clustered connections. Nature Neuroscience, 15(11), 1498–1505.
Markov, N.T., Ercsey-Ravasz, M., Van Essen, D.C., Knoblauch, K., Toroczkai, Z., & Kennedy, H. (2013). Cortical high-density counterstream architectures. Science, 342(6158), 1238406.
Mascaro, M., & Amit, D.J. (1999). Effective neural response function for collective population states. Network: Computation in Neural Systems, 10(4), 351–373.
Mazzucato, L., Fontanini, A., & La Camera, G. (2015). Dynamics of multistable states during ongoing and evoked cortical activity. The Journal of Neuroscience, 35(21), 8214–8231.
McKenzie, S., Frank, A.J., Kinsky, N.R., Porter, B., Rivière, P. D., & Eichenbaum, H. (2014). Hippocampal representation of related and opposing memories develop within distinct, hierarchically organized neural schemas. Neuron, 83(1), 202–215.
Meunier, D., Lambiotte, R., & Bullmore, E.T. (2010). Modular and hierarchically modular organization of brain networks. Frontiers in Neuroscience, 4.
Mishkin, M., Suzuki, W.A., Gadian, D.G., & Vargha-Khadem, F. (1997). Hierarchical organization of cognitive memory. Philosophical Transactions of the Royal Society B: Biological Sciences, 352(1360), 1461–1467.
Mongillo, G., Amit, D.J., & Brunel, N. (2003). Retrospective and prospective persistent activity induced by hebbian learning in a recurrent cortical network. European Journal of Neuroscience, 18(7), 2011–2024.
Nørrelykke, S.F., & Flyvbjerg, H. (2011). Harmonic oscillator in heat bath: exact simulation of time-lapse-recorded data and exact analytical benchmark statistics. Physical Review E, 83(4), 041103.
Packer, A.M., & Yuste, R. (2011). Dense, unspecific connectivity of neocortical parvalbumin-positive interneurons: a canonical microcircuit for inhibition? The Journal of Neuroscience, 31(37), 13260–13271.
Palop, J.J., & Mucke, L. (2009). Epilepsy and cognitive impairments in alzheimer’s disease. Archives of Neurology, 66(4), 435– 440.
Peyrache, A., Dehghani, N., Eskandar, E.N., Madsen, J.R., Anderson, W.S., Donoghue, J.A., & et al. (2012). Spatiotemporal dynamics of neocortical excitation and inhibition during human sleep. Proceedings of the National Academy of Sciences, 109(5), 1731–1736.
Pfeiffer, B.E., & Foster, D.J. (2015). Autoassociative dynamics in the generation of sequences of hippocampal place cells. Science, 349(6244), 180–183.
Raaijmakers, J.G., & Shiffrin, R.M. (1981). Search of associative memory. Psychological Review, 88(2), 93.
Rainer, G., Rao, S.C., & Miller, E.K. (1999). Prospective coding for objects in primate prefrontal cortex. The Journal of Neuroscience, 19(13), 5493–5505.
Renart, A., Brunel, N., & Wang, X.-J. (2003). Mean-field theory of irregularly spiking neuronal populations and working memory in recurrent cortical networks. In Feng, J. (Ed.) Computational Neuroscience: A Comprehensive Approach (pp. 431–490). Chapman and Hall, CRC Press.
Renart, A, de la Rocha, J., Bartho, P., Hollender, L., Parga, N., Reyes, A., & et al. (2010). The asynchronous state in cortical circuits. Science, 327(5965), 587–590.
Rhodes, T., & Turvey, M.T. (2007). Human memory retrieval as lévy foraging. Physica A: Statistical Mechanics and its Applications, 385(1), 255–260.
Richardson, M.J., & Gerstner, W. (2005). Synaptic shot noise and conductance fluctuations affect the membrane voltage with equal significance. Neural Computation, 17(4), 923– 947.
Risken, H. (1984). Fokker-Planck Equation: Springer.
Rohrer, D., & Wixted, J.T. (1994). An analysis of latency and interresponse time in free recall. Memory & Cognition, 22(5), 511–524.
Romani, S., Pinkoviezky, I., Rubin, A., & Tsodyks, M. (2013). Scaling laws of associative memory retrieval. Neural Computation, 25(10), 2523–2544.
Russo, E., & Treves, A. (2012). Cortical free-association dynamics: Distinct phases of a latching network. Physical Review E, 85(5), 051920.
Schwarz, G., & et al. (1978). Estimating the dimension of a model. The Annals of Statistics, 6(2), 461–464.
Sederberg, P.B., Miller, J.F., Howard, M.W., & Kahana, M.J (2010). The temporal contiguity effect predicts episodic memory performance. Memory & Cognition, 38(6), 689–699.
Shadlen, M.N., & Newsome, W.T. (1998). The variable discharge of cortical neurons: implications for connectivity, computation, and information coding. The Journal of Neuroscience, 18(10), 3870–3896.
Solway, A., Diuk, C., Córdova, N., Yee, D., Barto, A., Niv, Y., & et al. (2014). Optimal behavioral hierarchy. PLoS Computational Biology, e1003779(8).
Sompolinsky, H., & Kanter, I. (1986). Temporal association in asymmetric neural networks. Physical Review Letters, 57(22), 2861.
Sporns, O. (2011). Networks of the Brain: MIT press.
Stein, R.B. (1967). Some models of neuronal variability. Biophysical Journal, 7(1), 37.
Stern, M., Sompolinsky, H., & Abbott, L. (2014). Dynamics of random neural networks with bistable units. Physical Review E, 90(6), 062710.
Steyvers, M., & Tenenbaum, J.B. (2005). The large-scale structure of semantic networks: statistical analyses and a model of semantic growth. Cognitive Science, 29(1), 41–78.
Takeuchi, D., Hirabayashi, T., Tamura, K., & Miyashita, Y. (2011). Reversal of interlaminar signal between sensory and memory processing in monkey temporal cortex. Science, 331(6023), 1443–1447.
Troyer, A.K., Moscovitch, M., & Winocur, G (1997). Clustering and switching as two components of verbal fluency: evidence from younger and older healthy adults. Neuropsychology, 11(1), 138.
Troyer, A. K., Moscovitch, M., Winocur, G., Leach, L., & Freedman, M. (1998). Clustering and switching on verbal fluency tests in alzheimer’s and parkinson’s disease. Journal of the International Neuropsychological Society, 4(2), 137–143.
Tuckwell, H.C. (2005). Introduction to theoretical neurobiology: Volume 2, nonlinear and stochastic theories: Cambridge University Press.
Uhlenbeck, G.E., & Ornstein, L.S. (1930). On the theory of the brownian motion. Physical Review, 36(5), 823.
van der Meer, M.A., Johnson, A., Schmitzer-Torbert, N.C., & Redish, A.D. (2010). Triple dissociation of information processing in dorsal striatum, ventral striatum, and hippocampus on a learned spatial decision task. Neuron, 67(1), 25–32.
van Vreeswijk, C., & Sompolinsky, H. (1996). Chaos in neuronal networks with balanced excitatory and inhibitory activity. Science, 274(5293), 1724–1726.
Verechtchaguina, T., Sokolov, I.M., & Schimansky-Geier, L. (2006). First passage time densities in resonate-and-fire models. Physical Review E, 73(3), 031108.
Viswanathan, G., Buldyrev, S. V., Havlin, S., Da Luz, M., Raposo, E., & Stanley, H.E (1999). Optimizing the success of random searches. Nature, 401(6756), 911– 914.
Voss, R.F. (1992). Evolution of long-range fractal correlations and 1/f noise in DNA base sequences. Physical Review Letters, 68(25), 3805.
Wang, X.-J. (1999). Synaptic basis of cortical persistent activity: the importance of nmda receptors to working memory. The Journal of Neuroscience, 19(21), 9587–9603.
Wilson, H.R., & Cowan, J.D (1972). Excitatory and inhibitory interactions in localized populations of model neurons. Biophysical Journal, 12(1), 1.
Wixted, J.T., & Rohrer, D. (1994). Analyzing the dynamics of free recall: an integrative review of the empirical literature. Psychonomic Bulletin & Review, 1(1), 89–106.
Xue, M., Atallah, B.V., & Scanziani, M. (2014). Equalizing excitation-inhibition ratios across visual cortical neurons. Nature, 511(7511), 596–600.
Yizhar, O., Fenno, L.E., Prigge, M., Schneider, F., Davidson, T.J., O’Shea, D.J., & et al. (2011). Neocortical excitation/inhibition balance in information processing and social dysfunction. Nature, 477 (7363), 171–178.
Acknowledgments
This work was supported by Australian Research Council (ARC) Discovery Project grants (P.G.).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interest
The authors declare that they have no conflict of interest.
Additional information
Action Editor: Michael Breakspear
Appendices
Appendix A: Mean-field analysis methods
1.1 A.1 Derivation of the mean-field dynamics
In our model, the ensemble of neurons can be divided into several sub-populations, including the excitatory neural modules \(a, b {\dots } h\) and the inhibitory population, within each of which the neurons share the same dynamic properties and, in the statistical sense, the same structure of connections with the rest neurons in the network. In the mean-field approach, it is often the collective behaviour of such neural sub-populations (for instance, the steady-state mean firing rates) instead of the behaviour of individual neurons that is of interest, and the steady-state mean firing rate of any sub-population can be obtained by studying a typical neuron in it as we will show. For this purpose, we simplify our notations in Eqs. (1)–(4) by dropping the subscript i, which gives the dynamic equation of a typical neuron in the sub-population α
where the recurrent current is
Note that we have also dropped the subscript i and superscript α in \(s_{ij}^{\alpha \beta }(t)\), this is done by neglecting the conduction delay in our analysis (\(d_{ij}^{\alpha \beta } \approx 0\)), which is the only dependence of \(s_{ij}^{\alpha \beta }(t)\) on i and α in Eq. (5). The recurrent current I αβ(t) can be expressed as the sum of its mean \(\bar {I}^{\alpha \beta }\) and fluctuation δ I αβ(t). This is done by expressing the gating variable in Eq. (17) as the sum of its mean and fluctuation as \(s_{j}^{\beta }(t) = \bar {s}^{\beta } + \delta {s}_{j}^{\beta }(t)\), which gives
where in the last step we have introduced the average number of connections \(\bar {C}^{\alpha \beta }=\left \langle {\sum }_{j=1}^{N^{\beta }}a^{\alpha \beta } \right \rangle =N^{\beta }\Pr (a^{\alpha \beta } = 1)\). Note that the actual number of connections to different neurons are different, following a binomial distribution (Erdős-Rényi random graph). By only considering the average number of connections \(\bar {C}^{\alpha \beta }\) in the current mean-field approach, the moments of the binomial distribution higher than one are neglected. Substituting Eqs. (19) into Eq. (15) and, after some rearrangement, we have
where the effective leaky conductance is \(\tilde {g}_{L}^{\alpha } = g_{L} + {\sum }_{\beta }\bar {C}^{\alpha \beta }J^{\alpha \beta } \bar {s}^{\beta }\), the effective membrane time constant is \(\tilde {\tau }_{m}^{\alpha } = C/\tilde {g}_{L}^{\alpha }\), and the steady-state membrane potential in the absence of threshold is \(V^{\alpha }_{ss} = (g_{L}V_{L} + {\sum }_{\beta }\bar {C}^{\alpha \beta }J^{\alpha \beta } \bar {s}^{\beta }V^{\beta }_{rev} + I^{\alpha }_{ext}) / \tilde {g}_{L}^{\alpha }\) (Renart et al. 2003). Note that to avoid dealing with the multiplicative dependence of the fluctuation in the recurrent current δ I αβ(t) on the membrane potential V α(t), the membrane potential V α(t) in Eq. (19) is approximated by its mean value \(\bar {V}^{\alpha }\) (Renart et al. 2003), that is,
The mean membrane potential \(\bar {V}^{\alpha }\) is different from the steady-state membrane potential in the absence of threshold \(V^{\alpha }_{ss}\) due to the mechanisms of resetting membrane potential to V r t for an absolute refractory period τ r e f immediately after firing in the integrate-and-fire neuron model. An approximate expression of \(\bar {V}^{\alpha }\) is Renart et al. (2003)
where v α is the average firing rate of the sub-population α.
To simplify the dynamics of the gating variable described by Eq. (5), we neglect the conduction delay \( \left (d_{ij}^{\alpha \beta } \approx 0\right )\) and the saturation contributed by two consecutive spikes \(t^{\beta }_{j,k}\) and \(t^{\beta }_{j,k+1}\), since the firing rate of interest is much lower than \(\left (\tau _{d}^{\beta }\right )^{-1} \approx 200\) Hz. And we further approximate the time-course of the neurotransmitter concentration as a delta function, that is
where A β is a normalization factor to keep \({\int }_{-\infty }^{+\infty } \hat {s}_{j}^{\beta }(t) dt\) unchanged under the simplification Eq. (24) and \(\hat {s}_{j}^{\beta }(t)\) is the time course of the gating variable \(s_{j}^{\beta }(t)\) in response to a single spike coming at t = 0 with the initial condition \(s_{j}^{\beta }(-\infty ) = 0\). This normalization factor is given by,
Substituting these simplifications into Eq. (5) gives
Now we want to cast the above dynamic equation of gating variable into a dynamic equation of the fluctuations in recurrent current \(\delta \tilde {I}^{\alpha }(t)\), because it is \(\delta \tilde {I}^{\alpha }(t)\), rather than the gating variable \(s^{\beta }_{j}(t)\), that directly affects the membrane potential as shown in Eq. (1). For this purpose, we substitute \(s_{j}^{\beta }(t) = \bar {s}^{\beta } + \delta {s}_{j}^{\beta }(t)\) into Eq. (26), multiply both sides by \(\tau _{d}^{\beta }a^{\alpha \beta }J^{\alpha \beta }\) and sum over j; this yields
To simply Eq. (27), firstly note that the number of synapses into one neuron is large (\(\bar {C}^{\alpha \beta } \gg 1\)) and that the individual synaptic couplings are weak, by which we mean the peak amplitude of an unitary EPSP or IPSP due to recurrent input (<1m V) is much less than what is required to excite the post-synaptic neuron from resting potential to fire (20m V). With the above two conditions, we can apply the diffusion approximation to simplify the dynamics with an assumption that the presynaptic neurons in the sub-population β fire approximately as independent Poisson processes (Stein 1967; Tuckwell 2005; Richardson and Gerstner 2005). The diffusion approximation simplifies the last term on the right hand side of Eq. (27) into Gaussian white noise,
where η(t)β is zero-mean Gaussian white noise with unit variance. Substituting the diffusion approximation Eqs. (28) into Eq. (27) gives,
where letting the first term on the right hand side be zero gives a self-consistent solution for the stationary mean of the gating variable, \(\bar {s}^{\beta } = A^{\beta }v^{\beta }\tau _{d}^{\beta }\).
Multiplying both sides of Eq. (29) by \(-\left (\bar {V}^{\alpha }-V_{rev}^{\beta }\right )\) gives the dynamic equation governing the fluctuation of each individual recurrent current \(\delta \tilde {I}^{\alpha \beta }\)
where \(\sigma _{I}^{\alpha \beta } = \tau _{d}^{\beta }J^{\alpha \beta }A^{\beta }\sqrt {\bar {C}^{\alpha \beta }v^{\beta }}|\bar {V}^{\alpha }-V_{rev}^{\beta }|\) and any extra minus sign is absorbed into the Gaussian white noise η β(t). Note that Eq. (30) describes an Ornstein-Uhlenbeck (O-U) process (Uhlenbeck and Ornstein 1930) for the fluctuations in each recurrent current component \( \delta \tilde {I}^{\alpha \beta }(t)\). The last step necessary for obtaining the dynamic equation governing the fluctuation in total recurrent current \( \delta \tilde {I}^{\alpha }(t) ={\sum }_{\beta } \delta \tilde {I}^{\alpha \beta }(t)\) is to introduce an effective synaptic time constant \(\tilde {\tau }_{d}^{\alpha }\) to replace \(\tau _{d}^{\beta }\) in Eq. (30) (Fourcaud and Brunel 2002) so that we can sum Eq. (30) over β as
where in the last step we have used the fact that the sum of several independent zero mean Gaussian processes is still a zero mean Gaussian process, that is, \({\sum }_{\beta } \sigma _{I}^{\alpha \beta } \eta ^{\beta }(t) = \sigma _{I}^{\alpha } \eta ^{\alpha }(t)\) with a new variance \(\left (\sigma _{I}^{\alpha }\right )^{2} = {\sum }_{\beta } \left (\sigma _{I}^{\alpha \beta }\right )^{2}\). The effective synaptic time constant \(\tilde {\tau }_{d}^{\alpha }\) is chosen to preserve the stationary variance of \(\delta \tilde {I}^{\alpha }\) (Fourcaud and Brunel 2002). Note that the stationary solution of the O-U process described by Eq. (30) is a zero-mean Gaussian process with a variance of \(\left (\sigma _{I}^{\alpha \beta }\right )^{2}/\left (2\tau _{d}^{\beta }\right )\) (Uhlenbeck and Ornstein 1930). Therefore, preserving the total stationary variance summed over the sub-populations in the effective O-U process of Eq. (32) gives the equation for \(\tilde {\tau }_{d}^{\alpha }\)
The effective synaptic time constant approximation holds valid when the difference among the individual time constants and variances are within an order of magnitude (Fourcaud and Brunel 2002), which is the case in our model.
1.2 A.2 Numerical solution to the stationary mean firing rate
For a typical neuron in the sub-population α, Eq. (20) and (32) together describe a bi-variate Markov process (Gillespie 1991; Risken 1984). To obtain the mean firing rate v α of the sub-population α is equivalent to find the expected firing rate of this Markov process. In the case where \(k = \sqrt {\tilde {\tau }_{d}^{\alpha }/\tilde {\tau }_{m}^{\alpha }} \ll 1\), an analytical solution to the expected firing rate of the Markov process can be obtained via perturbation theory to the first order of k (Fourcaud and Brunel 2002). However, when k is comparable to 1, which is quite often the case in the conductance-based model, no analytical solution has been found, although extrapolation to the second order in k and extrapolation between small and large k limits have been attempted in similar models (Brunel and Sergi 1998; Brunel and Latham 2003). Therefore, to obtain the accurate mean firing rate when k≈1, we resort to numerical solution via efficient direct simulation of Eq. (20) and (32). Firstly, note that these two equations can be readily casted into a single stochastic harmonic oscillator by differentiating Eq. (20) once and eliminating the variable \(\delta \tilde {I}^{\alpha }\), which gives
where \( \hat {V}^{\alpha } = V^{\alpha } - V_{ss}^{\alpha }\). An exact simulation scheme has been developed for such harmonic oscillator driven by Gaussian white noise (Nørrelykke and Flyvbjerg 2011). After removing the numerical singularity, the simulation scheme can be applied to Eq. (34) to obtain the mean firing rates (see Appendix B). This numerical solution essentially provides a stochastic non-linear input-output relation v o u t = Φ(v i n ), where v i n is the vector of the input firing rates to the neuron sub-populations and v o u t is the resultant vector of the output firing rates. This is conceptually equivalent to the input-output relations used in firing-rate models, such as the simple sigmoid function (Wilson and Cowan 1972). To solve the stationary mean firing rates of the sub-populations in the recurrent network is to find the roots of the self-consistent equation v = v i n = v o u t , or, equivalently,
where v = [v α]. Solving the above self-consistent equation can be casted as a stochastic optimization problem with an objective function |Φ(v)−v| to be minimized, which can be readily solved by stochastic optimization algorithms such as the pattern search algorithm (Hooke and Jeeves 1961). Starting from different initial values for the optimization algorithm, different local minimums are found and the residue errors are calculated, which is defined as |Φ(v)−v|/|v|; only the local minimums with a residue error less than 20 % are accepted as solutions to Eq. (35). The solutions are summarized in Table 2.
Approximate non-stationary firing rate dynamics of the neuron sub-populations are commonly used for stability analysis in current based integrate-and-fire neuron circuit models to bypass the difficulty of analyzing the exact non-stationary dynamics (Amit and Brunel 1997; Mascaro and Amit 1999; Renart et al. 2003). However, in our conductance based model, the time-dependency of the system parameters \(\tilde {\tau }_{m}^{\alpha }\), \(\tilde {\tau }_{d}^{\alpha } \) and \(\tilde {g}^{\alpha }_{L}\) in Eq. (20) renders such approximate method invalid, because the approximate dynamics assume that such system parameters are constant. Nonetheless, without stability analysis, the above mean-field results for fixed points, when compared with results from simulating the full spiking neural circuit model, still provide valuable insights into the dynamics of the latter.
Qualitatively, from the perspective of attractor dynamics, the IRIs may be understood as the escape times of a “particle” in a simplified multi-well potential being driven by noise, as suggested by Litwin-Kumar and Doiron (2012). Based on Kramer’s classic reaction-rate theory (Hänggi et al. 1990), the time it takes for the particle to escape from one well into another is mainly controlled by the statistical properties of the noise and the depths of the two wells. As shown by the mean-field analysis (Eqs. (30) and (33)), the noise can be approximated as Gaussian white noise, and its intensity depends on several factors, including the synaptic time constants, coupling strengths, presynaptic firing rates, number of synapses (connection probabilities and network size) and the mean membrane potential. And the potential well depths may also depend on such a broad range of system parameters. Especially, among all the parameters, the coupling strengths and hierarchical connection probabilities play an important role in determining the IRI timescales, as demonstrated in Figs. 6 and 7, where the switching/transition frequencies are, by definition, inversely correlated with the IRI timescale. However, Kramer’s reaction-rate theory predicts an exponential tail of the escape time distribution in the low noise intensity limit (Hänggi et al. 1990), which is qualitatively different from the log-normal distributions observed in our model as well as in a recent firing-rate network model (Stern et al. 2014), and it remains an open problem to explain the generative mechanisms of such a log-normal distribution.
The observation that the transition probability decreases while the transition duration (IRI) increases as the connection probability decreases (Fig. 3) might also be explained from the perspective of attractor dynamics. As strongly indicated by the simulation results in Fig. 2 and the mean-field analytical result in Table 2, a transition from one clustered firing state to another (mean-field states of the first kind) is achieved by the system via going through one or multiple intermediate disordered states (mean-field states of the second kind). For example, in Fig. 2b, the level-3 transition happening between 302 and 305 seconds consists of obviously more intermediate disordered states than the previous transitions at lower levels. This suggests that, when two excitatory neural modules are connected with a smaller probability (i.e., at a higher hierarchy level), their corresponding clustered firing states may be further apart in the attractor landscape, and as a result the transition between them will require a longer duration to go through the intermediate attractors and have a smaller probability to happen.
Appendix B: Numerical methods for simulating stochastic harmonic oscillator
Here we describe in details the method used to simulate the stochastic process described by Eq. (34) in the main text, which is modified from the methods in Nørrelykke and Flyvbjerg (2011) to remove the numerical singularity at k = 1 in the original expressions. Firstly, we introduce the definitions of the parameters that will be used. The cyclic frequency of the undamped oscillator is \(\omega ^{\alpha }_{0}=1/\sqrt {\tilde {\tau }_{d}^{\alpha }\tilde {\tau }_{m}^{\alpha }}\) and the characteristic time τ α of the exponential decrease with time that the momentum of the “particle” undergoes in the absence of all but friction forces is
The diffusion constant is defined as
And the cyclic frequency of the damped oscillator is \(\omega ^{\alpha } = \sqrt {(\omega ^{\alpha }_{0})^{2}-1/(2\tau ^{\alpha })^{2}}\), which is real for under-damped oscillator. However, substituting Eq. (36) into the expression shows that ω α is zero when \(\tilde {\tau }_{d}^{\alpha } = \tilde {\tau }_{m}^{\alpha }\) and always imaginary otherwise, indicating an over-damped oscillator. For mathematical convenience, we keep the imaginary frequency ω α. Note that the difficulty in obtaining analytical solution to the mean first passage time of the bi-variate Markov process described by Eqs. (20) and (32) in the main text is consistent with the difficulty in solving it for an over-damped stochastic harmonic oscillator (Verechtchaguina et al. 2006).
With the above definitions, together with suitable time step Δt, t j = jΔt, \(\hat {V}^{\alpha }_{j} = \hat {V}^{\alpha }(t_{j})\), \(d\hat {V}^{\alpha }/dt = Y^{\alpha }\), and \(Y^{\alpha }_{j} = Y^{\alpha }(t_{j})\), the recursive relation for simulation is given by
where
where \({\xi }_{j}^{\alpha }\) and \({\zeta }_{j}^{\alpha }\) are independent zero-mean Gaussian white noise with unit variance, and \(\text {sinc}(x) = \sin (x) / x\) is known as the unnormalized cardinal sine function. Note that the above simulation scheme is exact for time steps of arbitrary size (Nørrelykke and Flyvbjerg 2011). In practice, a time step of Δt = 1 ms is chosen for a balance between performance and temporal resolution. Analytical solution to special cases when k≪1 (Fourcaud and Brunel 2002) is used for code testing.
With the above simulation scheme, the mean first passage time of the dynamic system described by Eq. (34) in the main text can be obtained with the following procedure. Firstly the initial value of the membrane potential is set to be at the reset potential, or, equivalently, \({\hat {V}}_{0}^{\alpha } = V_{rt} - {V}_{ss}^{\alpha }\). And the initial value \({Y}_{0}^{\alpha }\) is self-consistently chosen from the stationary solution of \({Y}_{t}^{\alpha }\), which is a zero-mean Gaussian distribution with the variance given by Chandrasekhar (1943)
Then, for each trial, the time ΔT α it takes for the system to reach the threshold \({\hat {V}}_{th}^{\alpha } = V_{th} - {V}_{ss}^{\alpha }\) for the first time is recorded (first passage time). Finally, the sum of the refractory period τ r e f and the mean first passage time averaged over trials 〈ΔT α〉 gives an estimate on the inverse of the mean firing rate, that is
In practice, 5000 trails are used for each estimation.
Appendix C: Bayesian model comparison
Here we describe the detailed procedures of the Bayesian comparison between the ex-Gaussian model and the log-normal model of the IRI distributions. Firstly, the distribution of IRIs data is extracted from Fig. 1 in study by Rohrer and Wixted (1994) and the data is in the form of a histogram. We denote the bin width of the histogram as ΔB and the i th bin centre as B i = B 1+(i−1)ΔB, where \(i = 1, 2 {\dots } M\) and M is the total number of bins. And we denote the number of entries (IRIs) in the i th bin as d i . In the data, ΔB = 1 second, B 1 = 0.5 second, M = 20 and the sample size \(N = {\sum }_{i = 1}^{M} d_{i} = 1186\).
The two models for the distribution of IRIs to be compared are the log-normal distribution and the ex-Gaussian distribution. The probability density function (PDF) for a log-normal distribution is
where μ 1 and σ 1 are the distribution parameters. The PDF for a ex-Gaussian distribution is
where λ 2, μ 2 and σ 2 are the distribution parameters and \(H(x) = 2/\sqrt {\pi } {\int }^{+\infty }_{x} e^{- t^{2}} dt\), which is the complementary error function. For convenience, we refer to the two PDF models as f j (x;𝜃 j ) or f j (𝜃 j ), where j = 1,2 is the model index and 𝜃 1 = (μ 1, σ 1) and 𝜃 2 = (λ 2, μ 2, σ 2) are the model parameter vectors.
Given the sample size N, i.e. the total number of entries in all of the bins, the expected number of entries t i, j in the i th bin given by the j th model is
where p i, j is the probability for an observed IRI to be within the i th bin given by the j th model, which can be obtained by
Note that, in the above expression, the term on the denominator renormalizes the PDF over the range of bins, that is, the distribution is truncated after the last bin. With enough number of bins and a sample size large enough, the observed number of entries in the i th bin should follow a Poisson distribution with the parameter t i, j . Therefore, the likelihood for t i, j expected and d i actually observed entry in the i th bin given by the j th model is \(L_{i,j}(\boldsymbol {\theta }_{j}) = e^{-t_{i,j}}t_{i,j}^{d_{i}}/d_{i} !\). Summing this over all the bins, taking the logarithm and discarding model independent terms give the log-likelihood of the observed histogram under the j th model with parameter vector 𝜃 j
With this expression, we can readily use the Schwarz criterion (Schwarz et al. 1978) to approximate the Bayes factor B 12, that is
where \(\hat {\boldsymbol {\theta }}_{1}\) and \(\hat {\boldsymbol {\theta }}_{2}\) are the maximum likelihood estimators of the model parameter vectors that maximize the log-likelihood of the observed histogram as in Eq. (49), N is the sample size and k 1 and k 2 are the dimensions of the model parameter vectors 𝜃 1 and 𝜃 2, respectively. This approximation should provide a reasonable indication for the model comparison in large samples (N≫k 1, k 2) (Kass and Raftery 1995), and in our case N is around 103 while k 1 = 2, k 2 = 3.
Rights and permissions
About this article
Cite this article
Gu, Y., Gong, P. The dynamics of memory retrieval in hierarchical networks. J Comput Neurosci 40, 247–268 (2016). https://doi.org/10.1007/s10827-016-0595-7
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10827-016-0595-7