Volterra dendritic stimulus processors and

2 downloads 0 Views 5MB Size Report
Sep 1, 2014 - noise sources affect the precision in encoding and decoding of sensory stimuli and the functional ... to stochastic transduction processes (Song et al., 2012), and ion channel ...... a divide and conquer strategy is ubiquitous in engineering circuits and leads to ..... The effect of contrast on the transfer properties ...
ORIGINAL RESEARCH ARTICLE published: 01 September 2014 doi: 10.3389/fncom.2014.00095

COMPUTATIONAL NEUROSCIENCE

Volterra dendritic stimulus processors and biophysical spike generators with intrinsic noise sources Aurel A. Lazar* † and Yiyin Zhou † Department of Electrical Engineering, Columbia University, New York, NY, USA

Edited by: Mark D. McDonnell, University of South Australia, Australia Reviewed by: Ron Meir, Technion, Israel Zoran Tiganj, Boston University, USA *Correspondence: Aurel A. Lazar, Department of Electrical Engineering, Columbia University, 500 W. 120th Street, New York, NY 10027, USA e-mail: [email protected] † The authors’ names are listed in alphabetical order.

We consider a class of neural circuit models with internal noise sources arising in sensory systems. The basic neuron model in these circuits consists of a dendritic stimulus processor (DSP) cascaded with a biophysical spike generator (BSG). The dendritic stimulus processor is modeled as a set of nonlinear operators that are assumed to have a Volterra series representation. Biophysical point neuron models, such as the Hodgkin-Huxley neuron, are used to model the spike generator. We address the question of how intrinsic noise sources affect the precision in encoding and decoding of sensory stimuli and the functional identification of its sensory circuits. We investigate two intrinsic noise sources arising (i) in the active dendritic trees underlying the DSPs, and (ii) in the ion channels of the BSGs. Noise in dendritic stimulus processing arises from a combined effect of variability in synaptic transmission and dendritic interactions. Channel noise arises in the BSGs due to the fluctuation of the number of the active ion channels. Using a stochastic differential equations formalism we show that encoding with a neuron model consisting of a nonlinear DSP cascaded with a BSG with intrinsic noise sources can be treated as generalized sampling with noisy measurements. For single-input multi-output neural circuit models with feedforward, feedback and cross-feedback DSPs cascaded with BSGs we theoretically analyze the effect of noise sources on stimulus decoding. Building on a key duality property, the effect of noise parameters on the precision of the functional identification of the complete neural circuit with DSP/BSG neuron models is given. We demonstrate through extensive simulations the effects of noise on encoding stimuli with circuits that include neuron models that are akin to those commonly seen in sensory systems, e.g., complex cells in V1. Keywords: Volterra dendritic stimulus processors, biophysical spike generators, noise, neural encoding, neural decoding, functional identification, Hodgkin-Huxley neuron, phase response curve

1. INTRODUCTION Intrinsic noise sources are diverse and appear on many levels of a neuronal system ranging from electrical to chemical noise sources (Faisal et al., 2008; Destexhe and Rudolph-Lilith, 2012) and from single cells to networks of neurons. At the cellular and subcellular level, variability in biochemical reactions leads to stochastic transduction processes (Song et al., 2012), and ion channel fluctuations (Neher and Sakmann, 1976; White et al., 1998) result in variability in spike generation and propagation (Faisal and Laughlin, 2007). At the network level, probabilistic quantal release of neurotransmitters (Katz, 1962), background synaptic activity (Destexhe et al., 2003; Jocobson et al., 2005) and variability in timing of spikes from presynaptic neurons (Faisal and Neishabouri, 2014) are sources of stochastic fluctuation of synaptic conductances (Destexhe et al., 2001) that are believed to have a major impact on spike time variability (Yarom and Hounsgaard, 2011). The existence of sources of noise also leads to variability in the spike times even when neurons are subject to the same, repeated inputs (Calvin and Stevens, 1968; Berry et al., 1997; de Ruyter van Steveninck et al., 1997). Spikes are the primary form of carriers of

Frontiers in Computational Neuroscience

information in the nervous system and their timing is thought to be relevant to the message neurons need to convey (Rieke et al., 1999). Therefore, the variability of spike timing may reduce or damage the information being transmitted. It is quite remarkable, however, that sensory systems manage to be very robust even if they are subject to interference due to noise. Visual and auditory systems are two examples in which the stimuli are highly time varying. These systems have been reported to convey information with high spike timing precision (Butts et al., 2007; Kayser et al., 2010). Noise may be useful in facilitating signal detection (McDonnell and Ward, 2011). Still, interference due to noise poses an important limit on how well sensory systems can represent input stimuli. It is not clear how intrinsic noise sources affect the representation of sensory inputs based on spike times, and how they impact the functional identification of sensory neurons. We study the representation of sensory stimuli using a novel neural circuit model, that extends previously proposed models (Lazar et al., 2010; Lazar and Slutskiy, 2014, in press) in terms of architectural complexity and the existence of intrinsic noise sources. Our base level circuit architecture consists of

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 1

Lazar and Zhou

Neural circuits with intrinsic noise sources

two interconnected neurons, each with two cascaded stages. The first stage comprises two types of dendritic stimulus processors. The first dendritic stimulus processor performs nonlinear processing of input stimuli in the feedforward path leading to the spike generator. The second dendritic stimulus processor performs nonlinear processing in the feedback loop whose inputs are spike trains generated by biophysical spike generators (BSGs). The BSGs constitute the second stage of the base level circuit. Our nonlinear dendritic stimulus processors describe functional I/O relationships between the dendritic outputs in the first stage and inputs that are either sensory stimuli or spikes generated by BSGs. DSPs are modeled using Volterra series. Volterra series have been used for analyzing nonlinear neuronal responses in many contexts (Lu et al., 2011; Eikenberry and Marmarelis, 2012), and have been applied to the identification of single neurons in many of sensory areas (Benardete and Kaplan, 1997; Theunissen et al., 2000; Clark et al., 2011). Volterra dendritic processors can model a wide range of nonlinear effects commonly seen in sensory systems (Lazar and Slutskiy, in press). Here, in addition, we introduce nonlinear interactions between neurons in the feedback and cross-feedback paths. This gives rise to interesting neural processing capabilities directly in the spike domain, e.g., coincidence detection (Agmon-Snir et al., 1998; Stuart and Häusser, 2001). The relationships described here by the Volterra model are functional and do not address the underlying circuit/dendritic tree level interactions. However, the latter have recently been subject to intense investigations (London and Häusser, 2005; Wohrer and Kornprobst, 2009; Werblin, 2011; Xu et al., 2012; Yonehara et al., 2013; Zhang et al., 2013). Conductance-based, biophysical spike generators are well established models that have been extensively used in studies of neuronal excitability and in large simulations of spiking neural networks (Izhikevich, 2007). Following Lazar (2010), we use formal BSG models to represent sensory stimuli under noisy conditions. We formulate the encoding, decoding and functional identification problems under the neural encoding framework of Time Encoding Machines (TEMs). In this modeling framework the exact timing of spikes is considered to carry information about input stimuli (Lazar and Tóth, 2004). The separation into dendritic stimulus processors and spike mechanisms mentioned above allows us to study synaptic inputs and spike generation mechanisms separately, and hence independently model the intrinsic noise sources of each component. We incorporate two important noise sources into a general single-input multi-output neural circuit model. The first is a channel noise source that arises in spike generation (White et al., 2000). The second is a synaptic noise source due to a variety of fluctuating synaptic currents (Manwani and Koch, 1999). Based on the rigorous formalism of TEMs, we show how noise arising in dendritic stimulus processors and in biophysical spike generators is related to the measurement error in generalized sampling. Dendritic stimulus processing and spike generation can then be viewed as a generalized sampling scheme that neurons utilize to represent sensory inputs (Lazar et al., 2010). Contrary to traditional sampling where the signal amplitude is sampled at clock times, neurons asynchronously sample all stimuli.

Frontiers in Computational Neuroscience

We systematically investigate how the strength of noise sources degrades the faithfulness of stimulus representation and the quality of functional identification of our proposed class of neural circuits. Furthermore, since the representation is based on spike timing, it is natural to investigate how spike timing variability affects the precision in representing the amplitude information of sensory stimuli. The work presented here requires a substantial amount of investment in the mathematical formalism employed throughout. There are a number of benefits in doing so, however. Formulating the problem of stimulus encoding with a neural circuit with intrinsic noise sources as one of generalized sampling, i.e., of taking noisy measurements is of interest to both experimentalists and theoreticians alike. Understanding that the problem of neural decoding and functional identification are dual to each other is key to building on either or both. Finding how many repeat experiments need to be performed for a precise quantitative identification of Volterra kernels is of great value in neurophysiology. A further qualitative insight of our work is that for neural circuits with arbitrary connectivity, feedforward kernels are typically easier to estimate than feedback kernels. Finally, our finding that some key nonlinear neural circuits are tractable for detailed noise analysis suggests a wide reaching analytical methodology.

2. MODELING NONLINEAR NEURAL CIRCUITS, STIMULI, AND NOISE We present in Section 2.1 the general architecture of the neural circuits considered in this paper. In Section 2.2 we discuss the modeling of the space of stimuli. Volterra DSPs are the object of Section 2.3. Finally, in Section 2.4 we provide models of BSGs with intrinsic noise sources. 2.1. NEURAL CIRCUIT ARCHITECTURE

The general architecture of the neural circuit considered here is shown in simplified form in Figure 1. It consists of two neurons with a common time-varying input stimulus. With added notational complexity the neural circuit in Figure 1 can easily be extended in two ways. First, multiples of such circuits can encode a stimulus in parallel (see Section 2.1 in the Supplementary Material). In this case only pairs of neurons are interconnected through the feedback kernels. Second, more neurons can be considered in the neural circuit of Figure 1; all these neurons can be fully interconnected through feedback loops. Each neuron i, i = 1, 2, receives a single time-varying input stimulus u1 (t). The modeling of the input stimulus is discussed in Section 2.2. The output of each of the biophysical spike generators (BSGs) is a spike sequence denoted by (tk1 ) and (tl2 ), k, l ∈ Z. The input stimulus u1 (t) is first processed by a feedforward Dendritic Stimulus Processor (feedforward DSP) (Lazar and Slutskiy, in press). The feedforward DSP models the aggregated effect of processing in the neural circuits in the prior stages and in the dendritic tree of neuron i = 1, 2. For example, if the neurons in the model circuit are considered to be Retinal Ganglion Cells (RGCs), then the feedforward Volterra DSP models the processing that takes place in the outer- and inner-plexiform layers of the retina as well as in the dendritic trees of an RGC (Werblin, 2011; Masland, 2012). The feedforward

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 2

Lazar and Zhou

Neural circuits with intrinsic noise sources

FIGURE 1 | Diagram of the architecture of the neural circuits.

DSPs are modeled here as second order Volterra expansion terms (Volterra, 1930). The first order terms h11i 1 (t) in the feedforward DSPs are linear filters typically used in modeling receptive fields. The second order terms h11i 2 (t1 , t2 ) model nonlinear operations on the stimulus u1 (t). A second group of Volterra DSPs models the cross-feedback interactions between the two neurons. Instead of time-varying stimuli, the output spikes generated by the BSGs are the inputs to these DSPs. We therefore refer to these as feedback Dendritic Stimulus Processors (feedback DSPs). The output spikes of each individual neuron i are processed by the first order term 2ji h1 (t), i, j = 1, 2, i  = j. In addition, output spikes from both neurons interact nonlinearly through the second order terms 2ji h2 (t1 , t2 ), i, j = 1, 2, i  = j. The summed responses from the 2ji first order feedback DSP h1 and the second order feedback DSP 2ji h2 are fed back to neuron i as additional dendritic currents. The dendritic currents consisting of the output of the DSPs with added noise are subsequently encoded by biophysical spike generators. BSGs are biophysically realistic axon hillock spike generator models that are governed by a set of differential equations with multiple types of ion channels (Hodgkin and Huxley, 1952; Izhikevich, 2007). The detailed BSG models are introduced in Section 2.4. The spike times of output spikes generated by the BSGs are assumed to be observable. We identify two intrinsic noise sources of the proposed neural circuit. First, the feedforward DSPs and the feedback DSPs are affected by additive Gaussian white noise. This noise arises from the combined effect along the path from sensory transduction to synaptic integration and includes synaptic background noise and stochasticity in the dendritic tree (Manwani and Koch, 1999;

Frontiers in Computational Neuroscience

Fellous et al., 2003; Destexhe and Rudolph-Lilith, 2012). Since the outputs of the feedforward and feedback DSPs are additively combined, we consider, for simplicity, a single source of additive Gaussian white noise. Second, the ion channels of the BSGs are intrinsically stochastic and introduce noise in the spike generators (White et al., 2000; Hille, 2001). 2.2. MODELING SIGNAL SPACES

Two signal spaces will be considered here. The first, models the space of input signals to feedforward DSPs. The second models the space of input spikes to feedback DSPs. These spaces will be formally described below. 2.2.1. Modeling the space of input stimuli

We model the space of input stimuli as a Reproducing Kernel Hilbert Space (RKHS) (Berlinet and Thomas-Agnan, 2004). RKHSs are versatile vector spaces for modeling signals arising in computational neuroscience, signal processing and machine learning. For example, auditory signals, olfactory signals and visual signals can readily be modeled as band-limited functions of an RKHS with a sinc or Dirichlet kernel (Lazar et al., 2010; Lazar and Slutskiy, 2013). A particular choice of RKHSs in this article is the space of trigonometric polynomials. The computational advantage of working on the space of trignometric polynomials has been discussed (Lazar et al., 2010) and is closely related to the algorithmic tractability of the Fourier series in the digital domain. If the biological signals have unknown bandwidth with a spectrum that falls off fast enough, many Sobolev spaces might be a suitable choice of RKHS (Berlinet and ThomasAgnan, 2004; Lazar and Pnevmatikakis, 2009). In such spaces the norm may include the derivative of the signal, i.e., the rate of

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 3

Lazar and Zhou

Neural circuits with intrinsic noise sources

change of the signal that many neurons are sensitive to Kim et al. (2011). The space of trigonometric polynomials is defined as below. Definition 2.1. The space of trigonometric polynomials H11 is a function space whose elements are functions defined on the domain D1 = [0, S1 ], S1 ∈ R+ , of the form L1 

u1 (t) =

where el1 l2 (t1 , t2 ) =

1 j l1 11 t1 j l2 11 t2 e L e L , S1

(7)

are a set of functions forming an orthonormal basis.

H21 is again an RKHS with RK ul el (t),

(1)

l = −L1

1

K21 (t1 , t2 ; s1 , s2 )

L 

=

1

L 

el1 l2 (t1 − s1 , t2 − s2 ).

(8)

l1 = −L1 l2 = −L1

where 1 jl 1 t el (t) = √ e L1 , l = −L1 , · · · , L1 , S1

(2)

are a set of orthonormal basis functions. 1 denotes the bandwidth and L1 is the order of the space.

H11 endowed with the inner product:  u1 , v1  =

D1

u1 (t)v1 (t)dt

(3)

is a Hilbert Space. Intuitively, the basis functions el (t), l = −L1 , . . . , L1 , can be interpreted as a set of discrete spectral lines uniformly spaced in the frequency domain between −1 and 1 . For a given signal u1 (t), the amplitude of its spectral lines is determined by the coefficients ul , l = −L1 , . . . , L1 . Remark 2.2. Functions in H11 are periodic over R with period 1 S1 = 2πL . Therefore, the domain D1 covers exactly one period of 1 the function. Note that the ul ’s are closely related to the Fourier coefficients of the periodic signal u1 (t), and can thereby be very efficiently computed via the Fast Fourier Transform.

H11 is an RKHS with reproducing kernel (RK) K11 (t; s) =

el (t − s).

(4)

l = −L1

It can be easily verified that the RK satisfies the reproducing property u1 ( · ), K11 (t; ·) = u1 (t), ∀u1 ∈ H11 , t ∈ D1 .

(5)

Definition 2.3. We shall also consider the tensor product space H21 on the domain D2 = [0, S1 ] × [0, S1 ], whose elements are of the form 1

u2 (t1 , t2 ) =

L 

1

L 

ul1 l2 el1 l2 (t1 , t2 ),

l1 = −L1 l2 = −L1

Frontiers in Computational Neuroscience

Definition 2.4. Let h1 ∈ L1 (D1 ), where L1 denotes the space of Lebesgue integrable functions. The operator P 1 : L1 (D1 ) → H11 given by  1 (P h1 )(t) = h1 (s)K11 (t; s)ds, D1

is called the projection operator from L1 (D1 ) to H11 . Similarly, let h2 (t1 , t2 ) ∈ L1 (D2 ), the operator P 1 : L1 (D2 ) → H21 (by abuse of notation) given by  (P h2 )(t1 , t2 ) = 1

D2

h2 (s1 , s2 )K21 (t1 , t2 ; s1 , s2 )ds1 ds2 ,

is called the projection operator from L1 (D2 ) to H21 . 2.2.2. Modeling the space of spikes

1

L 

Note that we use the subscript to indicate the dimension of the domain of functions, i.e., the number of variables the functions in the RKHS have, and use the superscript 1 to indicate the input space. Projections of functions onto the RKHSs introduced here can be defined as follows:

(6)

The feedback kernels of the neural circuit in Figure 1 receive as inputs spike trains generated by the BSGs. Spike trains are often modeled as sequences of Dirac delta pulses and, consequently, the outputs of linear feedback kernels are the result of superposition of their impulse responses (Keat et al., 2001; Pillow et al., 2008; Lazar et al., 2010). Dirac delta pulses have infinite bandwidth. Spikes generated by the BSGs, however, have limited effective bandwidth. Following (Lazar and Slutskiy, 2014) spikes are modeled to be the RK of an one-dimensional Hilbert space H12 at spike time occurrence. Here H12 is a space of trigonometric polynomials whose order L2 , period S2 and bandwidth 2 may differ from the input stimulus space H11 , where 2 shall be larger than the bandwidth assumed for the feedback kernel, and S2 is much larger than the support of the feedback kernel (Lazar and Slutskiy, 2014). A spike at time tki of neuron i can then be expressed in functional form as K12 (tki ; t), where the superscript indicates that the RK belongs to the spike input space.

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 4

Lazar and Zhou

Neural circuits with intrinsic noise sources

Due to the reproducing property, single or pairs of input spikes have the property    h1 (t) ∗ K12 tki ; t =

     h1 (t − s)K12 tki ; s ds = P 2 h1 t − tki

D1

and  D2

     j j h2 (t − s1 , t − s2 ) K22 tki , tl ; s1 , s2 ds1 ds2 = P 2 h2 t − tki , t − tl

for i, j = 1, 2, i  = j. The operator P 2 is similarly defined to P 1 above; it denotes, however, the projection onto the space of spikes. Thus, not surprisingly, incoming spikes directly readL2 → ∞, out2 the projection of the feedback kernels. By letting 2 P h1 (t − tk ) shall converge to h1 (t − tk ) in L norm as the RK converges to the sinc function and the RKHS becomes the space of band-limited signals (Lazar et al., 2010). A more detailed analysis is available in Lazar and Slutskiy (2014). This formalism will be employed for solving the functional identification problem formulated in Section 4.1.

1996; Dayan and Abbott, 2001). In addition, the two filters are quadrature pair in phase. Both filters are illustrated in Figure 2A. The response of applying the input stimulus u1 on the temporal filters with impulse response gc and gs is given by D1 gc (t − s)u1 (s)ds

and D1 gs (t − s)u1 (s)ds, respectively. The responses of the two linear filters of the complex cell model are squared and summed to produce the phase invariant measure vi (Carandini et al., 2005), where  v i (t) =  =  +

=

2.3.1. Feedforward Volterra dendritic stimulus processors

The feedforward DSPs are modeled as up to second order terms in the Volterra series. The feedforward DSPs take continuous signals in the stimulus space as inputs, while the output can be expressed as (see also Figure 1) 

 D1

h11i 1 (t − s)u1 (s)ds +

D2

h11i 2 (t − s1 , t − s2 ) u1 (s1 )u1 (s2 )ds1 ds2 , (9)

11i 1 1 where h11i 1 ∈ L (D1 ) and h2 ∈ L (D2 ) denote, respectively, the first and second order Volterra kernels, i = 1, 2. They are assumed to be real, causal and bounded-input bounded-output 11i (BIBO)-stable. It is also assumed that both h11i 1 and h2 have finite 11i memory. In addition, h2 is assumed, without loss of generality, 11i to be symmetric, i.e., h11i 2 (t1 , t2 ) = h2 (t2 , t1 ).

Example 2.5. We present here a Volterra DSP that is akin to a model of dendritic stimulus processing of complex cells in the primary visual cortex (V1). The difference is that the complex cells operate spatio-temporally, whereas in the example given below they operate temporally. We first consider two first order kernels based on Gabor functions,    (t − 0.13)2 cos 2π · 10 · (t − 0.13) , gc (t) = exp − 2 · 0.0005    (t − 0.13)2 sin 2π · 10 · (t − 0.13) . gs (t) = exp − 2 · 0.0005 The two filters are Gaussian modulated sinusoids, that are typically used to model receptive fields of simple cells in the primary visual cortex (V1) where the variables denote space instead of time (Lee,

Frontiers in Computational Neuroscience

D2

D2

D2



2.3. VOLTERRA DENDRITIC STIMULUS PROCESSORS

As mentioned in Section 2.1, two forms of dendritic stimulus processing appear in our model.

D1

 =

2

D2

gc (t − s)u1 (s)ds

 +

D1

2 gs (t − s)u1 (s)ds

gc (t − s1 )h1 (t − s2 )u1 (s1 )u1 (s2 )ds1 ds2 gs (t − s1 )gs (t − s2 )u1 (s1 )u1 (s2 )ds1 ds2

gc (t − s1 )gc (t − s2 ) + gs (t − s1 )gs (t − s2 )

(10) 

u1 (s1 )u1 (s2 )ds1 ds2 h11i 2 (t − s1 , t − s2 )u1 (s1 )u1 (s2 )ds1 ds2 ,

where h11i 2 (t1 , t2 ) = gc (t1 )gc (t2 ) + gs (t1 )gs (t2 ). Therefore, the operation performed by a complex cell can be modeled with a second order Volterra kernel. h11i 2 is shown in Figure 2B. We now take a closer look at the operation of the second order kernel. The two dimensional convolution of the second order kernel with u2 (t1 , t2 ) is shown in Figure 2C. It is important to note that, since the second order kernel has finite memory, it may not have enough support to cover the entire domain D2 for u2 (t1 , t2 ). For example, as illustrated in Figure 2C, the output of the second order feedforward DSP at time t is given by the integral of the product of u2 (t1 , t2 ) and a rotated h11i 2 with the origin shifted to (t, t) [see also (10)]. Since the shift is along the diagonal, only u2 (t1 , t2 ) in the domain that is contained within the black lines is multipled by nonzero values of h11i 2 . u2 (t1 , t2 ) elsewhere in the domain is always multiplied by zero in evaluating the output. Therefore, the output of the second order filter only contains information about u2 within the domain located in between the black lines in Figure 2C. This has implications on decoding the signal (see also Remark 3.11 in Section 3.2) 2.3.2. Feedback Volterra dendritic stimulus processors

As already mentioned, the feedback DSPs do not operate on stimuli directly but rather on spikes generated by BSGs. We assume 2ji 2ji that h1 ∈ L1 (D1 ), h2 ∈ L1 (D2 ), i  = j, are real, causal, BIBOstable and have finite memory. In addition, we assume that these kernels are effectively band-limited (see also Section 2.2.2). In  functional form we denote a train of spikes as k K12 (tki ; t). The output of the feedback DSP i amounts to        2ji j 2ji j P 2 h1 P 2 h2 t − tl + t − tl , t − tki (11) l∈Z

k∈Z l∈Z

with j  = i.

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 5

Lazar and Zhou

Neural circuits with intrinsic noise sources

FIGURE 2 | Examples of Volterra kernels. (A) First order kernels of quadrature pair of Gabor functions modeling the receptive fields of simple cells. (B) Second order kernel modeling receptive fields of complex cells. (C) The mechanics of the two dimensional convolution operation between the u2 (S 1 = 0.8, D2 = [0, 0.8] × [0, 0.8]) and h11i 2 . u2 (t1 , t2 ) = u1 (t1 )u1 (t2 ) is shown in the background. The inset shows the second order Volterra kernel h11i 2

In particular, the inputs to the second order term of the feedback DSPs are generated by two neurons. This allows for modeling nonlinear interactions between the two neurons in the spike domain. 2.3.3. Overall output from DSPs

The overall inputs (without noise) to the two BSGs in Figure 1 are  vi (t) =

D

 h11i 1 (t − s)u1 (s)ds +

D2

h11i 2 (t − s1 , t − s2 ) u1 (s1 )

     2ji j 2ji P 2 h1 P 2 h2 t − tl + u1 (s2 )ds1 ds2 + l∈Z

l∈Z k∈Z

  j t − tl , t − tki , and i, j = 1, 2, i  = j.

(12) The system of Equations (12) above functionally describe the post-synaptic aggregate currents that are injected into the BSG i. There are a variety of noise sources to be considered. Synaptic variability of feedforward DSPs adds noise sources to the current input to the BSGs. These include thermal noise, synaptic background noise, etc. (Jonston, 1927; Calvin and Stevens, 1968; Manwani and Koch, 1999; Fellous et al., 2003; Destexhe and Rudolph-Lilith, 2012). Feedback DSP kernels may themselves be subject to intrinsic noise sources that may lead to variability in the spike generation process. Intrinsic variability of BSG spike times can, e.g., contribute to the variability of the aggregate current driving the axon hillock in feedback loops. Overall, the combined effect of DSP noise sources is modeled as Gaussian white noise processes that are added to the feedforward and feedback DSP outputs. The sum total of signal and noise represents the aggregate current input to the BSGs (see Figure 1). Formal DSP noise models will be incorporated directly into the BSG model presented in the next section.

Frontiers in Computational Neuroscience

rotated 180◦ around origin [see also (B)]. (h11i 2 is only shown in a restricted domain and is zero elsewhere). For t = 0.3, the output of the convolution is the integral of the product of the rotated Volterra kernel and the signal underneath. Since the convolution is evaluated on the diagonal t = t1 = t2 , the second order kernel shifts, as t increases, along the arrow on the diagonal. See also Supplementary Figure 5E.

2.4. BIOPHYSICAL SPIKE GENERATORS

2.4.1. BSGs and phase response curves

We consider biophysically realistic spike generators such as the Hodgkin-Huxley, Morris-Lecar, Connor-Stevens neurons (Hodgkin and Huxley, 1952; Connor and Stevens, 1971; Morris and Lecar, 1981). The class of BSGs can be expressed in vector notation as   dxi (13) = f i xi , I i , i = 1, 2, dt where xi are the state variables, fi are vector functions of the same dimension, and I i are the constant bias currents in the voltage equation of each BSG. Each input current vi (t) is applied to the neuron i by additive coupling to the voltage equation, typically the first of the set of ordinary differential equations, i.e.,    T dxi = f i xi , I i + vi (t), 0 , i = 1, 2, dt

(14)

where 0 is a row vector of appropriate size. We assume that the neuron is periodically spiking when no external input is applied. This can be satisfied by a constant bias current I i additively coupled onto the voltage equation. The use of I i is necessary to formulate the encoding for the single neuron case, and this assumption will be relaxed later in this article. A large enough bias current induces a periodic oscillation of the biophysical spike generator. Therefore, the phase response curve (PRC) is well defined for this limit cycle (Izhikevich, 2007). We denote the PRC of the limit cycle induced by the bias current         T I i as ψ i t, I i = ψ1i t, I i , ψ2i t, I i , · · · , ψNi i t, I i with   appropriate dimension N i , where ψni t, I i , n = 1, 2, · · · , N i , are the PRCs associated with the  nthstate variable. Without loss of generality, we assume that ψ1i t, I i is always the PRC associated with the voltage variable. An example of a Hodgkin-Huxley neuron model of a BSG can be found in Section 2.2 in the Supplementary Material.

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 6

Lazar and Zhou

Neural circuits with intrinsic noise sources

2.4.2. Channel noise in BSGs

As shown in Figure 1, we consider BSGs with noise sources in the ion channels. The noise arises due to thermal fluctuations (White et al., 2000; Hille, 2001) as the finite number of ion channels in the BSGs open and close stochastically. The differential equations that govern the dynamics of the BSGs in (14) are deterministic. The set of stochastic differential equations (SDEs) below represent their stochastic counterpart (Lazar, 2010):     d Yi = f i Yi , I i dt + Bi Y i d Z i (t), i = 1, 2, (15) where Bi is a matrix with state dependent values, dZi = T  vi dt, dW2i , dW3i , · · · , dWPi i , and Wpi (t), p = 2, · · · , Pi , are independent Brownian motion processes. Note that Pi does not necessarily have to be equal to N i , the number of state variables. The first element in the stochastic differential dZi is the aggregate dendritic input vi dt driving the voltage equation. The other entries in dZi are noise terms that reflect the stochastic fluctuation in the ion channels / gating variables. Randomness is often added to BSGs by setting Bi = I, where I is a N i × N i identity matrix. The later setting can be viewed as adding subunit noise (Goldwyn and Shea-Brown, 2011). Recently, it has been suggested that a different way of adding channel noise into the BSGs may result in more accurate stochastic behavior (Goldwyn and Shea-Brown, 2011; Goldwyn et al., 2011; Linaro et al., 2011; Orio and Soudry, 2012). The SDEs in (15) are of general form and do not preclude them. In fact, by setting Bi to be a block matrix with blocks equal to be the square root of the diffusion matrix for each ion channel, the channel SDE model (Goldwyn et al., 2011; Orio and Soudry, 2012) can easily be incorporated into (15). Finally, we note that, under appropriate technical conditions the SDE formulation applies to BSGs with voltage-gated ion channels as well as other types of ion channels. The conditions require that the BSG model can be treated mathematically as a system of SDEs of the form (15) and that the latter satisfies the assumptions of Section 2.4.1.

Note that in the system of Equations (16) the two output spikes trains tki , i = 1, 2, k ∈ Z, are the observables. Due to the intrinsic noise sources in the DSPs and in the BSGs, spike timing jitter may be observed from trial to trial by repeatedly applying the same stimulus to the neural circuit (see Section 2.3 in the Supplementary Material).

3. ENCODING, DECODING, AND NOISE In Section 3.1 we present the mathematical encoding formalism underlying the neural circuit in Figure 1. We formulate stimulus decoding as a smoothing spline optimization problem and derive an algorithm that reconstructs the encoded signal in Section 3.2. Finally, we analyze the effect of noise on stimulus decoding in Section 3.3. 3.1. ENCODING

In this section, we formulate a rigorous stimulus encoding model based on the neural circuit shown in Figure 1. The input of the circuit is a signal u1 modeling a typical sensory stimulus as described in Section 2.2.1. The neural circuit generates a multidimensional spike train that is assumed to be observable. We establish model equations by first describing the I/O relationship (i.e., the t-transform) of a single BSG. We then provide the t-transform of the entire neural circuit model that maps the input stimulus amplitude into a multidimensional spike timing sequence. 3.1.1. The I/O of the BSG

In the presence of a bias current I i and absence of external inputs, each BSG in Figure 1 is assumed to be periodically spiking. Provided that the inputs are small enough, and by using the PRC, the BSG dynamics of spike generation can be described in an one-dimensional phase space (Lazar, 2010). Definition 3.1. A neuron whose spike times (tki ), k ∈ Z, i = 1, 2, verify the system of equations 

tki

2.4.3. Overall encoding of the neural circuit model

where



⎤ v i dt + dW1i ⎢ ⎥ dW2i ⎢ ⎥ ⎥, d Zi = ⎢ . ⎢ ⎥ . . ⎣ ⎦ i dWPi

with vi (t) given by Equation (12).

Frontiers in Computational Neuroscience

    T ψ i s − tki + τ i s − tki , I i , I i

     Bi xi s − tki + τ i s − tki , I i , I i d Zi (s)     = T i I i − tki + 1 − tki ,

Taking into account the dendritic input from the feedforward DSPs and feedback DSPs, the encoding by the neural circuit model under the two noise sources is given by two systems of SDEs. With the Brownian motion W1i modeling the DSP white noise, the encoding of neuron i, i = 1, 2, can be expressed as     d Yi = f i Y i , I i dt + Bi Yi d Zi (t),

tki + 1

(17)

where (16)

      T (18) dτ i t − tki , I i = ψ i t − tki + τ i t − tki , I i , I i      Bi xi t − tki + τ i t − tki , I i , I i d Zi (t),     with τ i 0, I i = 0 and xi t, I i the periodic solution to (13) with bias current I i , is called a Project-Integrate-and-Fire (PIF) neuron with random thresholds. In (17), [·]T denotes transpose and T i (I i ) is the period of limit cycle with bias current I i .

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 7

Lazar and Zhou

Neural circuits with intrinsic noise sources

As its name suggests, the PIF projects a weighted version of the input embedded in noise and the ion channel noise associated with the gating variables (Bi dZi ) onto the PRCs of the corresponding gating variables on a time interval between two consecutive spikes. Note that the integrand in (17) is identical to the RHS of (19). τ i (t, I i ) on the LHS of (19) denotes the phase deviation and is driven by the perturbation on the RHS. The LHS of (17) represents the phase deviation measurement performed by the PIF neuron. The RHS of (17) provides the value of the measurement and is equal to the difference between the inter-spike interval and the period of the limit cycle. The BSG and the PIF neuron with random thresholds are, to the first order, I/O equivalent (Lazar, 2010). In Lazar (2010) it was also shown that a good approximation to the PIF neuron is the reduced PIF with random threshold. The functional description of the reduced PIF is obtained by setting the phase deviation in (17) to zero. Definition 3.2. The reduced PIF neuron with random threshold is given by the equations N  

tki + 1

i n = 1 tk

     ψni s − tki , I i bin1 xi s − tki , I i v i (s)ds

  = T i (I i ) − tki + 1 − tki + εki ,

(19)

where (εki ), k ∈ Z, is a sequence of independent Gaussian random variables with zero mean and variance      2 E εki Ii

=

Pi  p=1



tki + 1 tki

⎡ ⎣

Ni 

⎤2

     ψni s − tki , I i binp xi s − tki , I i ⎦ ds. (20)

n=1

For reasons of notational simplicity and without loss of generality, and unless otherwise stated, we shall assume here that B = I (N i = Pi ). The reduced PIF (rPIF) with random threshold can now be written as 

tki + 1

tki

      ψ1i s − tki , I i vi (s)ds = T i I i − tki + 1 − tki + εki , (21)

Definition 3.3. The reduced PIF neuron with conditional PRC and random threshold is given by the system of equations 

tki

where ∈ Z, i = 1, 2, is a sequence of independent Gaussian random variables with zero mean and variance    Ni   2 E εki (I i ) =

   ψ1i s − tki , bik vi (s) − bik + I0i ds = εki ,

      Ni  2 bik = E εki

i n = 1 tk



ψni (s − tki , I i )

2

ds.

(22)

Frontiers in Computational Neuroscience



 2 ds, ψni s − tki , bik

(24)

and ψ1i (s, bik ) is the conditional PRC (Kim and Lazar, 2012). The conditional PRC formulation above allows us to separate BSG inputs into a constant bias current and fluctuations around it on short inter-spike time intervals. The bias current can be estimated between consecutive spikes, making the deviation from the limit cycle small in each inter-spike interval even for strong inputs. Moreover, by considering the conditional PRCs, the assumption that BSGs oscillate in the absence of input can be nearly dropped. Thus, it is not required for BSGs to always be on a limit cycle. Only when the neuron enters the limit cycle do we consider formulating the encoding using the rPIF model with conditional PRCs. Remark 3.4. Note that by parametrizing each of the PRCs with bik , the variance of the error in (24) depends on the estimated PRC on each inter-spike interval. In conjunction with (23), we see that the variability of spike times depends on the strength of the input to the BSGs. 3.1.2. The t-transform of the neural circuit

The overall encoding by the neural circuit model can be expressed as 

tki + 1

  ψ1i s − tki , bik vi (s)ds

  = bik − I i

tki + 1

  ψ1i s − tki , bik ds + εki , i = 1, 2, k ∈ Z.

Substituting (12) into the above, we have 

The above analysis assumes that the inputs are weak and therefore the BSGs operate on a limit cycle. Stronger signals can be taken into account by considering a manifold of PRCs associated with a wide range of limit cycles (Kim and Lazar, 2012). By estimating the limit cycle and hence its PRC using spike times, we have the following I/O relationship for each of the BSGs.

tki + 1

i n = 1 tk

tki

tki + 1

(23)

  where bik = [T i ]−1 tki + 1 − tki , k ∈ Z, is the total estimated bias current on the inter-spike interval [tki , tki + 1 ], I0i is an initial bias that brings the neuron close to the spiking region in the absence of input and (by abuse of notation) εki , k ∈ Z, i = 1, 2, is a sequence of independent Gaussian random variables with zero mean and variance

tki

(εki ), k

tki + 1

tki + 1

h11i 1 (s − r)u1 (r)drds

D1

tki

 +

  ψ1i s − tki , bik

tki + 1

tki

ψ1i

  i i s − tk , bk

D2

h11i 2 (s − r1 , s − r2 )u1 (r1 )

u1 (r2 )dr1 dr2 ds

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 8

Lazar and Zhou

Neural circuits with intrinsic noise sources

  = bik − I i

  ψ1i s − tki , bik ds

tki + 1

tki







tki + 1

i l ∈ Z tk



i ∈ H1 such that and φ2k 2 i T2ki u2 = u2 , φ2k H1 , for all u2 ∈ H21 .

   2ji j ψ1i s − tki , bik P 2 h1 (s − tl )ds tki + 1

i l ∈ Z m ∈ Z tk

2

(25) Therefore, (26) can be rewritten in inner product form:

    2ji j i ψ1i s − tki , bik P 2 h2 s − tl , s − tm ds

+ εki , i, j = 1, 2, i  = j. We arrived at the following. Lemma 3.5. The model of encoding in Figure 1 is given in operator form by

T1ki u1 + T2ki u2 = qik + ki , i = 1, 2, k ∈ Z,

(26)

where u1 ∈ H11 , u2 ∈ H21 , u2 (t1 , t2 ) = u1 (t1 )u1 (t2 ), and, T1ki : H11 → R and T2ki : H21 → R are bounded linear functionals given by 

T1ki u1 =

tki



T2ki u2 =

tki + 1

tki + 1

tki

 ϕki (s)

D1

 ϕki (s)

 qik = (bik − I i )

D2

h11i 2 (s − r1 , s − r2 )u2 (r1 , r2 )dr1 dr2 ds,

tki + 1

tki



h11i 1 (s − r)u1 (r)drds,



ϕki (s)ds −

  2ji j ϕki (s) P 2 h1 (s−tl )ds

tki + 1

i l ∈ Z tk



  2ji j i ϕki (s) P 2 h2 s − tl , s − tm ds,

tki + 1

i l ∈ Z m ∈ Z tk



ψ i (t − tki , bik ) ϕki (t) = 1

2  12 E εki and ki , k ∈ Z, are independent random variables with normal distribution N (0, 1) and j = 1, 2, j  = i. Equation (26) is called the t-transform (Lazar and Tóth, 2004) of the neural circuit in Figure 1. Remark 3.6. The t-transform describes the mapping of the input stimulus u1 into the spike timing sequence (tki ), i = 1, 2, k ∈ Z. Thus, the t-transform shows how the amplitude information of the input signal is related to or transformed into the time information contained in the sequence of output spikes generated by the neural circuit.

i i H1 + u2 , φ2k H1 = qik + εki . u1 , φ1k 1

Recall that inner products are projections that are typically interpreted as measurements. In the Equation (27) above, the signals i and φ i , u1 and u2 are projected onto the sampling functions φ1k 2k respectively. We also note that traditional amplitude sampling of a bandlimited signal u1 at times (tn ), n ∈ Z, can be expressed as u1 ( · ), sinc (tn − ·)L2 (R) = u1 (tn ), 1

( t) is the impulse response of the ideal where sinc (t) = sin πt low pass filter bandlimited to 1 or in other words, the kernel of the RKHS of finite-energy band-limited functions (Lazar and Pnevmatikakis, 2009). Thus, the neural encoding model described by the Equation (27) can be interpreted as generalized i sampling with noisy measurements with sampling functions φ1k i and φ2k . The formulation of the encoding model can easily be extended to the case when M neural circuits encode a stimulus in parallel. This is shown schematically in Supplementary Figure 1. A left superscript was added in the figure to each of the components to indicate the circuit number.

3.2. DECODING

In the previous section, we showed that the encoding of a signal u1 by the neural circuit model with feedforward and feedback DSPs and BSGs can be characterized by the set of t-transform Equations (26). We noticed that the Equations (26) are nonlinear in u1 due to the second order Volterra term. However, by reinterpreting the second order term as linear functionals T2ki on the higher dimensional tensor space H21 , (26) implies that the measurements taken by each of the neurons are the sum of linear measurements in two different vector spaces [see also Equations (27)]. In this section we investigate the decoding of signals encoded with the neural circuit in Figure 1. The purpose of decoding is to recover from the set of spike times the original signals, u1 (t) and u2 (t1 , t2 ), that respectively belong to the two different vector spaces H11 and H21 . We formulate the decoding problem as the joint smoothing spline problem 

1

Frontiers in Computational Neuroscience



uˆ 1 , uˆ 2 =

⎧ ⎨ argmin u1 ∈ H11 ,u2 ∈ H12

λ1 u1 2H1 + λ2 u2 2H1



1

2

⎫ ni  2  2 ⎬  + T1ki u1 + T2ki u2 − qik , (28) ⎭

We provide here further intuition behind the Equations (26). By the Riesz representation theorem (Berlinet and Thomasi ∈ H1 such that Agnan, 2004), there exists functions φ1k 1 i T1ki u1 = u1 , φ1k H1 , for all u1 ∈ H11 ,

(27)

2

i=1k=1

where ni + 1 is the number of spikes generated by BSG i = 1, 2.

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 9

Lazar and Zhou

Neural circuits with intrinsic noise sources

Theorem 3.7. The solution to (28) is of the form i

2  n 

uˆ 1 (t) =

i cki φ1k (t)

i=1 k=1

(29)

i

n 2  

uˆ 2 (t1 , t2 ) =

i cki φ2k (t1 , t2 ),

i=1 k=1 i (t) = T i K 1 i i 1 where φ1k 1k 1|t and φ2k (t1 , t2 ) = T2k K2|t1 , t2 , i = 1, 2, i k = 1, · · · , n ,  T c = c11 , · · · , cn11 , c12 , · · · , cn22 is the solution of the system of linear equations



 (1 + 2 )2 + λ1 1 + λ2 2 c = (1 + 2 ) q,

where q =



q11 , · · ·

, q1n1 , q21 , · · ·

i =

, q2n2

T

(30)

, and

12 11 i i 22 , i = 1, 2, 21 i i

mn i

 kl

3.3. EFFECT OF NOISE ON STIMULUS DECODING

= φikm , φiln .

Proof: Proof of the theorem follows the Representer Theorem (Berlinet and Thomas-Agnan, 2004) and is given in detail in Appendix. Remark 3.8. When λ1 = λ2 , the solution c amounts to c = (1 + 2 + λ1 I)−1 q, where I is an identity matrix of appropriate dimensions. Remark 3.9. Although (29) solves (28), in practice a minimum number of spikes is needed to obtain a meaningful estimate of the original signal. A minimum bound for the number of measurements/spikes can be derived in the noiseless case. Clearly, the bound has to be larger than the dimension of the space. This may require the signal to be encoded by a circuit with a larger number of neurons than the two shown in Figure 1 (Lazar and Slutskiy, in press). A number of such neural circuits in parallel can be used to encode input stimuli as shown in the Supplementary Figure 1. Theorem 3.7 can be easily extended to solving the smoothing spline problem 



uˆ 1 , uˆ 2 =

⎧ ⎨ argmin u1 ∈ H11 , u2 ∈ H12

λ1 u1 2H1 + λ2 u2 2H1



M  n  2  

1

m

T1ki u1 + m T2ki u2 − m qik

m=1 i=1 k=1

Frontiers in Computational Neuroscience

In this section, we investigate the effect of noise sources (i) on spike timing of the reduced PIF neuron, and (ii) on the decoding of stimuli encoded with a neural circuit. We will also present the effect of an alternative noise source model on both spike timing and stimulus decoding. 3.3.1. Effect of noise on measurement and spike timing errors of the reduced PIF neuron

As suggested by (22), the variance of the measurement error of the reduced PIF neuron is directly related to the PRC of the associated limit cycle. We first characterize the variance of the measurement error due to each individual noise source parametrized by the bias current I i . We then evaluate the spike timing variance between the spike trains generated by the Hodgkin-Huxley neuron and the reduced PIF neuron again as a function of the bias current I i . We start with a brief description of the key elements of Hodgkin-Huxley neuron and the PIF neuron. We consider the stochastic Hodgkin-Huxley equations   dYi = f i Y i , I i dt + dZ i (t),

(31)

where fi is defined as in Section 2.2 of the Supplementary Material with additional normalization such that the unit of time is in seconds instead of milliseconds and the unit of voltage is in Volts instead of milivolts as conventionally used. Zi (t) takes the form ⎡

2

m i

+

Remark 3.10. Since u2 (t1 , t2 ) = u1 (t1 )u1 (t2 ) = u2 (t2 , t1 ), u2 belongs to a subspace of H21 whose elements are symmetric functions. We also note that since the second order ker i feedforward  nels are symmetric, the sampling functions φ2k (t1 , t2 ) , i = 1, 2, k = 1, · · · , ni , also belong to the same subspace. Therefore, if the 1 sampling functions span the subspace of symmetric  1  functions in H2 , 1 u2 can readily be reconstructed with only L + 1 2L + 1 mea2  surements/spikes, rather than 2L1 + 1 , the dimension of H21 . Remark 3.11. The reconstruction of u2 (t1 , t2 ) on D2 strongly depends on the support (in practice the finite memory) of the kernels h11i 2 , i = 1, 2 (see also Figure 2C). In the reconstruction example of the Supplementary Figure 5, we show that uˆ 2 approximates u2 well in the restricted domain where h11i 2 is nonzero. Outside this restricted domain, h11i vanishes and u is not well recovered as suggested by the 2 2 large error in the Supplementary Figure 5E.

and

where m = 1, 2, . . . , M, denotes the circuits number in Supplementary Figure 1. In addition, if the circuits consist of only first order feedforward kernels, then only u1 (t) can be reconstructed. Similarly, if the circuits are comprised of only the second order feedforward kernels, then u2 (t1 , t2 ) can be reconstructed but not u1 (t).

⎫ 2 ⎬ ⎭

,

www.frontiersin.org

⎤ v i dt + σ1i dW1i ⎢ ⎥ σ2i dW2i ⎥. dZ i (t) = ⎢ i i ⎣ ⎦ σ3 dW3 i i σ4 dW4

September 2014 | Volume 8 | Article 95 | 10

Lazar and Zhou

Neural circuits with intrinsic noise sources

Here Wni (t) are independent standard Brownian motion processes and σni , n = 1, 2, 3, 4, are associated scaling factors. The variance of the measurement error of the reduced PIF neuron due to each Brownian motion process Wni , n = 1, · · · , 4, is given by [see also Equation (22)]         2 2 i I i = σni E εkn

tki + 1



tki

 2 ψni s − tki , I i ds. (32)

We show in Figure 3A the variance of the measurement error in (32) associated with each source of noise of the reduced PIF neuron for the unitary noise levels σni = 1, n = 1, 2, 3, 4. The variances given by (32) are plotted as a function of the bias current I i . Clearly, the noise arising in dendritic stimulus processing (W1i ) induces the largest error, and together with noise in the potassium channels (W2i ), these errors are about two magnitudes larger in variance than those induced by the noise sources in the sodium channels (W3i , W4i ). The above analysis is based on the analytical derivation of the measurement error in (32) for the rPIF neurons. The measurement error is closely related, however, to the spike timing variation of the BSGs subject to noise sources. A variance of 10−6 in Figure 3A corresponds to a standard deviation of 1 ms in spike timing. In practice the error between the spike times of the Hodgkin-Huxley neuron and the reduced PIF neuron can be directly evaluated.

FIGURE 3 | Variance of the measurement and spike timing errors. (A) Error measurement variances computed from the PRCs of the Hodgkin-Huxley neuron [Equation (32)]. Each individual variance is parametrized by the bias current I i . (B) Error variance between spike times generated by the noiseless Hodgkin-Huxley neuron and its reduced PIF counterpart. (C) The spike timing error variance due to each source of noise, obtained from simulations of the Hodgkin-Huxley neuron follow the pattern of the theoretically derived measurement error shown in (A). The spike timing error variances are obtained by setting, at each time, one of the σn ’s to a nonzero value and the rest

Frontiers in Computational Neuroscience

In order to do so, we randomly generated a weak bandlimited dendritic input. All evaluations were based on encoding a signal with the Hodgkin-Huxley neuron model described above with internal noise sources and bias current I i . The spike times (tki ) of the Hodgkin-Huxley neuron were recorded. Starting from each spike time tki , we encoded the appropriate portion of the signal by the reduced PIF neuron until a spike r tki + 1 was generated. The difference between r tki + 1 and tki + 1 is the error in approximating the encoding using the reduced PIF formulation. This process was repeated for each I i . We computed the variance of the errors based on some 3000–5000 spikes generated in encoding the input. In Figure 3B, the variance of the spike timing error r tki + 1 − i tk + 1 for σn = 0, n = 1, 2, 3, 4, is shown. Since the reduced PIF is an approximation (even under noiseless conditions) and, although small, the error is nonzero. From Figure 3B, the variance of the spike timing error is on the order of 10−9 . We shall evaluate the spike timing error variance of the intrinsic noise sources in a range much larger than 10−9 . We also tested to what extent each individual source of noise contributes to the variance of spike timing as suggested by the theoretical analysis depicted in Figure 3A. Indeed, the error variance obtained through simulations in Figure 3C follows the basic pattern shown in Figure 3A. Figure 3C was obtained by setting one of the σn ’s to a nonzero value and the rest to 0 (the nonzero values were σ1 = σ2 = 0.01, σ3 = σ4 = 0.1). Each nonzero value was picked to be large enough so that the error variance in the

to zero. The spikes generated by the Hodgkin-Huxley neuron are compared with the spikes generated by its reduced PIF counterpart. The variance of the differences between two spike times are normalized by the nonzero σn mentioned before. (D) The spike timing variance due to the simultaneous presence of multiple noise sources approximates the sum of spike timing variances due to individual noise sources. Blue curve shows the spike timing variance obtained by simulating Hodgkin-Huxley equations using nonzero values for all σn , n = 1, 2, 3, 4. Red curve shows the sum of spike timing variances obtained in (C) with proper scaling.

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 11

Lazar and Zhou

Neural circuits with intrinsic noise sources

absence of noise (Figure 3B) becomes negligible, and at the same time, it was small enough such that the states of the neurons did not substantially deviate from the limit cycle. To compare the with the ones in Figure 3A we normalized the error variance obtained in simulations by σn . Next, we tested whether the variance of spike timing due to presence of multiple noise sources is truly the summation of error variances due to individual noise sources. We simulated the Hodgkin-Huxley equations with σ1 = σ2 = 0.005, σ3 = σ4 = 0.05. The total spike timing error variance shown in Figure 3D (blue curve) is very close to the sum of error variances in Figure 3C with proper scaling (red curve in Figure 3D). As suggested by the above analysis, the reduced PIF neuron with random thresholds largely captures the encoding of stimuli by BSGs subject to intrinsic noise sources. 3.3.2. Effect of noise on stimulus decoding

In order to quantitatively explore how noise impacts signal decoding, we recovered from spikes the signal encoded by the noisy neural circuit of Supplementary Figure 1. We started with the base-level noise-less case described in Section 3.2 of the Supplementary Material (M = 4) and proceeded to introduce individual noise terms with a range of scaling factors. For example, we set σ2i = σ3i = σ4i = 0 and varied σ1i . We also tested the case when 10σ1i = 10σ2i = σ3i = σ4i for the aggregated effect on stimulus recovery. We choose to use σ3i and σ4i 10 times larger than σ1i and σ2i so that each noise source introduced a similar error. In all simulations, the Euler-Maruyama scheme (Kloeden and Platen, 1992) was used for the numerical integration of the SDEs. We performed 20 encoding and decoding experiments. Each time, the input stimulus was generated by randomly picking from a Gaussian distribution the real and imaginary parts of the coefficients ul in (1). We further constrained the stimuli to be real-valued. (An example is given in Supplementary Figure 5.) For each noise level, the input signal was encoded/decoded. The mean Signal-to-Noise Ratio (SNR) across 20 experiments is reported for each noise level. The SNR for the reconstruction of u1 was computed as SNR = 10 log10

u1 2 , u1 − uˆ 1 2

3.3.3. Effect of an alternative noise model on spike timing and stimulus decoding

Biologically, the effect of channel noise on the operation of the BSGs is due to the ON-OFF activity of a finite number of ion channels. The Hodgkin-Huxley equations and the noise terms used in Section 3.3.2 correctly capture the dynamics in the limit of infinitely many channels. Recent research, however, suggests that the model equations may not correctly model the ion current fluctuations for a finite number of channels (Goldwyn and Shea-Brown, 2011). We consider here an alternative stochastic formulation of the Hodgkin-Huxley model that more precisely captures the ion channel kinetics. By using a finite number of ion channels the strength of noise amplitude becomes directly related to the actual number of ion channels. Therefore, the free variables are only the number of potassium and sodium channels that are both biologically meaningful. The successful use of an alternative noise model as described below also suggests that our analysis can be applied to a wide range of stochastic formulations of BSGs based on SDEs. We shall construct here stochastic ion channels using conductance noise rather than subunit noise as in the previous Sections (Goldwyn and Shea-Brown, 2011; Goldwyn et al., 2011). This stochastic Hodgkin-Huxley system is simulated using a diffusion approximation following (Orio and Soudry, 2012). The system of SDEs can be expressed as     d Y i = f i Y i , I i dt + B i Y i dZ i (t), where Yi has 14 state variables and the full system can be found in Section 3.3 of the Supplementary Material. Here i = 1 for simplicity. The variance of the measurement error is now given by (20). We can decompose the variance into three terms as 2  2  2   2 i i i + E εkK + E εkNa , E εki = E εkV

(33)

where u1 is the original signal and uˆ 1 is its reconstruction. Note that the spike time occurrences generated for the same signal are different for each noise level. Since the sampling functions are spike time dependent, the number of measurements/spikes may not be the same for each noise level. Moreover, at times, the sampling functions may not fully span the stimulus space. To reduce the uncertainty caused by the stimulus dependent sampling we averaged our SNR data over 20 different signals. Figure 4A shows the SNR of the reconstruction of signal u1 (t) against different noise strength. Figure 4B shows the SNR of the reconstruction of signal u21 (t) = u2 (t, t). The reconstruction SNR in Figure 4A largely matches the inverse ordering of noise strength of each of the individual noise sources shown in Figure 3A. DSP noise sources degrade the reconstruction

Frontiers in Computational Neuroscience

performance most strongly while noise sources associated with gating variables m and h have a much smaller effect for the same variance level. Since the variance of measurement error is the sum of error variance in each variable, the case when 10σ1 = 10σ2 = σ3 = σ4 = σ exhibits the lowest performance.

i , εi , εi where εkV kK kNa are measurement errors associated with the noise in the DSP, in potassium channels and in sodium channels, respectively. i is quantitatively the same as that in Section 3.3.2, we As εkV i and ε i . The variance of the errors can be focus here on εkK kNa respectively expressed as

   2 i (I i ) E εkK =

5  

tki + 1

i p = 2 tk

www.frontiersin.org



6 

ψni (s − tki , I i )binp



i

x

(s − tki , I i )

  2 ds,

n=2

September 2014 | Volume 8 | Article 95 | 12

Lazar and Zhou

Neural circuits with intrinsic noise sources

FIGURE 4 | SNR reconstruction error of encoded signals with a total of M = 2 circuits (4 neurons). Color legend: (Blue) σ1i = σ , σ2i = σ3i = σ4i = 0. (Green) σ2i = σ , σ1i = σ3i = σ4i = 0. (Red) σ3i = σ , σ1i = σ2i = σ4i = 0. (Black) σ4i = σ , σ1i = σ2i = σ3i = 0. (Magenta) 10σ1i = 10σ2i = σ3i = σ4i = σ . In-sets (on

and   2 i E εkNa (I i ) =

15  

tki + 1

i p = 6 tk



14  n=7

ψni

  2    i i i i i i ds. s − tk , I bnp x (s − tk , I )

Note that bnp , n = 1, · · · , 14, p = 2, 3, · · · , 15, are functions that dependent on either the number of potassium channels NNa or the number of sodium channels NK , and the states of the neuron.

Frontiers in Computational Neuroscience

the left) are typical reconstructions that yield corresponding SNR indicated by arrows. The top left in (A) shows an example of reconstruction (green) whose SNR is 25 dB when compared to the original signal (blue). (A) SNR of reconstruction of u1 (t). (B) SNR of reconstruction of u12 (t) = u2 (t, t).

i 2 i We first evaluate (E εkNa )(I ) using the PRCs. The PRCs are obtained by letting NNa = NK = ∞ and thereby making the system deterministic. Since the measurement error variance for fixed I i is proportional to (NNa )−1 , it is shown in Figure 5A as a function of the bias current Ii for NNa = 1. Similarly, the

i 2 variance of the measurement error E εkK (I i ) for NK = 1 is shown in Figure 5A, and it is proportional to (NK )−1 for a fixed I i . We notice that, when the number of channels is the same, the measurement error due to the sodium channels is of the same order of magnitude with the measurement error due to the potassium channels. However, the number of sodium channels is

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 13

Lazar and Zhou

Neural circuits with intrinsic noise sources

FIGURE 5 | The variance of the measurement and spike timing error associated with the sodium channels (blue) and the potassium channels (red) of the Hodgkin-Huxley equations with alternative noise sources parametrized by the bias current I. (A) The variance of the measurement error computed from PRCs of Hodgkin-Huxley equations, with NNa = 1 and NK = 1. Actual variance with different number of ion channels is inversely proportional to NNa and NK , respectively. (B) Spike timing variance obtained in simulations by comparing the spike times generated by the Hodgkin-Huxley neuron with channel noise and the spike times generated by its reduced PIF counterpart. Blue curve is obtained by using NNa = 5 × 104 , NK = ∞, and normalized to 1 sodium channel. Red curve is obtained by using NK = 5 × 104 , NNa = ∞, and normalized to 1 potassium channel.

typically 3–4 times larger than the number of potassium channels. Therefore, in contrast to the previous case, the error induced by sodium channels shall be larger than that induced by potassium channels. We also analyzed in simulations the difference between spike times generated by the alternative stochastic formulation of the Hodgkin-Huxley equations and those generated by the corresponding reduced PIF neuron. We used in simulation NNa =

i 2  i 4 5 × 10 , NK = ∞, to obtain the variance E εkNa (I ) and scaled it by NNa to compare it with Figure 5A. Similarly, we used

i 2 i (I ). NK = 5 × 104 , NNa = ∞, to obtain the variance E εkK The spike timing variances of error across different I i are shown in Figure 5B The pattern of similarity between variances in Figures 5A,B suggest that the reduced PIF with random threshold associated with this formulation of Hodgkin-Huxley equations is highly effective in capturing the encoding under internal noise sources. We now show how ion channel noise sources affect the decoding of the input signal. We varied the number of sodium channels NNa and fixed the number of potassium channels to be NK = 0.3NNa , a ratio typically used for Hodgkin-Huxley neurons with the alternative noise source model. By decoding the input

Frontiers in Computational Neuroscience

FIGURE 6 | SNR of reconstruction of the signals. (A) SNR of u1 (t). (B) SNR of u12 (t) = u2 (t, t). (Blue) NNa = N, NK = 0.3NNa . (Green) NNa = N, NK = ∞. (Red) NNa = ∞, NK = N.

stimulus we show how increasing the number of ion channels improves the faithfulness of signal representation. The SNR of the reconstruction of u1 (t) and u21 (t) are depicted in Figure 6. SNR goes down to about 4 dB when 600 sodium channels and 180 potassium channels are used. This corresponds to a membrane area of about 10 µm2 with a density of 60 µm2 in sodium channels and 18 µm2 in potassium channels (Goldwyn et al., 2011). We also tested the reconstruction for the case when one type of ion channels is infinitely large, i.e., deterministic, while varying the number of ion channels of the other type. The result is also shown in Figure 6. The noise from the dendritic tree shall have similar effect on the representation since the voltage equation is the same as in Section 3.3.2.

4. FUNCTIONAL IDENTIFICATION AND NOISE In Section 4.1 we show how to functionally identify the feedforward and feedback DSPs of the circuit in Figure 1 under noisy conditions. We assume here that the BSGs have already been identified using a methodology such as the one developed in Lazar and Slutskiy (2014). In Section 4.2 we discuss the effect of noise parameters on the quality of DSP identification. 4.1. FUNCTIONAL IDENTIFICATION

In the decoding problem analyzed in Section 3.2, we reconstructed unknown input stimuli by assuming that the neural circuit in Figure 1 is known and the spike trains are observable. In contrast, the objective of the functional identification problem investigated in this section is to estimate the unknown circuit parameters of the feedforward and feedback DSPs from I/O data.

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 14

Lazar and Zhou

Neural circuits with intrinsic noise sources

The I/O data is obtained by stimulating the circuit with controllable inputs and by measuring the time occurrences of the output spikes. This basic methodology has been a standard practice in neurophysiology for inferring the function of sensory systems (Hubel and Wiesel, 1962). We assume here that either the BSGs are known in functional form or the family of PRCs associated with the BSGs have already been identified (Lazar and Slutskiy, 2014). Although decoding and functional identification are seemingly two different problems, they are closely related. By exploiting the commutative property of linear operators, we can rearrange the diagram of the neural circuit model of Figure 1 into the form shown in Figure 7. We notice that the outputs of Figure 7 and those of Figure 1 are spike time equivalent, as long as the RKs K12 and K22 have large enough bandwidth. In what follows we will evaluate the four Volterra terms, i.e., the four dendritic currents feeding the BSG of Neuron 1 in Figure 7. Formally, the first order (feedforward) Volterra term can be written as (Lazar and Slutskiy, in press) 

 D1

h11i 1 (t − s)u1 (s)ds =

D1

u1 (t − s)(P 1 h11i 1 )(s)ds.

(34)

Similarly, the second order (feedforward) Volterra term amounts to  D2

h11i 2 (t − s1 , t − s2 ) u2 (s1 , s2 ) ds1 ds2

FIGURE 7 | Diagram of the neural circuit that is spike equivalent with the one in Figure 1 highlighting the between neural decoding and functional identification. the input stimuli and the DSP projections are reordered

Frontiers in Computational Neuroscience

(35)

timing duality Note that to reflect

 =

D2

  u2 (t − s1 , t − s2 ) P 1 h11i (s1 , s2 ) ds1 ds2 . 2

The above equations suggest that the projections of the feedforward kernels can be re-interpreted as inputs, whereas the signals u1 and u2 can be treated as feedforward kernels. In Section 2.2.2 we introduced two RKHSs, H12 and H22 , for modeling two different spaces of spikes. The elements of H12 are functions defined over the domain [0, S2 ] with 2ji

S2 ≥ supp h1

!

+ max

 ! tki + 1 − tki

i = 1,2,k ∈ Z

.

The period S2 is large enough to ensure that any spike that arrives 2ji supp{h1 } seconds prior to the arrival of tki , or earlier, will not affect the output of the feedback kernel on the inter-spike time interval [tki , tki + 1 ]. Thus, such spikes will not introduce additional error terms in the t-transform evaluated on the inter-spike time interval [tki , tki + 1 ]. Note that the domain [0, S2 ] of the functions in H12 may not be the same as the domain of the input space H11 . However, such a domain can be shifted on a spike by spike basis to [tki + 1 − S2 , tki + 1 ] for the inter-spike time interval [tki , tki + 1 ]. This is important for mitigating the practical limitation of modeling the stimuli as periodic functions in H11 . The response of the first-order feedback term to its spiking input on the inter-spike time interval [tki , tki + 1 ] in Figure 7 amounts to (i  = j)

that the unknowns are the DSP projections. The input stimuli u11 (t), u21 (t1 , t2 ), and the kernel representation of spikes (see also Section 2.2.2) are intrinsic to the neural circuit. The DSP projections are interpreted as inputs.

www.frontiersin.org

September 2014 | Volume 8 | Article 95 | 15

Lazar and Zhou



2ji

P 2 h1



Neural circuits with intrinsic noise sources



(t) ∗

l:tki + 1 −S2

j ≤ tl



j

K12 (tl , t) = < tki + 1

l:tki + 1 −S2

j ≤ tl

2ji

j

(P 2 h1 )(t − tl ).

for identification, the t-transform (26) can be written as

< tki + 1

(36)

m

j l:tki + 1 −S2 ≤ tl

  2ji j L2 P 2 h1 (t − tl ) −→

 j l:tki + 1 −S2 ≤ tl

< tki + 1

2ji

j

h1 (t − tl ) < tki + 1 2ji

if 2 is at least larger than the effective bandwidth of h1 and L2 → ∞. Similarly, the response of the second-order feedback kernel to its spiking input on the inter-spike time interval [tki , tki + 1 ] amounts to      2ji j P 2 h2 t − tl , t − tni j l:tki + 1 −S2 ≤ tl

L2

−→

i i i < tki + 1 n:tk + 1 −S2 ≤ tn < tk + 1





2ji

h2

i

i

 m 1i L1k [P 1 h11i 1 ]

=

m 1i L2k [P 1 h11i 2 ]

=

m ti k+1

m ti k



m ti k+1

m ti k

 m i ϕk (s)

D1

1 11i um 1 (s − r)(P h1 )(r)drds,

 m i ϕk (s)

D2

1 11i um 2 (s − r1 , s − r2 )(P h2 )

(r1 , r2 )dr1 dr2 ds,

i

(37) if 2 is large enough and L2 → ∞. Combining (34), (36), (36), and (37), for each spike interval [tki , tki + 1 ], the aggregated output current of the DSPs of Neuron i in Figure 7, shall converge to the DSP aggregated output current of Neuron i in Figure 1 for large enough 2 . A direct consequence of this equivalence is that, under the same additive Gaussian white noise and channel noise in the BSGs, the t-transform of the circuit in Figure 7 and in Figure 1 are identical. Note that the outputs of the feedforward kernels are always equivalent; the equivalence of the outputs of the feedback kernels requires, however, the use of large enough bandwidth 2 . Otherwise, the equivalence in the t-transform is invalid and an additional noise term appears in the t-transform of the Neuron 1 in Figure 7. The projections of the Volterra DSP kernels of Figure 7 are interpreted as inputs, while the input stimuli and the train of RKs at spike times replace the impulse response of the corresponding filters. Therefore, the functional identification problem has been transformed into a dual decoding problem, where the objects to decode are the set of projections of Volterra DSP kernels and the neural circuit is comprised of “stimulus DSP kernels” and “spike DSP kernels” with the same BSGs and noise sources. The only difference is that, instead of a Single-Input MultiOutput decoding problem, the identification was transformed into a Multi-Input Multi-Output decoding problem. In addition, multiple trials using different stimuli are needed; this procedure is illustrated in block diagram form in Figure 8. By stimulating the neural circuit with multiple stimuli in the functional identification setting, multiple neural circuits effectively encode the projections of the DSP kernels. We are now in the position to derive the t-transform of Neuron 1 in Figure 7. Assuming that m = 1, · · · , M, trials are performed

Frontiers in Computational Neuroscience

(38)

1 m 1i for i, j = 1, 2, i  = j, k ∈ Z. Here m L1i 1k : H1 → R, L2k : 1 H2 → R are bounded linear functionals associated with the 2 2 m 2i feedforward DSP kernels, and m L2i 1k : H1 → R, L2k : H2 → R are bounded linear functionals associated with the feedback DSP kernels for each trial m. The above functionals are defined as

  j t − tl , t − tni

l:tki + 1 −S2 ≤ tl < tki +1 n:tk + 1 −S2 ≤ tn < tk + 1 j

2ji

= m qik + m ki ,

It is clear from Section 2.2.2 that 

2ji

1 11i m 1i 1 11i m 2i 2 m 2i 2 L1i 1k [P h1 ] + L2k [P h2 ] + L1k [P h1 ] + L2k [P h2 ]

2ji m 2i L1k [P 2 h1 ]





=

m ti k+1

2ji m i ϕk (s)(P 2 h1 )

m ti k

j

l:m tki + 1 −S2 ≤ m tl