This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Adaptive stimulus design for dynamic recurrent neural network models

R.Ozgur DORUK11footnotemark: 1 resat.doruk@atilim.edu.tr Kechen Zhang kzhang4@jhmi.edu Atilim University, Department of Electrical and Electronics Engineering, Kizilcasar Mahallesi, Incek, Golbasi, Ankara, 06836,TURKEY Department of Biomedical Engineering, Johns Hopkins School of Medicine, 720 Rutland Avenue, Ross 528, Baltimore , MD, 21205, USA Department of Biomedical Engineering, Johns Hopkins School of Medicine, 720 Rutland Avenue, Traylor 407, Baltimore , MD, 21205, USA
Abstract

We present a theoretical application of an optimal experiment design (OED) methodology to the development of mathematical models to describe the stimulus-response relationship of sensory neurons. Although there are a few related studies in the computational neuroscience literature on this topic, most of them are either involving non-linear static maps or simple linear filters cascaded to a static non-linearity. Although the linear filters might be appropriate to demonstrate some aspects of neural processes, the high level of non-linearity in the nature of the stimulus-response data may render them inadequate. In addition, modelling by a static non-linear input - output map may mask important dynamical (time-dependent) features in the response data. Due to all those facts a non-linear continuous time dynamic recurrent neural network that models the excitatory and inhibitory membrane potential dynamics is preferred. The main goal of this research is to estimate the parametric details of this model from the available stimulus-response data. In order to design an efficient estimator an optimal experiment design scheme is proposed which computes a pre-shaped stimulus to maximize a certain measure of Fisher Information Matrix. This measure depends on the estimated values of the parameters in the current step and the optimal stimuli are used in a maximum likelihood estimation procedure to find an estimate of the network parameters. This process works as a loop until a reasonable convergence occurs. The response data is discontinuous as it is composed of the neural spiking instants which is assumed to obey the Poisson statistical distribution. Thus the likelihood functions depend on the Poisson statistics. The model considered in this research has universal approximation capability and thus can be used in the modelling of any non-linear processes. In order to validate the approach and evaluate its performance, a comparison with another approach on estimation based on randomly generated stimuli is also presented.

keywords:
Optimal Design, Sensory Neurons, Recurrent Neural Network, Excitatory Neuron, Inhibitory Neuron, Maximum Likelihood Estimation
journal: Journal of Biological Physicslabel2label2footnotetext: This work is partially supported by Turkish Scientific and Technological Research Council (TÜBİTAK) 2219 Research Program.

1 Introduction

Optimal experiment design (OED) or shortly optimal design [1] is a sub-field of optimal control theory which concentrates design of an optimal control law aiming at the maximization of the information content in the response of a dynamical system related to its parameters. The statistical advantage brought by information maximization helps the researchers to generate the best input to their target plant/system that can be used in a system identification experiment producing estimates with minimum variance [2]. With the utilization of mathematical models in theoretical neuroscience research, the application of optimal experiment design in adaptive stimuli generation should be beneficial as it is expected to have better evaluations of the model specific parameters from the collected stimulus-response data. Though these benefits, the optimal experiment design have not found its place among theoretical or computational neuroscience research due to the nature of the models. As the stimulus-response relationship is naturally quite non-linear, computational complexity of the optimization algorithms utilized for an optimal experiment design will typically be very high and thus OED has not gained enough attraction during the past decades. However, thanks to the today’s computational powers of new microprocessors, it will be much easier to talk about a real optimal experiment design in neuroscience research ([3],[4]). In the past decades, some researchers had stimulated their models by Gaussian white noise stimuli [5], [6] and performed an estimation of input-output relationships of their model ([7] [8] and [9]). This algorithmically simpler approach is theoretically proven to be efficient in the estimation of models based on linear filters and their cascades. However, in [10], it is suggested that white noise stimuli may not be successful as a stimuli in the parametric identification of non-linear response models due to high level of parameter confounding (refer to [11] for a detailed description of the confounding phenomenon in non-linear models).

Concerning the applications of optimal experiment design to biological neural network models, there exist a limited amount of research. One such example is [12] where a static non-linear input output mapping is utilized as a neural stimulus-response model. The optimal design of the stimuli is performed by the maximization of the D-Optimal metric of the Fisher Information Matrix (FIM) [13] which reflects a minimization of the variance of the total parametric error of the model network. In the last research,the parameter estimation is based on the Maximum A Posteriori (MAP) Estimation methodology [14] which is linked to the Maximum Likelihood Estimation (ML) [15] approach. Two other successful mainly experimental work on applications of optimal experiment design to adaptive data collection are [16] and [17]. The experimental works successfully proven the efficiency of optimal designs for certain models in theoretical neuroscience. However, none of those studies explore fully dynamical non-linear models explicitly. Because of this deficiency, this research will concentrate on an application of the optimal experiment design to a fully dynamical non-linear model. The final goal is almost similar to that of [12].

The proposed model is a continuous time dynamical recurrent neural network (CTRNN) [18] in general and it also represents the excitatory and inhibitory behaviours [19] of the realistic biological neurons. Like in that of [20] and its derivatives, the CTRNN describes the dynamics of the membrane potentials of the constituent neurons. However, the channel activation dynamics is not directly represented. Instead it constitutes, a more generic model which can be applied to a network having any number of neurons. The dynamic properties of the neuron membrane is represented by time constants and the synaptic excitation and inhibition are represented as network weights (scalar gains). Though not the same, a similar excitatory-inhibitory structure is utilized in numerous studies such as [21, 22, 23]. As there isn’t sufficient amount of research on the application of OED to dynamical neural network models, it will be convenient to start with a basic network model having two neurons representing the average of excitatory and inhibitory populations respectively. The final goal is to estimate the time constants and weight parameters. The optimal experiment design will be performed by maximizing a certain metric of the FIM. The FIM is a function of stimulus input and network parameters. As the true network parameters are not known in the actual problem, the Information Matrix should depend on the estimated values of the parameters in the current step. An optimization on a time dependent variable like stimulus will not be easy and often its parametrization is required. In auditory neuroscience point of view, that can be done by representing the stimuli by a sum of phased cosine elements. If periodic stimulation is allowed, these can be formed as harmonics based on a base stimulation frequency. The optimally designed stimulus will be the driving force of a joint maximum likelihood estimation (JMLE) process which involves all the recorded response data. Unfortunately, the recorded response data will not be continuous. The reason for this is that, in vivo measurements of the membrane potentials are often very difficult and dangerous as the direct manipulation with the neuron in vivo may trigger the death of a neuron. Thus, in the real experimental set-up, the peaks of the membrane potentials are collected as firing instants. As a result, one will only have a neural spike train with the exact neural spiking times (timings of the membrane potential peaks) but no other data. This outcome prevents one to apply traditional parameter estimation techniques such as minimum mean square estimation (MMSE) as it will require continuous firing rate (is based on the membrane potential) data. Researches like [24], suggests that the neural spiking profile of sensory neurons obey the famous inhomogeneous Poisson distribution [25]. Under this assumption, the Fisher Information Matrix [12] and Likelihood functions [26, 27] can be derived based on Poisson statistical distribution. The optimization of a certain measure of Fisher Information Matrix and the Likelihood can be performed by readily available packages such as MATLAB® Optimization Toolbox (like well known fmincon algorithm).

There are certain challenges in this research. First of all, the limited availability of similar studies lead to the fact that this work is one of the first contributions on the applications of optimal experiment design to the dynamical neural network modelling. Secondly, we will most probably not be able to have a reasonable estimate just from a single spiking response data set as we do not have a continuous response data. This is also demonstrated in the related kernel density estimation research such as [28, 29, 30, 31]. From these sources, one will easily note that repeated trials and superimposed spike sequences are required to obtain a meaningfully accurate firing rate information from the neural response data. In a real experiment environment, repeating the trials with the same stimulus profile will not be appropriate as the repeated responses of the same stimulus are found to be attenuated. Because of this issue, a new stimulus should be designed each time based on the currently estimated parameters of the model and then it should be used in an updated estimation. These updated parameters are used in the next step to generate the new optimal stimulus. As a result one will have a new stimulus in each step and thus the risk of response attenuation is largely reduced. In a maximum likelihood estimation, the likelihood function will depend on the whole spiking data obtained throughout the experiment (or simulation). The parallel processing capabilities of MATLAB® (i.e. parfor) on multiple processor/core computers will help in resolving of those issues.

2 Models & Methods

2.1 Continuous Time Recurrent Neural Networks

The continuous time recurrent neural networks have a similar structure to that of the discrete time counterparts that are often met in artificial intelligence studies. In Figure 1, one can see a general continuous time network that may have any number of neurons.

Refer to caption
Figure 1: (A) A generic recurrent neural network structure. The stimulus means external inputs to the network. (B) A simple recurrent network with one excitatory unit and one inhibitory unit, with both units having nonlinear sigmoidal gain functions. Here each unit may represent a population of neurons.We assume that the recorded responses are inhomogeneous Poisson spike trains based on the continuous rate generated by the state of the excitatory unit.

The mathematical representation of this generic model can be written as shown below [18]:

τidVidt=Vi+j=1nWijgj(Vj)+k=1mCikIk\tau_{i}\frac{dV_{i}}{dt}=-V_{i}+\sum_{j=1}^{n}W_{ij}g_{j}\left(V_{j}\right)+\sum_{k=1}^{m}C_{ik}I_{k} (1)

where τk{\tau_{k}} is the time constant, Vk{V_{k}} is the membrane potential of the kth{k^{th}} neuron, Wkj{W_{kj}} is the synaptic connection weight between the kth{k^{th}} and jth{j^{th}} neurons Cki{C_{ki}} is the connection weight from ith{i^{th}} input to the kth{k^{th}} neuron and Ik{I_{k}} is the ith{i^{th}} input. The term gj(Vj){g_{j}\left(V_{j}\right)} is a membrane potential dependent function which acts as a variable gain on the synaptic inputs to from the jth{j^{th}} neuron to the kth{k^{th}} one. It can be shown by a logistic sigmoid function which can be shown as:

gj(Vj)=Γj1+exp(aj(Vjhj))g_{j}\left(V_{j}\right)=\frac{\Gamma_{j}}{1+\exp\left(-a_{j}\left(V_{j}-h_{j}\right)\right)} (2)

where Γj\Gamma_{j} is the maximum rate at which the jth{j^{th}} neuron can fire, hjh_{j} is a soft threshold parameter of the jth{j^{th}} neuron and aj{a_{j}} is a slope constant. This is the only source of non-linearity in (1). In addition it also models the activation-inactivation behaviour in more specific models of the neuron (like [20]). The work by [32] shows that (2) gives a relationship between the firing rate rj{r_{j}} and membrane potential Vj{V_{j}} of the jth{j^{th}} neuron. In sensory nervous system, some of neurons have excitatory synaptic connections while some have inhibitory ones. This fact is reflected to the model in (1) by assigning negative values to the weight parameters which are originating from neurons with inhibitory synaptic connections. In the introduction of this research, it is stated that it would be convenient to apply the theory to a basic network first of all due to the lack of related research and computational complexity. So a basic excitatory and inhibitory continuous time recurrent dynamical network can be written as shown in the following:

τeV˙e\displaystyle\tau_{e}\dot{V}_{e} =\displaystyle= Ve+weege(Ve)weigi(Vi)+weI\displaystyle-V_{e}+w_{ee}g_{e}\left(V_{e}\right)-w_{ei}g_{i}\left(V_{i}\right)+w_{e}I (3)
τiVi˙\displaystyle\tau_{i}\dot{V_{i}} =\displaystyle= Vi++wiege(Ve)wiigi(Vi)+wiI\displaystyle-V_{i}++w_{ie}g_{e}\left(V_{e}\right)-w_{ii}g_{i}\left(V_{i}\right)+w_{i}I (4)

where the subscripts e{{}^{\prime}e^{\prime}} and i{{}^{\prime}i^{\prime}} stands for excitatory and inhibitory neurons respectively. Starting from now on, we will have a single stimulus and it will be represented by the term I{I} which will be generated by the optimal design algorithm. In addition in order to suit the model equations to the estimation theory formalism the time constant may be moved to the right hand side as shown below:

ddt[VeVi]=[βe00βi]{[VeVi]+[weeweiwiewii][ge(Ve)gi(Vi)]+[wewi]I}\frac{d}{dt}\left[\begin{array}[]{c}V_{e}\\ V_{i}\end{array}\right]=\left[\begin{array}[]{cc}\beta_{e}&0\\ 0&\beta_{i}\end{array}\right]\left\{-\left[\begin{array}[]{c}V_{e}\\ V_{i}\end{array}\right]+\left[\begin{array}[]{cc}w_{ee}&-w_{ei}\\ w_{ie}&-w_{ii}\end{array}\right]\left[\begin{array}[]{c}g_{e}\left(V_{e}\right)\\ g_{i}\left(V_{i}\right)\end{array}\right]+\left[\begin{array}[]{c}w_{e}\\ w_{i}\end{array}\right]I\right\} (5)

where βe\beta_{e} and βi\beta_{i} are the reciprocals of the time constants τe\tau_{e} and τi\tau_{i}. They are taken to the right for easier manipulations of the equations. Note that this equation is written in matrix form to be conformed to the formal non-linear system forms. A descriptive illustration related to (5) is presented in Figure 1b. It should also be noted that, in (4) and (5) the weights are all assumed as positive coefficients and they have signs in the equation. So negative signs indicate that originating neuron is inhibitory (tend to hyper-polarize the other neurons in the network).

2.2 Inhomogeneous Poisson spike model

The theoretical response of the network in (4) will be the firing rate of the excitatory neuron as re=ge(Ve){r_{e}=g_{e}\left(V_{e}\right)}. In the actual environment, the neural spiking due to the firing rate re(t){r_{e}\left(t\right)} is available instead. While introducing this research, it is stated that this spiking events conform to an inhomogeneous Poisson process which is defined below:

Prob[N(t+Δt)N(t)=k]=eλλkk!\mbox{Prob}\left[N\left(t+\Delta t\right)-N\left(t\right)=k\right]=\frac{e^{-\lambda}\lambda^{k}}{k!} (6)

where

λ=tt+Δtre(τ)𝑑τ\lambda=\int_{t}^{t+\Delta t}r_{e}\left(\tau\right)d\tau (7)

is the mean number of spikes based on the firing rate re(t)r_{e}(t) which varies with time, and N(τ)N(\tau) indicates the cumulative total number of spikes up to time τ\tau, so that N(t+Δt)N(t)N\left(t+\Delta t\right)-N\left(t\right) is the number of spikes within the time interval [t,t+Δt){\left[t,t+\Delta t\right)}. In other words, the probability of having k{k} number of spikes in the interval (t,t+Δt){\left(t,t+\Delta t\right)} is given by the Poisson distribution above.

Consider a spike train (t1,t2,,tK)(t_{1},t_{2},\ldots,t_{K}) in the time interval (0,T)(0,T) (here 0t1t2tKT0\leq t_{1}\leq t_{2}\leq\ldots\leq t_{K}\leq T so tt and Δt\Delta t become t=0t=0 and Δt=T\Delta t=T). Here the spike train is described by a list of the time stamps for the KK spikes. The probability density function for a given spiking train (t1,t2,,tK)(t_{1},t_{2},\ldots,t_{K}) can be derived from the inhomogeneous Poisson process [26, 27]. The result reads:

p(t1,t2,,tK)=exp(0Tre(t)𝑑t)k=1Kre(tk,𝐱,θ)p\left(t_{1},t_{2},\ldots,t_{K}\right)=\exp\left(-\int_{0}^{T}r_{e}\left(t\right)dt\right)\prod_{k=1}^{K}r_{e}\left(t_{k},{\bf x},\theta\right) (8)

This probability density describes how likely a particular spike train (t1,t2,,tK)(t_{1},t_{2},\ldots,t_{K}) is generated by the inhomogeneous Poisson process with the rate function re(t,𝐱,θ)r_{e}\left(t,{\bf x},\theta\right). Of course, this rate function depends implicitly on the network parameters and the stimulus used.

2.3 Maximum Likelihood Methods and Parameter Estimation

The network parameters to be estimated are listed below as a vector:

θ=[θ1,,θ8]=[βe,βi,we,wi,wee,wei,wie,wii]\theta=\left[\theta_{1},\ldots,\theta_{8}\right]=\left[\beta_{e},\beta_{i},w_{e},w_{i},w_{ee},w_{ei},w_{ie},w_{ii}\right] (9)

which includes the time constants and all the connection weights in the E-I network. Our maximum-likelihood estimation of the network parameters is based on the likelihood function given by (8), which takes the individual spike timings into account. It is well known from estimation theory is that maximum likelihood estimation is asymptotically efficient, i.e., reaching the Cramér-Rao bound in the limit of large data size.

To extend the likelihood function in (8) to the situation where there are multiple spike trains elicited by multiple stimuli, consider a sequence of MM stimuli. Suppose the mm-th stimulus (m=1,,Mm=1,\ldots,M) elicits a spike trains with a total of KmK_{m} spikes in the time window [0,T][0,T], and the spike timings are given by Sm=(t1(m),t2(m),,tKm(m)){S}_{m}=\left(t_{1}^{(m)},t_{2}^{(m)},\ldots,t_{K_{m}}^{(m)}\right). By (8), the likelihood function for the spike train SmS_{m} is

p(Smθ)=exp(0Tre(m)(t)𝑑t)k=1Kmre(m)(tk(m))p\left(S_{m}\mid\theta\right)=\exp\left(-\int_{0}^{T}r_{e}^{(m)}\!\left(t\right)dt\right)\prod_{k=1}^{K_{m}}r_{e}^{(m)}\!\left(t_{k}^{(m)}\right) (10)

where re(m)r_{e}^{(m)} is the firing rate in response to the mm-th stimulus. Note that the rate function re(m)r_{e}^{(m)} depends implicitly on the network parameters θ\theta and on the stimulus parameters. The left-hand side of (10) emphasizes the dependence on network parameters θ\theta, which is convenient for parameter estimation. The dependence on the stimulus parameters will be discussed in the next section.

We assume that the responses to different stimuli are independent, which is a reasonable assumption when the inter-stimulus intervals are sufficiently large. Under this assumption, the overall likelihood function for the collection of all MM spike trains can be written as

L(S1,S2,,SMθ)=m=1Mp(Smθ)L\left({S}_{1},S_{2},\ldots,S_{M}\mid\theta\right)=\prod_{m=1}^{M}p\left(S_{m}\mid\theta\right) (11)

By taking natural logarithm, we obtain the log likelihood function:

l(S1,S2,,SMθ)=m=1M0Tre(m)(t)𝑑t+m=1Mk=1Kmlnre(m)(tk(m))l\left({S}_{1},S_{2},\ldots,S_{M}\mid\theta\right)=-\sum_{m=1}^{M}\int_{0}^{T}r_{e}^{(m)}\!\left(t\right)dt+\sum_{m=1}^{M}\sum_{k=1}^{K_{m}}\ln{r_{e}^{(m)}}\!\left(t_{k}^{(m)}\right) (12)

Maximum-likelihood estimation of the parameter set is given formally by

θ^ML=argmaxθl(S1,S2,,SMθ)\hat{\theta}_{ML}=\arg\max_{\theta}l\left({S}_{1},S_{2},\ldots,S_{M}\mid\theta\right) (13)

Numerical issues related to this optimization problem will be discussed in Sections 2.5 and 2.6. In addition, some discussion on the local maxima problems is provided in Section 3.3.

2.4 Objective function of optimal design of stimuli

The optimal design method generates the stimuli by maximizing a utility function, or an objective function. The basic idea is that these stimuli are designed so as to elicit responses that are most informative about the network parameters. In optimal design method, the utility function U(𝐱,θ)U({\bf x},\theta) depends on the stimulus parameters 𝐱\bf x, but typically also on the model parameters θ\theta. An intuitive explanation of the dependence on the model parameter is best illustrated with an example. Suppose we want to estimate a Gaussian tuning curve model with unknown parameters although we may have some idea about the sensible ranges of these parameters. To estimate the height of the tuning curve accurately, we should place a probing stimulus around the likely location of the peak. To estimate the width, the probing stimulus should go to where the tuning curve is likely to have the steepest slope. For the baseline, we should go for the lowest response. This simple example illustrates two facts: first, optimal design depends on our knowledge of possible parameter values; second, the elicited responses in an optimally design experiment are expected to vary over a wide dynamic range as different parameters are estimated.

Once the utility function U(𝐱,θ)U({\bf x},\theta) is chosen, the optimally designed stimulus may be written formally as:

𝐱^=argmax𝐱U(𝐱,θ)\hat{\bf x}=\arg\max_{\bf x}U({\bf x},\theta) (14)

where the network parameters θ\theta can be obtained by maximum-likelihood estimation from the existing spike data as described in the preceding section. Here the stimulus is specified by vector 𝐱\bf x, which is a set of parameters rather than the actual stimulus itself. Direct computation of the actual time-varying stimulus is not easy because no closed analytical form of the objective function is available and furthermore the computation of the optimal control input generally requires a backward integration or recursion. Instead of struggling with this difficulty, one can restrict the stimulus II to a well known natural form such as sum of phased cosines as shown below:

I=n=1NAncos(ωnt+ϕn)I=\sum_{n=1}^{N}A_{n}\cos\left(\omega_{n}t+\phi_{n}\right) (15)

where AnA_{n} is the amplitude, ωn\omega_{n} is the frequency of the nn-th Fourier component, and ϕn\phi_{n} is the phase of the component. We choose a base frequency ω1{\omega_{1}} and set the frequencies of all other components as the harmonics: ωn=nω1\omega_{n}=n\omega_{1} for n=1,,Nn=1,\ldots,N. Now the stimulus parameters can be summarized by the stimulus parameter vector:

𝐱=[A1,,AN,ϕ1,,ϕN]{\bf x}=[A_{1},\ldots,A_{N},\phi_{1},\ldots,\phi_{N}] (16)

We sometimes refer to 𝐱\bf x as the stimulus, with it understood that it really means a set of parameters that uniquely specify the actual stimulus II.

Some popular choices of the objective function are based on the Fisher information matrix, which is generally defined as:

Fij(𝐱,θ)=lnp(𝐫𝐱,θ)θilnp(𝐫𝐱,θ)θjF_{ij}\left({\bf x},\theta\right)=\left<\frac{\partial\ln p({\bf r}\mid{\bf x},\theta)}{\partial\theta_{i}}\frac{\partial\ln p({\bf r}\mid{\bf x},\theta)}{\partial\theta_{j}}\right> (17)

where p(𝐫𝐱,θ)p({\bf r}\mid{\bf x},\theta) is the probability distribution of the response 𝐫\bf r to a given stimulus 𝐱\bf x. One should here note that, the term p(𝐫𝐱,θ)p({\bf r}\mid{\bf x},\theta) is different from the terms p(t1,t2,,tK)p\left(t_{1},t_{2},\ldots,t_{K}\right) in (8) and p(Smθ)p\left(S_{m}\mid\theta\right) in (10). The latter two are derived from the spiking likelihood function presented in [26, 27] whereas the definition p(𝐫𝐱,θ)p({\bf r}\mid{\bf x},\theta) is derived from (6). However only the related results are provided in this text in order to save space. In (17), \left<\;\right> represents the weighted average over all possible responses 𝐫\bf r to all available sets of stimuli 𝐱\bf x. This average is calculated based on the response probability distribution p(𝐫𝐱,θ)p({\bf r}\mid{\bf x},\theta). Stimulus variable 𝐱\bf x can be represented by a group of parameters such as the Fourier series parameters in (15) which are An,ωnA_{n},\ \omega_{n} and ϕn\phi_{n}. The frequency parameter may or may not be fixed depending on the requirements or computational burden.

The Fisher information matrix reflects the amount of information contained in the noisy response 𝐫{\bf r} about the model parameters θ{\theta}, assuming a generative model given by the conditional probability p(𝐫𝐱,θ)p({\bf r}\mid{\bf x},\theta). So the stimulus designed by maximizing a certain measure of the Fisher information matrix (17) is expected to decrease the error of the estimation of the parameters θ{\theta}.

The utility function UU can be chosen as a scalar function of the Fisher information matrix 𝐅\bf F. One theoretically well founded popular choice is the D-optimal design:

U(𝐱,θ)=det𝐅(𝐱,θ)U({\bf x},\theta)=\det{\bf F}({\bf x},\theta) (18)

although the determinant of the Fisher information matrix is not always easy to optimize. The A-optimal design is based on the trace of the Fisher information matrix and is easier to optimize:

U(𝐱,θ)=tr𝐅(𝐱,θ)U({\bf x},\theta)={\rm tr\,}{\bf F}({\bf x},\theta) (19)

Another alternative is the E-optimal design where the objective function is the smallest eigenvalue of the Fisher information matrix. In this paper the A-optimality measure of the information matrix is preferred. There is an obvious reason for this preference. As the computational complexity of the optimization algorithms are expected to be high, the necessity of numerical derivative computation should be avoided as much as possible. Since it is not easy to evaluate the derivatives of the eigenvalues and determinants by any means other than numerical approximations it will be convenient to apply a criterion like A-optimality which has a direct relationship like the sums of the diagonal elements.

In the beginning of this section we mentioned that an optimally design stimulus is expected to depend on which parameter is supposed to be estimated. Since a scalar utility function in (18) or (19) depends on all the parameters θ=[θ1,θ2,]\theta=[\theta_{1},\theta_{2},\ldots], optimizing a single scalar function is sufficient to recover all the parameters. When a sequence of stimuli are generated by optimal design, the stimuli may sometimes alternate spontaneously as if the optimization is performed with respect to each of the parameters one by one [12].

In our network model, the recorded spike train has an inhomogeneous Poisson distribution with the rate function re(t){r_{e}\left(t\right)}. We write this rate as re(t,𝐱,θ)r_{e}\left(t,{\bf x},\theta\right) to emphasize that it is a time-varying function that depends on both the stimulus 𝐱\bf x and the network parameters 𝐱\bf x. For a small time window of duration Δt\Delta t and centered at time tt, the Fisher information matrix entry in (17) is reduced to:

Fij(t,𝐱,θ)=Δtre(t,𝐱,θ)re(t,𝐱,θ)θire(t,𝐱,θ)θjF_{ij}\left(t,{\bf x},\theta\right)=\frac{\Delta t}{r_{e}\left(t,{\bf x},\theta\right)}\frac{\partial{r_{e}}\left(t,{\bf x},\theta\right)}{\partial\theta_{i}}\frac{\partial{r_{e}}\left(t,{\bf x},\theta\right)}{\partial\theta_{j}} (20)

Since the Poisson rate function rer_{e} varies with time, the A-optimal utility function in (19) should be modified by including integration over time:

U(𝐱,θ)=0Ttr𝐅(t,𝐱,θ)𝑑t=0Tk=181re(t,𝐱,θ)(re(t,𝐱,θ)θk)2dtU\left({\bf x},\theta\right)=\int_{0}^{T}{\rm tr\,}{\bf F}(t,{\bf x},\theta)dt=\int_{0}^{T}\sum_{k=1}^{8}\frac{1}{r_{e}(t,{\bf x},\theta)}\left(\frac{\partial r_{e}(t,{\bf x},\theta)}{\partial\theta_{k}}\right)^{2}dt\\ (21)

Here the time window Δt\Delta t is ignored because it is a constant coefficient that does not affect the result of the optimization.

For convenience, we can also define the objective function with respect to a single parameter θk\theta_{k} as follows:

Uk(𝐱,θ)=0T1re(t,𝐱,θ)(re(t,𝐱,θ)θk)2𝑑tU_{k}\left({\bf x},\theta\right)=\int_{0}^{T}\frac{1}{r_{e}(t,{\bf x},\theta)}\left(\frac{\partial r_{e}(t,{\bf x},\theta)}{\partial\theta_{k}}\right)^{2}dt\\ (22)

The objective function in (21) is identical to k=18Uk\sum_{k=1}^{8}U_{k}.

The optimization of the D-optimal criterion in (18) is not affected by parameter rescaling, or changing the units of parameters. For example, changing the unit of parameter 1 (say, from msec-1 to sec-1) is equivalent to rescaling the parameter by a constant coefficient: θ1cθ1\theta_{1}\to c\theta_{1}. The effect of this transformation is equivalent to a rescaling of the determinant of the Fisher information matrix by a constant, namely, det𝐅(det𝐅)/c16\det{\bf F}\to(\det{\bf F})/c^{16}, which does not affect the location of the maximum of (18). By contrast, the criterion function in (19) or (21) are affected by parameter rescaling. A parameter with a smaller unit would tend to have larger derivative value and therefore contribute more to (21) than a parameter with a large unit. To alleviate this problem, we use UkU_{k} one by one to generate the stimuli. That is, stimulus 1 is generated by maximizing U1U_{1}, and stimulus 2 is generated by maximizing U2U_{2}, and so on. Once the 8th stimulus is generated by maximizing U8U_{8}, we go back and use U1U_{1} to generate the next stimulus, and so on. Finally, an alternative way to get rid of scale dependence is to introduce logarithm and use U=klnUkU=\sum_{k}\ln U_{k} as the criterion, which, however, may become degenerate when UkU_{k} approaches 0.

2.5 Gradient Computation

As just explained in the previous section about the computational issues in this research, the gradient computation decreases the computation durations considerably. The main issue with this fact is the lack of closed form expressions like in the case of static non-linear mappings as the model. In researches such as [33], [2] and [34] the gradients are computed as a self contained differential equation which is formed by taking the derivatives of the model equations (3) and (4) from both sides.

Compiling all the information in this section one can write the gradient of the Fisher Information Measure (i.e. the Fisher Information Matrix with a certain optimality criterion such as A-Optimality). In the beginning of this section, it is stated that the sensitivity levels of the firing rate w.r.to different network parameters are different and thus it would be convenient to maximize the fisher information for a single parameter at a time.

The optimization as expressed in (14) the optimal design problem is converted into a parameter optimization problem to optimize the amplitudes An{A_{n}}’s and ϕn{\phi_{n}}’s of the stimulus. For the sake of simplicity and modularity in programming, these equations can be written using their shorthand notations.

Let

𝐱=[x1,x2,,x2N]=[A1,,AN,ϕ1,,ϕN]{\bf x}=[x_{1},x_{2},\ldots,x_{2N}]=[A_{1},\ldots,A_{N},\phi_{1},\ldots,\phi_{N}] (23)

We write the state of the network as a vector: 𝐯=[Ve,Vi]T{\bf v}=[V_{e},V_{i}]^{\rm T}

and we need derivatives such as ddt𝐯𝐱\frac{d}{dt}\frac{\partial{\bf v}}{\partial\bf x}. Here, the idea is to use these derivatives as variables that can be solved directly from differential equations derived from the original dynamical equations in (5).

For the optimal design, one can write the following:

ddt𝐯𝐱=[βe00βi]{𝐯𝐱+[weeweiwiewii][ge(Ve)00gi(Vi)]𝐯𝐱+[wewi]I𝐱}\frac{d}{dt}\frac{\partial{\bf v}}{\partial\bf x}=\left[\begin{array}[]{cc}\beta_{e}&0\\ 0&\beta_{i}\end{array}\right]\left\{-\frac{\partial\bf v}{\partial\bf x}+\left[\begin{array}[]{cc}w_{ee}&-w_{ei}\\ w_{ie}&-w_{ii}\end{array}\right]\left[\begin{array}[]{cc}g^{\prime}_{e}\left(V_{e}\right)&0\\ 0&g^{\prime}_{i}\left(V_{i}\right)\end{array}\right]\frac{\partial\bf v}{\partial\bf x}+\left[\begin{array}[]{c}w_{e}\\ w_{i}\end{array}\right]\frac{\partial I}{\partial\bf x}\right\} (24)

where geg^{\prime}_{e} and gig^{\prime}_{i} are the derivatives of the gain functions geg_{e} and gig_{i} respectively, and the matrices derivatives are defined in the usual manner:

I𝐱=[IA1IAN,Iϕ1IϕN]\frac{\partial I}{\partial\bf x}=\left[\begin{array}[]{cccccc}\displaystyle\frac{\partial I}{\partial A_{1}}&\ldots&\displaystyle\frac{\partial I}{\partial A_{N}},&\displaystyle\frac{\partial I}{\partial\phi_{1}}&\ldots&\displaystyle\frac{\partial I}{\partial\phi_{N}}\end{array}\right] (25)

and

𝐯𝐱=[VeA1VeANVeϕ1VeϕNViA1ViANViϕ1ViϕN]\frac{\partial\bf v}{\partial\bf x}=\left[\begin{array}[]{cccccc}\displaystyle\frac{\partial V_{e}}{\partial A_{1}}&\ldots&\displaystyle\frac{\partial V_{e}}{\partial A_{N}}&\displaystyle\frac{\partial V_{e}}{\partial\phi_{1}}&\ldots&\displaystyle\frac{\partial V_{e}}{\partial\phi_{N}}\\ \mbox{}\\ \displaystyle\frac{\partial V_{i}}{\partial A_{1}}&\ldots&\displaystyle\frac{\partial V_{i}}{\partial A_{N}}&\displaystyle\frac{\partial V_{i}}{\partial\phi_{1}}&\ldots&\displaystyle\frac{\partial V_{i}}{\partial\phi_{N}}\end{array}\right] (26)

(24) can be written equivalently in the shorthand form:

ddt𝐯𝐱=𝐁{𝐯𝐱+𝐖𝐆𝐯𝐱+𝐰I𝐱}\frac{d}{dt}\frac{\partial\bf v}{\partial\bf x}={\bf B}\left\{-\frac{\partial\bf v}{\partial\bf x}+{\bf WG}\frac{\partial\bf v}{\partial\bf x}+{\bf w}\frac{\partial I}{\partial\bf x}\right\} (27)

where 𝐁=[βe00βi]{{\bf B}=\left[\begin{array}[]{cc}\beta_{e}&0\\ 0&\beta_{i}\end{array}\right]}, 𝐖=[weeweiwiewii]{{\bf W}=\left[\begin{array}[]{cc}w_{ee}&-w_{ei}\\ w_{ie}&-w_{ii}\end{array}\right]}, 𝐆=[ge(Ve)00gi(Vi)]{\bf G}=\left[\begin{array}[]{cc}g^{\prime}_{e}(V_{e})&0\\ 0&g^{\prime}_{i}(V_{i})\end{array}\right], and 𝐰=[wewi]{{\bf w}=\left[\begin{array}[]{c}w_{e}\\ w_{i}\end{array}\right]}.

In the evaluation of the Fisher Information Matrix and the gradients in estimation one will need the derivatives with respect to the parameters θ=[θ1,,θ8]=[βe,βi,we,wi,wee,wei,wie,wii]{\theta}=[\theta_{1},\ldots,\theta_{8}]=\left[\beta_{e},\beta_{i},w_{e},w_{i},w_{ee},w_{ei},w_{ie},w_{ii}\right].

It follows from the original dynamical equation in (5) that

ddt𝐯θk=𝐁{𝐯θk+𝐖𝐆𝐯θk}+𝐳k\frac{d}{dt}\frac{\partial\bf v}{\partial\theta_{k}}={\bf B}\left\{-\frac{\partial\bf v}{\partial\theta_{k}}+{\bf WG}\frac{\partial\bf v}{\partial\theta_{k}}\right\}+{\bf z}_{k} (28)

where 𝐯θk=[Veθk,Viθk]T\displaystyle{\frac{\partial\bf v}{\partial\theta_{k}}=\left[\frac{\partial V_{e}}{\partial\theta_{k}},\frac{\partial V_{i}}{\partial\theta_{k}}\right]^{\rm T}} and the last term 𝐳k{\bf z}_{k} refers to the extra components resulting from the chain rule of differentiation. These extra terms are presented in Table 1.

Table 1: The extra components 𝐳k{\bf z}_{k} in Equation (28)
kk Parameter θk\theta_{k} Extra term 𝐳k{\bf z}_{k} in Equation (28)
1 βe\beta_{e} [1000]{[VeVi]+𝐖[ge(Ve)gi(Vi)]+𝐰I}\left[\begin{array}[]{cc}1&0\\ 0&0\end{array}\right]\left\{-\left[\begin{array}[]{c}V_{e}\\ V_{i}\end{array}\right]+{\bf W}\left[\begin{array}[]{c}g_{e}\left(V_{e}\right)\\ g_{i}\left(V_{i}\right)\end{array}\right]+{\bf w}I\right\}
2 βi\beta_{i} [0001]{[VeVi]+𝐖[ge(Ve)gi(Vi)]+𝐰I}\left[\begin{array}[]{cc}0&0\\ 0&1\end{array}\right]\left\{-\left[\begin{array}[]{c}V_{e}\\ V_{i}\end{array}\right]+{\bf W}\left[\begin{array}[]{c}g_{e}\left(V_{e}\right)\\ g_{i}\left(V_{i}\right)\end{array}\right]+{\bf w}I\right\}
3 wew_{e} 𝐁[10]I{\bf B}\left[\begin{array}[]{c}1\\ 0\\ \end{array}\right]I
4 wiw_{i} 𝐁[01]I{\bf B}\left[\begin{array}[]{c}0\\ 1\\ \end{array}\right]I
5 weew_{ee} 𝐁[1000][ge(Ve)gi(Vi)]{\bf B}\left[\begin{array}[]{cc}1&0\\ 0&0\end{array}\right]\left[\begin{array}[]{c}g_{e}\left(V_{e}\right)\\ g_{i}\left(V_{i}\right)\end{array}\right]
6 weiw_{ei} 𝐁[0100][ge(Ve)gi(Vi)]{\bf B}\left[\begin{array}[]{cc}0&-1\\ 0&0\end{array}\right]\left[\begin{array}[]{c}g_{e}\left(V_{e}\right)\\ g_{i}\left(V_{i}\right)\end{array}\right]
7 wiew_{ie} 𝐁[0010][ge(Ve)gi(Vi)]{\bf B}\left[\begin{array}[]{cc}0&0\\ 1&0\end{array}\right]\left[\begin{array}[]{c}g_{e}\left(V_{e}\right)\\ g_{i}\left(V_{i}\right)\end{array}\right]
8 wiiw_{ii} 𝐁[0001][ge(Ve)gi(Vi)]{\bf B}\left[\begin{array}[]{cc}0&0\\ 0&-1\end{array}\right]\left[\begin{array}[]{c}g_{e}\left(V_{e}\right)\\ g_{i}\left(V_{i}\right)\end{array}\right]

The maximization of the Fisher Information Measure in trace form requires its gradients and they involve second order cross derivatives of the membrane potentials with respect to parameters (θk){\left(\theta_{k}\right)} and stimulus parameters (Il){\left(I_{l}\right)}.

Taking derivative of (27) with respect to θk\theta_{k}, we find:

ddt2𝐯𝐱θk=𝐁{2𝐯𝐱θk+𝐖𝐆2𝐯𝐱θk+𝐖𝐆diag(𝐯θk)𝐯𝐱}+𝐙k\frac{d}{dt}\frac{\partial^{2}\bf v}{\partial{\bf x}\partial\theta_{k}}={\bf B}\left\{-\frac{\partial^{2}\bf v}{\partial{\bf x}\partial\theta_{k}}+{\bf WG}\frac{\partial^{2}\bf v}{{\partial\bf x}\partial\theta_{k}}+{\bf WG^{\prime}}\,{\rm diag}\!\left(\frac{\partial\bf v}{\partial\theta_{k}}\right)\frac{\partial\bf v}{\partial\bf x}\right\}+{\bf Z}_{k} (29)

where 𝐆=[ge′′(Ve)00gi′′(Vi)]{\bf G}^{\prime}=\left[\begin{array}[]{cc}g^{\prime\prime}_{e}(V_{e})&0\\ 0&g^{\prime\prime}_{i}(V_{i})\end{array}\right], diag(𝐯θk)=[Veθk00Viθk]\displaystyle{\rm diag}\!\left(\frac{\partial\bf v}{\partial\theta_{k}}\right)=\left[\begin{array}[]{cc}\frac{\partial V_{e}}{\partial\theta_{k}}&0\\ 0&\frac{\partial V_{i}}{\partial\theta_{k}}\end{array}\right], and

2𝐯𝐱θk=[2VeA1θk2VeANθk2Veϕ1θk2VeϕNθk2ViA1θk2ViANθk2Viϕ1θk2ViϕNθk]\frac{\partial^{2}\bf v}{\partial{\bf x}\partial\theta_{k}}=\left[\begin{array}[]{cccccc}\displaystyle\frac{\partial^{2}V_{e}}{\partial A_{1}\partial\theta_{k}}&\ldots&\displaystyle\frac{\partial^{2}V_{e}}{\partial A_{N}\partial\theta_{k}}&\displaystyle\frac{\partial^{2}V_{e}}{\partial\phi_{1}\partial\theta_{k}}&\ldots&\displaystyle\frac{\partial^{2}V_{e}}{\partial\phi_{N}\partial\theta_{k}}\\ \mbox{}\\ \displaystyle\frac{\partial^{2}V_{i}}{\partial A_{1}\partial\theta_{k}}&\ldots&\displaystyle\frac{\partial^{2}V_{i}}{\partial A_{N}\partial\theta_{k}}&\displaystyle\frac{\partial^{2}V_{i}}{\partial\phi_{1}\partial\theta_{k}}&\ldots&\displaystyle\frac{\partial^{2}V_{i}}{\partial\phi_{N}\partial\theta_{k}}\end{array}\right] (30)

which is compatible with (26). The last term 𝐙k{\bf Z}_{k} is specified in Table 2

Table 2: The extra components 𝐙k{\bf Z}_{k} in Equation (29)
kk Parameter θk\theta_{k} Extra term 𝐙k{\bf Z}_{k} in Equation (29)
1 βe\beta_{e} [1000]{𝐯𝐱+𝐖𝐆𝐯𝐱+𝐰I𝐱}\left[\begin{array}[]{cc}1&0\\ 0&0\end{array}\right]\displaystyle\left\{-\frac{\partial\bf v}{\partial\bf x}+{\bf WG}\frac{\partial\bf v}{\partial\bf x}+{\bf w}\frac{\partial I}{\partial\bf x}\right\}
1 βi\beta_{i} [0001]{𝐯𝐱+𝐖𝐆𝐯𝐱+𝐰I𝐱}\left[\begin{array}[]{cc}0&0\\ 0&1\end{array}\right]\displaystyle\left\{-\frac{\partial\bf v}{\partial\bf x}+{\bf WG}\frac{\partial\bf v}{\partial\bf x}+{\bf w}\frac{\partial I}{\partial\bf x}\right\}
3 wew_{e} 𝐁[10]I𝐱{\bf B}\left[\begin{array}[]{c}1\\ 0\end{array}\right]\displaystyle\frac{\partial I}{\partial\bf x}
4 wiw_{i} 𝐁[01]I𝐱{\bf B}\left[\begin{array}[]{c}0\\ 1\end{array}\right]\displaystyle\frac{\partial I}{\partial\bf x}
5 weew_{ee} 𝐁[1000]𝐆𝐯𝐱{\bf B}\left[\begin{array}[]{cc}1&0\\ 0&0\end{array}\right]\displaystyle{\bf G}\frac{\partial\bf v}{\partial\bf x}
6 weiw_{ei} 𝐁[0100]𝐆𝐯𝐱{\bf B}\left[\begin{array}[]{cc}0&-1\\ 0&0\end{array}\right]\displaystyle{\bf G}\frac{\partial\bf v}{\partial\bf x}
7 wiew_{ie} 𝐁[0010]𝐆𝐯𝐱{\bf B}\left[\begin{array}[]{cc}0&0\\ 1&0\end{array}\right]\displaystyle{\bf G}\frac{\partial\bf v}{\partial\bf x}
8 wiiw_{ii} 𝐁[0001]𝐆𝐯𝐱{\bf B}\left[\begin{array}[]{cc}0&0\\ 0&-1\end{array}\right]\displaystyle{\bf G}\frac{\partial\bf v}{\partial\bf x}

To compute the gradients needed for optimizing the objective function based on Fisher Information or for performing maximum-likelihood estimation, we need to evaluate the derivatives of the mean firing rate re=ge(Ve){r_{e}=g_{e}\left(V_{e}\right)} with respect to the network parameters θk\theta_{k} in (9) or the stimulus parameters xjx_{j} in (23). The first and the second order derivatives are:

reθk=ge(Ve)Veθk\frac{\partial{r_{e}}}{\partial{\theta_{k}}}=g^{\prime}_{e}\left(V_{e}\right)\frac{\partial{V_{e}}}{\partial{\theta_{k}}} (31)
rexl=ge(Ve)Vexl\frac{\partial{r_{e}}}{\partial{x_{l}}}=g^{\prime}_{e}\left(V_{e}\right)\frac{\partial{V_{e}}}{\partial{x_{l}}} (32)
2rexlθk=ge′′(Ve)VexlVeθk+ge(Ve)2Vexlθk\frac{\partial^{2}r_{e}}{\partial x_{l}\partial\theta_{k}}=g^{\prime\prime}_{e}\left(V_{e}\right)\frac{\partial V_{e}}{\partial x_{l}}\frac{\partial V_{e}}{\partial\theta_{k}}+g^{\prime}_{e}\left(V_{e}\right)\frac{\partial^{2}V_{e}}{\partial x_{l}\partial\theta_{k}} (33)

These formulas are expressed in terms of the derivatives Veθk\frac{\partial{V_{e}}}{\partial{\theta_{k}}}, Vexl\frac{\partial{V_{e}}}{\partial{x_{l}}} and 2Vexlθk\frac{\partial^{2}V_{e}}{\partial x_{l}\partial\theta_{k}}, which are regarded as dynamical variables that can be solved from the three differential equations (27)–(29). Here the initial conditions were always assumed to be the equilibrium state. The initial values of the derivatives can be set to zero as recommended in [2].

The gradient of the Fisher Information Measure in (21) with respect to the stimulus parameters can be written as

Uxl\displaystyle\frac{\partial U}{\partial x_{l}} =\displaystyle= 0Txlk=181re(reθk)2dt\displaystyle\int_{0}^{T}\frac{\partial}{\partial x_{l}}\sum_{k=1}^{8}{\frac{1}{r_{e}}\left(\frac{\partial{r_{e}}}{\partial{\theta_{k}}}\right)^{2}}dt (34)
=\displaystyle= 0Tk=18{1re2rexl(reθk)2+2rereθk2rexlθk}dt\displaystyle\int_{0}^{T}\sum_{k=1}^{8}\left\{-\frac{1}{r_{e}^{2}}\frac{\partial r_{e}}{\partial x_{l}}\left(\frac{\partial r_{e}}{\partial\theta_{k}}\right)^{2}+\frac{2}{r_{e}}\frac{\partial r_{e}}{\partial\theta_{k}}\frac{\partial^{2}r_{e}}{\partial x_{l}\partial\theta_{k}}\right\}dt (35)

which the last expression is written in terms of the derivatives that are already evaluated by equations (31)–(33). If the Fisher Information is computed with respect to one parameter at a time as shown in (22), one rewrite the above by removing the summation as:

Ukxl\displaystyle\frac{\partial U_{k}}{\partial x_{l}} =\displaystyle= 0Txl1re(reθk)2𝑑t\displaystyle\int_{0}^{T}\frac{\partial}{\partial x_{l}}{\frac{1}{r_{e}}\left(\frac{\partial{r_{e}}}{\partial{\theta_{k}}}\right)^{2}}dt (36)
=\displaystyle= 0T{1re2rexl(reθk)2+2rereθk2rexlθk}𝑑t\displaystyle\int_{0}^{T}\left\{-\frac{1}{r_{e}^{2}}\frac{\partial r_{e}}{\partial x_{l}}\left(\frac{\partial r_{e}}{\partial\theta_{k}}\right)^{2}+\frac{2}{r_{e}}\frac{\partial r_{e}}{\partial\theta_{k}}\frac{\partial^{2}r_{e}}{\partial x_{l}\partial\theta_{k}}\right\}dt (37)

Lastly, for maximum likelihood estimation, one need the gradient of the log likelihood function of spike trains in (12):

lθk=m=1M0Tre(m)(t)θk(t)𝑑t+m=1Mk=1Kmre(m)(tk(m))1re(m)(tk(m))θk\frac{\partial{l}}{\partial{\theta_{k}}}=-\sum_{m=1}^{M}\int_{0}^{T}\frac{\partial{r_{e}^{(m)}}\!\left(t\right)}{\partial{\theta_{k}}}\!\left(t\right)dt+\sum_{m=1}^{M}\sum_{k=1}^{K_{m}}{{r_{e}^{(m)}}\!\left(t_{k}^{(m)}\right)}^{-1}\frac{\partial{r_{e}^{(m)}}\!\left(t_{k}^{(m)}\right)}{\partial{\theta_{k}}} (38)

which is written in terms of derivatives that can be evaluated by (31).

2.6 Other Numerical Issues Related to Optimization

Up to this point, the theoretical grounds of this research are presented. In order to achieve the results one needs two separate maximization algorithms targeting (13) and (14). Optimization algorithms have varieties in certain aspects. One of these classifications is the gradient requirements. The global optimization algorithms such as genetic algorithms (MATLAB® ga), simulated annealing (MATLAB® simulannealbnd) and pattern search (MATLAB® patternsearch) do not require gradient computations. However due to their stochastic nature, they do a lot of computations which increase the computation duration drastically. In addition, that stochastic nature leads to different results at the end of different runs. Disabling this randomness, might have adverse effects on the optimization performance and their ability of finding a global optimum solution. In order not to be trapped in these issues we will stick to gradient based optimization routines such as constrained interior-point gradient descent. MATLAB® provides these algorithms through its optimization toolbox. Interior point and similar methods are available in fmincon function. We can write the following facts about the optimization algorithm:

  • 1.

    MATLAB®’s fmincon allows the user to set constraints on the solution.In the optimal design the stimulus amplitudes An{A_{n}} will have an upper bound (numerical details will be discussed Section 3.2. The stimulus amplitude is expected to tend to the upper bound. The parameter estimation process will also be benefited from similar bounds as they are assigned to be positive in (5). In addition assignment of bounds on parameters will aid in prevention of overstimulation of the model.

  • 2.

    fmincon and similar algorithms are local optimizers. In order to find a good optimum, the algorithms are often repeated with multiple initial guesses and the best one is chosen according to the value of the objective and gradient value at the termination point. This is especially important in the optimal design part as Fisher Information measure in (21) may have lots of local minima. The same factors do exist for likelihood function (12) however as the number of samples M{M} increases the likelihood function tend to converge to the same optimum for different initial guesses. See Section 3.3 for a detailed discussion of this issue.

  • 3.

    The local optimization algorithms need the gradient of the objective functions. In MATLAB®’s fmincon and other similar packages have the ability to compute the gradients numerically. However this capability may not be an advantage in this sort of complicated problems for two main reasons. First of all, numerical gradient computation increases the risk of singularities in the solution. Secondly, the computational duration increases dramatically. It is often noted that, the numerical differentiation in fmincon at least doubles the computational duration. This is unacceptable in the context of this research.

  • 4.

    Longer simulation times due to the computational complexity of the overall problem can be eliminated to a certain extent (about the 5 times shorter duration) by utilizing the MATLAB®’s parallel computation facilities (such as replacing standard for loops by parfor loops).

  • 5.

    The MATLAB® version used in this research is R2013b and the fmincon version included in this package has the interior-point method of constrained non-linear optimization by default. This is not the only method available in fmincon and it can always be modified by changing the algorithm option. However, the interior-point method appears to be the most efficient among all available options.

2.7 Procedural Information

In this section, we will summarize the overall procedure to give an insight about how the optimal design and parameter estimation algorithm works together in a automatized loop. Before giving a start, it is worth to discuss the initial parameter problem. In the beginning of an experiment (or simulation in our case), one usually has no idea about the true or approximate values of the network parameter vector θ\theta. However, the optimal design always needs a current estimate. Thus one should assign a randomly chosen initial parameter vector. A good rule to get that value is to draw one sample from a parametric subspace of which members are uniformly distributed between the lower and upper bounds of the parameters. One can find the details about the parametric bounds in Table 3 Setting a parametric bound on the stimulus amplitude coefficients (AnA_{n} in (15)) might be critical. Too large amplitudes may lead to instabilities or very large responses which may break the optimal design procedures. Because of this fact, an upper bound on AnA_{n}’s will most often required. A few number of initial simulations will be required to determine a nominal value for those bounds. If no instabilities are detected one will be fine with the decided value of the bounds on AnA_{n} (which is termed as AmaxA_{\max}) and can continue the optimization with these bounds. Mostly, the optimization procedure will fail when the assigned value of the bounds are too large.

The algorithmic details of the overall process are summarized as shown below:

  1. 1.

    Set i=1i=1

  2. 2.

    Generate a random parameter vector θ0\theta^{0} from a uniform distribution between θmin\theta_{min} and θmax\theta_{max} (from Table 3).

  3. 3.

    Set θ^=θ0\hat{\theta}=\theta^{0}

  4. 4.

    Set k=1k=1

  5. 5.

    Optimize UkU_{k} based on θ^\hat{\theta} to generate stimulus 𝐱\bf x

  6. 6.

    Using the stimulus 𝐱\bf x perform a maximum likelihood estimation to find a new estimate of θ\theta and replace θ^\hat{\theta} by the new estimate.

  7. 7.

    Set kk+1k\rightarrow{k+1} and go to Step 5 . If k>8k>8 set ii+1i\rightarrow{i+1}

  8. 8.

    If i>Nitri>N_{itr} stop and report the result as θ^final=θ^\hat{\theta}^{final}=\hat{\theta} otherwise go to Step 4

The results related to the application of all theoretical and algorithmic developments in the Sections 2.1 to 2.7 will be presented and discussed in Sections 3 and 4.

3 Results

In this section, we will summarize the functional and numerical details of the combined optimal design - parameter estimation algorithm and present the results in comparison with the random stimuli tests. This section is divided into several sections discussing the numerical details of an example application, statistical properties of the optimal stimuli, accuracy of the parameter estimation and parameter confounding phenomenon.

3.1 Details of the example problem

This section is devoted to the detailed presentation of the simulation set-up. An numerical example will be presented which will demonstrate our optimal design approach. In the example application, the algorithms presented in Sections 2.3 and 2.4 are applied to probe an EI network.

In order to verify the performance of the parameter estimation we have to compare the estimates with their true values. So we will need a set of reference values of the model parameters in (5). These are shown in Table 3.

Table 3: The true values, lower and upper bounds of parameters of model in (5). Mean values and standard deviations of the estimates for the case M=120M=120 are also shown for convenience.
Parameter βe{\beta_{e}} (1/s\nicefrac{{1}}{{s}}) βi{\beta_{i}} (1/s\nicefrac{{1}}{{s}}) we{w_{e}} (kΩ\Omega) wi{w_{i}} (kΩ\Omega) wee{w_{ee}} (mV\cdots) wei{w_{ei}} (mV\cdots) wie{w_{ie}} (mV\cdots) wii{w_{ii}} (mV\cdots)
True value (θ)\left(\theta\right) 50{50} 25{25} 1.0{1.0} 0.7{0.7} 1.2{1.2} 2.0{2.0} 0.7{0.7} 0.4{0.4}
Lower bound (θmin)\left(\theta_{min}\right) 0 0 0 0 0 0 0 0
Upper bound (θmax)\left(\theta_{max}\right) 100 100 2 2 3 3 3 3
Mean (OED) 49.936249.9362 25.184225.1842 1.00161.0016 0.69750.6975 1.22671.2267 2.06902.0690 0.71200.7120 0.45620.4562
STD (OED) 0.86200.8620 1.43731.4373 0.03570.0357 0.06780.0678 0.06900.0690 0.18720.1872 0.09490.0949 0.19870.1987
Mean (RAND) 50.068550.0685 25.226525.2265 0.99790.9979 0.71490.7149 1.24691.2469 2.07142.0714 0.78440.7844 0.56940.5694
STD (RAND) 1.62711.6271 1.96011.9601 0.03940.0394 0.09960.0996 0.10660.1066 0.22710.2271 0.17080.1708 0.37350.3735

Our model in (5) has two more important components which are the gain functions ge(Ve)g_{e}\left(V_{e}\right) and gi(Vi)g_{i}\left(V_{i}\right). These are obtained by setting jj in (2) by either ’ee’ or ’ii’. So one has 66 additional parameters [Γe,ae,he,Γi,ai,hi]\left[\Gamma_{e},a_{e},h_{e},\Gamma_{i},a_{i},h_{i}\right] which have direct effect on the neural model behaviour. This research targeted the estimation of the network parameters only. Because of that, the parameters of the gain functions are kept as fixed and they have the values Γe=100,ae=0.04,he=70,Γi=50,ai=0.04,hi=35\Gamma_{e}=100,a_{e}=0.04,h_{e}=70,\Gamma_{i}=50,a_{i}=0.04,h_{i}=35.

This set of parameters (gain functions and Table 3) allows the network to have a unique equilibrium state for each stationary input. To demonstrate the excitatory and inhibitory characteristics of our model, we can stimulate the model with a square wave (pulse) stimulus as shown in Figure 2A. The resultant excitatory and inhibitory neural membrane potential responses (Ve(t)V_{e}\left(t\right) and Vi(t)V_{i}\left(t\right)) are shown in Figure 2B and Figure 2C. It can be said that, the network has shown both transient and sustained responses. In Figure 2D, the excitatory firing rate response re(t)r_{e}\left(t\right) which is related to excitatory potential as re(t)=ge(Ve(t))r_{e}\left(t\right)=g_{e}\left(V_{e}\left(t\right)\right) is shown. The response Vi(t)V_{i}\left(t\right) is slightly delayed which leads to the depolarization of excitatory unit until t=250t=250ms. This delay is also responsible from the subsequent re-polarization and plateau formation in the membrane potential of excitatory neuron. The firing rate re(t)r_{e}\left(t\right) is higher during excitation and lower in subsequent plateau and repolarization phases (Figure 2D).

Refer to caption
Figure 2: The network model in Figure 1B in response to a square-wave stimulus. The states of the excitatory and inhibitory units, VeV_{e} and ViV_{i}, are shown, together with the continuous firing rate of the excitatory unit, re=ge(Ve)r_{e}=g_{e}\left(V_{e}\right). The firing rate of the excitatory unit (bottom panel) has a transient component with higher firing rates, followed by a plateau or sustained component with lower firing rates.

The optimisation of the stimuli requires that the maximum power level in a single stimulus is bounded. This is a precaution to protect the model from potential instabilities due to over-stimulation. In addition, if applied in a real environment the experiment subject will also be protected from such over-stimulations. As the amplitude parameter is assumed positive, assigning an upper bound defined as AmaxA_{max} should be enough. This is applied to all stimulus amplitudes (i.e. AnA_{n}). In this research, a fixed setting of Amax=120μA_{max}=120\ \muA is chosen. The lower bound is obviously Amin=0A_{min}=0. For the phase ϕn\phi_{n}, no lower or upper bounds are necessary as the cosine function itself has already been bounded as (1cos1)\left(-1\leq\cos\leq 1\right). The frequencies of the stimulus components are kthk^{t}h harmonics of a base frequency fbasef_{base} (ωk=2πk×fbase)\left(\omega_{k}=2\pi{k}\times{f_{base}}\right). Since we have a simulation time of Topt=3T_{opt}=3 seconds, we have a reasonable choice of 3.333.33 Hz which is in fact equal to 103\frac{10}{3}. So we have chosen an integer relationship between the stimulation frequency and simulation time. The number of stimulus components NN is chosen as N=5N=5 which is found to be reasonable concerning speed and performance balance. These are reasonable choices specifically for this research because utmost importance is first given to computational performance. The first few simulations showed that choosing NN at a higher setting then N=5N=5 leads to longer simulation durations which are not desirable. In addition there seems no significant advantage of a larger setting for NN. For the base frequency fbasef_{base}, few different values other than fbase=3.33f_{base}=3.33 Hz are tried (fbase=1f_{base}=1 Hz, fbase=5f_{base}=5 Hz, fbase=10f_{base}=10 Hz) but the optimization worked best at the mentioned frequency.

It is well known that optimization algorithms such as fmincon requires an initial guess of the optimum solution. A suitable choice for the initial guesses can be their assignment from a set of initial conditions generated randomly between the optimization bounds. In the optimization of stimuli, the initial amplitudes can be uniformly distributed between [0,Amax]\left[0,A_{max}\right] and phases can be uniformly distributed between [π,π]\left[-\pi,\pi\right]. Although we do not have any constraints on the phase parameter, we limit the initial phase values to a safe assumed range. In the analysis of the optimal phase results we will wrap the resultant phase values to a range [π,π]\left[-\pi,\pi\right] using modulo function. Thus assuming an initial range in the same interval will be meaningful.

We follow a similar strategy for the parameter estimation based on maximum likelihood method. The multiple initial guesses will be chosen from a set of values uniformly distributed between the lower and upper bounds defined in Table 3.

In Section 2.3, one recall from (11) that the likelihood estimation should produce better results when the number of samples (i.e. MM in (11)) increases. Because of this fact, the likelihood function is based on data having all spikes generated since the beginning of the simulation. The number of repeats determines MM. If simulation is repeated NitrN_{itr} times (NitrN_{itr} iterations), one will have an MM value of M=8NitrM=8N_{itr} due to the fact that each iteration has 88 optimal designs sub-steps (with respect to each parameter θk\theta_{k}. Read Section 2.4). So, if one has 15 iterations Nitr=15N_{itr}=15, M=120M=120 which means likelihood has 120120 samples. This also means that optimal design and subsequent parameter estimation will also be repeated 120120 times.

3.2 Statistics of optimally designed stimuli

Having all necessary information from Section 3.1, one can perform an optimal design and obtain a sample optimal stimulus and associated neural responses Figure 3. It is noted that the optimal stimulus in Figure 3 top panel has a periodic variation as it is modelled as a phased cosines form (as it is equivalent to real valued Fourier series).

Refer to caption
Figure 3: An example of an optimally designed stimulus with a duration of 3 sec (top panel), . The responses of the excitatory and the inhibitory units in the network are shown below, followed by an example of spike trains generated by an inhomogeneous Poisson process according to the continuous firing rate of the excitatory unit (re)\left(r_{e}\right). Driven by the same stimulus, the response of eight derivatives variables, namely, the derivatives of the firing rate rer_{e} with respect to all the network parameters, are shown as red curves. These derivatives were solved directly from differential equations (27). The sensitivity derivatives shown in this figure above are evaluated at the true values of parameters shown in Table 3. This evaluation is performed to show the level differences between the sensitivity of response to each individual parameter.

In addition to the fundamental responses, the second half of Figure 3 displays the variation of the parametric sensitivity derivatives reθi\frac{\partial{r_{e}}}{\partial{\theta_{i}}} which are generated by integrating (28) and then substituting to (31). The variation of the sensitivity derivatives support the idea of optimization of the Fisher Information Metric with respect to a single parameter (see (22)) as the sensitivity (or gradient) of the Fisher Information Metric in (22) with respect to a single parameter θk\theta_{k} varies widely from parameter to parameter. . It appears that, the weight parameters have a very high sensitivity which are more than 1010 times that of the β\beta_{*} parameters. Also some of the parameters affecting the behaviour of inhibitory (I) unit βi,wi\beta_{i},w_{i} and also the E-I interconnection weights wei,wiew_{ei},w_{ie} have a reverse behaviour. This fact appears as a negative values variation of the sensitivity derivatives. Self inhibition coefficient wiiw_{ii} does not show this behaviour as it represent an inhibitory effect on the inhibitory neuron potential (which favours excitation).

One of the most distinguishing result related to optimal design is the statistics of the optimal stimuli (i.e. the amplitudes AnA_{n} and phases ϕn\phi_{n}). This analysis can be performed by generating the histograms of AnA_{n} and ϕn\phi_{n} from available data as shown in Figure 4. The Figure 4A shows the flat uniform distribution of the random stimuli amplitudes and phases. This is expected as the stimuli is generated directly from a uniform distribution. Concerning the optimal stimuli, the histograms shown in Figure 4B reveals that optimal design has a tendency to maximize the amplitude of the stimuli towards the upper bound. This reassures that optimal design tends to maximize the stimulus power which is expected increase the efficiency of the parameter estimation process and it distinguishes optimal stimuli from their randomly generated counterparts.

The results mentioned above coincide the findings of [35]. Here, what happens is related to the topological boundary property [36] associated with the optimal design results. It is found in [35] that, maximum firing rate response is always lies on the topological boundary of the collection of all allowable stimuli provided that the neurons have increasing gain functions and convergent synaptic connections between layers. Another interpretation of this result can be expressed as follows: The maximum and minimum responses of each individual neuron should arise from the stimulus boundary (from all available sets of stimuli) and the entire boundary of the pattern of responses elicited in a particular layer should also a result of the stimulation in the boundary level. This result is valid regardless of the neuron being feed-forward or recurrent. For recurrent neural networks we assume that the network is stable. Some mathematical proofs related to the stated results are also presented in [35]. In Figure 4B, the amplitudes of the Fourier stimulus components are collected at the boundary level which is Amax=120A_{\max}=120. This is the topological boundary of the set of stimuli defined by (15).

Refer to caption
Figure 4: Histograms of the stimulus Fourier amplitudes A1,,A5A_{1},\ldots,A_{5} and phases ϕ1,ϕ5\phi_{1},\ldots\phi_{5}. (A) Random stimuli were generated by choosing their Fourier amplitudes and phases randomly from uniform distributions. A total of 12,000 random stimuli were used in each plot. (B) The optimally designed stimuli showed some structures in the distributions of their Fourier amplitudes and phases, which differ radically from a uniform distribution. A total of 12,000 optimally designed stimuli were used in each plot.

Given a dataset consisting of stimulus-response pairs, we can always use maximum-likelihood estimation to fit a model to the data to recover the parameters. Maximum-likelihood estimation is known to be asymptotically efficient in the limit of large data size, in the sense that the estimation is asymptotically unbiased (i..e, average of the estimates approaches the true value) and has minimal variance (i.e., the variance of the estimates approaches the Cramér-Rao lower bound).

We found that maximum likelihood obtained from the optimally design stimuli was always much better than that obtained from the random stimuli (see Figure 5). It also reveals that, the likelihood value increases as the number of stimuli increases. For any given number of stimuli, the optimally designed stimuli always yielded much greater likelihood value than the random stimuli. The minimum difference between the likelihood values (the minimum from the optimal design and the maximum from the random stimuli based test) was typically about two times greater than the standard deviation of either estimates except for the case with 2424 samples (Nitr=3&M=24)\left(N_{itr}=3\&M=24\right). Even in this case, this violation appear only on one sample. In addition, it can be easily deduced from the box diagram in Figure 5 that the difference between 7575th and 2525th percentiles (33st and 11rd quartiles) yield a value which is larger than three times the standard deviation of either estimate. The standard deviations of the maximum likelihood values are also larger in the random stimuli based tests. Those results are certain evidences of the superiority of an optimal design over the random stimuli based tests. The difference between the maximum values of the two likelihoods (from optimal and random stimuli) becomes more significant as the number of samples MM increases.

Refer to caption
Figure 5: Optimally designed stimuli yield greater likelihood values for maximum-likelihood parameter estimation compared to random stimuli. (A) Histogram of the optimized likelihood values for 100 trials with random stimuli is compared with that for 100 trails with optimally designed stimuli. Note that the best value from the random trials was not even close to the worst value from the optimal design trials. Here each trial contained a sequence of 120 stimuli, which were generated either randomly or by optimal design. Each likelihood value was obtained by maximizing the likelihood function in equation xxx using the response data elicited by all 120 stimuli. (B) As the number of stimuli increases, the likelihood function also increases, following an approximate linear relationship. The optimal design yielded better likelihood values than random stimuli regardless of the number of stimuli. The boxplot shows 25%, 50% (median) and 75% percentiles, with the whiskers outside of the boxes indicating the minimum and maximum values. The straight lines were obtained by linear regression on the median values.

It would also be convenient to stress the fact that the greater the likelihood, the better the fitting of the data to the model. This fact is demonstrated by two regression lines imposed on the box diagram Figure 5. One of those lines correspond to the optimal design and the other correspond to the random stimuli based tests. Both regression lines path through the origin point (0, 0) approximately. This means that the ratio of the log likelihood values in the two cases is approximately a constant, regardless of the number of stimuli. The regression lines are represented by equations l=257.0896M+76.1500l=257.0896M+76.1500 for optimal and 174.5375M+31.1000174.5375M+31.1000 for the random stimuli. So the two lines have a slope ratio of approximately 1.47301.4730. The significance of this number can be explained by a simple example. If one desires to attain the same level of likelihood with M=120M=120 optimally designed stimuli, the required number of random stimuli to be generated is equal to a value about 180180.

Another statistical comparison of the maximized likelihoods can be performed by Wilcoxon rank-sum tests [37, 38, 39]. When applied, one will be able to see the difference which is highly significant. Regardless of the number of samples the p-values remained at least 103010^{-30} times smaller than the widely accepted probability significance threshold of p=0.05p=0.05 or 5%5\%.

The likelihood function provides an overall measure of how well a model fits the data. We have also tested the mean errors of individual parameters relative to their true values. The main finding is that, for each individual parameter, the error is typically smaller for the optimally designed stimuli than the error for the random stimuli. This result can be observed from the bar charts presented in Figure 6. The heights of the bars show the mean error levels for randomly and optimally generated stimuli respectively. One can get the benefits of the rank-sum test on the statistical properties of the parameter estimates. Computation of rank-sum p-values for each individual parameter corresponding to the case of 120120 samples (M=120)\left(M=120\right) yields:

p(βe)=4.4711×105p(βi)=9.0988×103p(we)=9.8928×101p(wi)=6.8591×103p(wee)=1.9874×103p(wei)=5.5709×103p(wie)=2.2302×105p(wii)=2.8541×108\begin{array}[]{c}p\left(\beta_{e}\right)=4.4711\times 10^{-5}\\ p\left(\beta_{i}\right)=9.0988\times 10^{-3}\\ p\left(w_{e}\right)=9.8928\times 10^{-1}\\ p\left(w_{i}\right)=6.8591\times 10^{-3}\\ \ p\left(w_{ee}\right)=1.9874\times 10^{-3}\\ p\left(w_{ei}\right)=5.5709\times 10^{-3}\\ p\left(w_{ie}\right)=2.2302\times 10^{-5}\\ p\left(w_{ii}\right)=2.8541\times 10^{-8}\end{array} (39)

The above result showed that for M=120M=120 the differences are statistically significant for 77 parameters. For different values one can refer to Figure 6. In this illustration, the statistical significance of the difference between optimal and random stimuli based tests are indicated as an asterisk placed above the bars. Any of the cases with asterisk means that, the associated sample size led to a result where optimal design is significantly better.

Refer to caption
Figure 6: Errors of individual network parameters obtained by maximum likelihood estimation tended to decrease as the number of stimuli increased. The optimal design yielded smaller errors for all parameters in all the cases tested, although the differences were not always statistically significant.An asterisk (*) at the top of an optimal design bar indicates that the difference from the neighboring random bar is statistically significant at the level p<0.05p<0.05 in the ranksum test.

Here an interesting result is the statistical non-significance of the third parameter wew_{e} regardless of the sample size MM. This might be associated with the parameter confounding phenomenon that is discussed in Section 3.4.

For convenience, one can see the mean values and standard deviations of the estimates obtained from 100100 optimal and random stimuli in Table 3. These are obtained with M=120M=120 samples. In general, mean values seem to be comparable for both stimuli however the standard deviations of estimates from optimal stimuli are smaller then that obtained from random stimuli. The mean values of parameters wiew_{ie} and wiiw_{ii} are also closer to the true values when obtained from optimal stimuli.

3.3 Problem of local maxima during optimization

We need optimization in two places: maximum likelihood estimation of parameters, and optimal design of stimuli. Due to speed and computational complexity considerations one needs to utilize gradient base local optimizers such as MATLAB® fmincon. These algorithms often needs initial guesses and not all of the initial guesses will converge to a true value. This issue may especially appear in the cases where the objective function involves dynamical model (differential or difference equations). In such cases, the problem of multiple local maxima might occur in the objective function which often requires multiple initial guesses to be provided to the solver. So an analysis on this issue may reveal useful information.

Suppose there are nn repeats or starts from different initial values. Let pp be the probability of finding the “correct” solution in an individual run with random initial guess. Then in nn repeated runs, the probably that at least one run will lead to the “correct” solution is

Prob (“correct”)=1(1p)n\mbox{Prob (``correct'')}=1-(1-p)^{n} (40)

The probability pp can be estimated from a pairwise test or directly from the values of the likelihood and Fisher Information Metric. In the test of the likelihood function, one can achieve the goal by starting the optimization from KK different initial guesses and checking the number of solutions which stay in an error bound 10%10\% for each individual parameter with respect to the solution leading to the highest likelihood value. In other words, to pass the test the following criterion should be satisfied for each individual parameter θ=θi\theta=\theta_{i} in (9):

|θ^bestθ^|θ10%\frac{\left|\hat{\theta}_{best}-\hat{\theta}\right|}{\theta}\leq{10}{\%} (41)

where θ^best\hat{\theta}_{best} is the local optimum solution having the highest objective (likelihood) value and θ^\hat{\theta} is the estimated value of θ\theta. If the above is satisfied for all θi\theta_{i}, this result is counted as one pass.

So for maximum likelihood estimation with M=120M=120, the data suggests a probability of p0.85p\approx 0.85 and to get a 99%99\% correct rate we will only need n=3n=3 repeats. This is a result obtained from 1010 multiple initial guesses per 2020 different stimuli configurations (total of 200200 occurrences). Here, we have a high probability of obtaining a global maxima and thus we may get rid of multiple initial guesses requirement in the estimation of θ\theta.

For the optimal design part the problem is expected to be harder as the stimulus amplitudes tend to the upper bound AmaxA_{max}. This should be due to the increasing level of response as the stimulus approaches the upper bound. It should also be remembered that, the Fisher Information measure is computed with respect to a single parameter θi\theta_{i} as shown in (22). Thus it will here be convenient to analyse the respective Fisher Information metric Uk(𝐱,θ𝐢)U_{k}\left(\bf{x},\theta_{i}\right) as defined w.r.to each parameter θi\theta_{i}. Like in the case of likelihood analysis, we do the analysis on 2020 samples (i.e. 2020 stimuli) separately for amplitudes (An)\left(A_{n}\right) and phases (ϕn)\left(\phi_{n}\right). The criterion for passing the test is similar to that of (41) after replacing θ^\hat{\theta} by 𝐱\bf x from (16) and θ^best\hat{\theta}_{best} by 𝐱best\mathbf{x}_{best} with 𝐱best\mathbf{x}_{best} being the stimulus parameter yielding the largest value of Uk(𝐱,θ𝐢)U_{k}\left(\bf{x},\theta_{i}\right). Note that, since we do not have any concept of ”true stimulus parameters” we will use 𝐱best\mathbf{x}_{best} in the denominator of (41).

After doing the analysis for amplitudes AnA_{n}, one can see that highest probability value p=0.8625p=0.8625 is obtained for U1U_{1} whereas the smallest value is obtained for U5U_{5} as p=0.175p=0.175. The second and third smallest values are U3U_{3} and U6U_{6} having p=275p=275 and p=0.35p=0.35 respectively. The indices 11,33,55 and 66 correspond to the βe\beta_{e},wew_{e},weew_{ee} and weiw_{ei}. This means that the lowest probability values occur at the parameters wew_{e},weew_{ee} and weiw_{ei}. This result might be interesting as those three parameters have strong correlations at least with one other network parameter (see Section 3.4). The required number of repeats appears to be n=23n=23 for the worst case (p=0.175p=0.175 for U5U_{5}).

For the phase parameters ϕn\phi_{n}, different initial conditions lead to different values. This is an expected situation as fmincon or similar functions are local optimizers. In fact, the stochastic nature of the global optimizing routines such as the genetic algorithms or simulated annealing will also lead to a similar result when initial populations are provided after each run. Because of this outcome, the solution yielding the largest value of Fisher Information Metric (Uk)\left(U_{k}\right) among all runs with different initial conditions should be preferred in the actual application.

Modern parallel computing facilities will ease the implementation of optimization with multiple random guesses.

3.4 Parameter confounding

The errors of some parameters tend to be correlated (see Figure 7).

Refer to caption
Figure 7: Some network parameters are approximately confounded. Here confounding means that a change of one parameter can be compensated by a proper change of another parameter, such that the stimulus-response relation of the whole network is kept approximately the same. (A) Correlation coefficient matrix of all possible pairs of network parameters obtained by maximum likelihood estimation from either optimal design data or random stimulus data. Since each matrix is symmetric, only a half needs to be shown. Here data from optimal design trials are shown in the upper triangle, whereas data from the random trials are shown in the lower triangle. Each data point was based on 100 repeated trials each containing a sequence of 120 stimuli. (B) Same data as in panel A, except that the absolute values are shown. (C) The three pairs of parameters with the highest correlation coefficients in the upper triangle in panel B are shown in the scatter plots. Data from both optimal design and random trials are shown. Black curves are theoretical predictions according to equations (44), (47) and (48). Green crosshairs are centered at the true parameter values.

Parameter confounding may explain some of the correlations. The idea is that different parameter may compensate each other such that the network behaves in similar ways, even though the parameter values are different. It is known that in individual neurons, different ion channels may be regulated such that diverse configurations may lead to neurons with similar neuronal behaviours in their electrical activation patterns [40]. Similar kind of effect also exists at the network level [10].

Here we will consider the original dynamical equations and demonstrate how parameter confounding might arise. We first emphasize that different parameters in our model are distinct and there is no strict confounding at all. The confounding is approximate in nature.

From the correlation analysis on the optimal design data (Figures 7A and 7B), three pairs of parameters stand out with the strongest correlations. These pairs are (βe(\beta_{e}, we)w_{e}), (βi,wei)(\beta_{i},w_{ei}) and (wee,wei)(w_{ee},w_{ei}).

We will offer an intuitive heuristic explanation based on the idea of parameter confounding.

We here rewrite the dynamical equations for convenience as shown below:

Ve˙\displaystyle\dot{V_{e}} =\displaystyle= βeVe+βe{weege(Ve)weigi(Vi)}+βeweI\displaystyle-\beta_{e}V_{e}+\beta_{e}\left\{w_{ee}g_{e}(V_{e})-w_{ei}g_{i}(V_{i})\right\}+\beta_{e}w_{e}I (42)
Vi˙\displaystyle\dot{V_{i}} =\displaystyle= βiVi+βi{wiege(Ve)wiigi(Vi)}+βiwiI\displaystyle-\beta_{i}V_{i}+\beta_{i}\left\{w_{ie}g_{e}(V_{e})-w_{ii}g_{i}(V_{i})\right\}+\beta_{i}w_{i}I (43)

Example 1: Confounding of the parameter pair (βe(\beta_{e}, we)w_{e}).

The external stimulus II drives the first equation (42) through the weight βewe\beta_{e}w_{e}. If this product is the same, the drive would be the same, even though the individual parameters are different. For example, if βe\beta_{e} is increased by 10% from its true value while wew_{e} is decreased by 10% from its true value, then the product stays the same, so that the external input provides the same drive to (42). Of course, any deviation from the true parameter values also leads to other differences elsewhere in the system. Therefore, the confounding relation is only approximate and not strict. This heuristic argument gives an empirical formula:

β^ew^e=βewe\hat{\beta}_{e}\hat{w}_{e}=\beta_{e}w_{e} (44)

where βe\beta_{e} and wew_{e} refer to the true values of these parameters, whereas β^e\hat{\beta}_{e} and w^e\hat{w}_{e} refer to the estimated values.

Example 2: Confounding of the parameter pair (wei,βi(w_{ei},\beta_{i}).

These two parameters appear separately in different equations, namely, weiw_{ei} appearing only in (42) while βi\beta_{i} appearing only in (43). To combine them, we need to consider the interaction of these two equations. To simplify the problem, we consider a linearised system around the equilibrium state:

Ve˙\displaystyle\dot{V_{e}} =\displaystyle= βeVe+βe{weekeVeweikiVi}+βeweI+Ce\displaystyle-\beta_{e}V_{e}+\beta_{e}\left\{w_{ee}k_{e}V_{e}-w_{ei}k_{i}V_{i}\right\}+\beta_{e}w_{e}I+C_{e} (45)
Vi˙\displaystyle\dot{V_{i}} =\displaystyle= βiVi+βi{wiekeVewiikiVi}+βiwiI+Ci\displaystyle-\beta_{i}V_{i}+\beta_{i}\left\{w_{ie}k_{e}V_{e}-w_{ii}k_{i}V_{i}\right\}+\beta_{i}w_{i}I+C_{i} (46)

where kek_{e} and kik_{i} are the slopes of the gain functions, and CeC_{e} and CiC_{i} are extra terms that depend on the equilibrium state and other parameters. Note that ViV_{i} appears in (45) only once, in the second term in the curly brackets. Since ViV_{i} also satisfies (46), we solve for ViV_{i} in terms of V˙i\dot{V}_{i} from (46) and find a solution of the form: Vi=cV˙i/βi+aV_{i}=c\dot{V}_{i}/\beta_{i}+a where cc is a constant. Substitution into (45) shows that the parameter combination wei/βiw_{ei}/\beta_{i} scales how strongly V˙i\dot{V}_{i} influences this equation. Thus we have a heuristic confounding relation:

w^ei/β^i=wei/βi\hat{w}_{ei}/\hat{\beta}_{i}=w_{ei}/\beta_{i} (47)

Example 3: Confounding of the parameter pair (wee,wei)(w_{ee},w_{ei}).

These two parameters both appear in the curly brackets in (42). We have a heuristic confounding relation:

w^eege(V¯e)w^eigi(V¯i)=weege(V¯e)weigi(V¯i)\hat{w}_{ee}g_{e}(\bar{V}_{e})-\hat{w}_{ei}g_{i}(\bar{V}_{i})=w_{ee}g_{e}(\bar{V}_{e})-w_{ei}g_{i}(\bar{V}_{i}) (48)

where V¯e\bar{V}_{e} and V¯i\bar{V}_{i} are the equilibrium states. If this equation is satisfied, we expect that the term in the curly brackets in (42) would be close to a constant (the right-hand side of (48)) whenever the state VeV_{e} and ViV_{i} are close to the equilibrium values. When the state variables vary freely, we expect this relation to hold only as a very crude approximation.

Simulation results show that these three confounding relation can qualitatively account for the data scattering. The random data follow the same pattern (see Figure 7C) although they appear to have more scattering compared to the optimal design based data. Although the confounding relations are not strictly valid, their offer useful approximate explanations that are based on intuitive argument and are supported by the data as shown in Figure 7.

The theoretical slopes are always smaller, suggesting that the heuristic theory only accounts for a portion of the correlation. It is likely that there are approximate confounding among more than those three pairs of parameters namely (βe(\beta_{e}, we)w_{e}), (βi,wei)(\beta_{i},w_{ei}) and (wee,wei)(w_{ee},w_{ei}).

4 Discussion

4.1 Summary of the results

The main results of this paper can be summarized as follows.

  1. 1.

    We have implemented an optimal design algorithm for generating stimuli that can efficiently probe the E-I network, which describes the dynamic interaction between an excitatory neuronal population and an inhibitory neuronal population. The data has been used to model the auditory system etc.

  2. 2.

    The dynamical network allows both transient and sustained response components (Figure 2)

  3. 3.

    Derivatives are computed directly by differential equations derived from the original system (Figure 3)

  4. 4.

    Optimally designed stimuli have patterns. The amplitude tend to be saturated, which can lead to faster search because only boundary values need to be checked (Figure 4)

  5. 5.

    The optimally designed stimuli elicited responses that have more dynamic range (more variance in the histogram of amplitude of the stimulus waveform I(t)I\left(t\right))

  6. 6.

    The optimal design yield much better parameter estimation in term of likelihood (Figure 5), and in the errors of the individual parameters (Figure 6).

  7. 7.

    We studied parameter confounding which is thought to be a major source of the error (Figure 7).

  8. 8.

    Significance of the results can readily lead to practical applications. For example in the modelling of the auditory networks known works all used stationary sound stimuli. However, it is beneficial to include time dependency as realistic neurons and their networks have several dynamic features.

4.2 Defence of this Research

Choice of the model:

The chosen model is an appropriate and simple model which has only two neurons with one being excitatory and one being inhibitory. This choice is reasonable in the context of this research as the availability of previous work targeting a similar goal is quite limited. In addition, the known ones mostly based on the static feed-forward network designs. In order to verify our optimal design strategy we have to start from a simpler model. The theoretical development and results of this work can very easily be adapted to alternative and/or more complicated models. Here the most critical parameter is the number of data generating neurons (neurons where the spiking events are recorded by electrode implantations) and the computational complexity. The former is a procedural aspect of the experiment whereas the latter is directly related to the instrumentation. In addition, the universal approximating nature of the CTRNN’s is an advantage on this manner.

Speed of computation:

In this work, we aim at investigation of the computational principles without trying the maximizing the computation speed. Right now on a single PC, having a Intel® Core™i7 processor with 66 cores, the computational duration for optimization of a single stimulus is about 1515 minutes. Surely, this is an average value as the number of steps required to converge to an optimum depends on certain conditions such as the value of objective value, gradient, constraint violation and step size. A similar situation exist for the optimization of the likelihood. However in this case, the optimization times of the likelihood will vary due to its increasing size as all the historical spiking is taken into account (see (12)) leading to a function with gradually increasing complexity. Although that is not the only fact contributing to the computational times, the average duration of optimization tend to increase with the size of likelihood MM. An average value for the observed duration of likelihood optimization is 3838 minutes. As a result optimization of one stimulus and subsequent likelihood estimation requires a duration of about 5353 minutes. This is approximately one hour. So one complete run with a sample size of M=120M=120 is completed in a duration about 2828 hours. Changing the value of the sample size MM will have a direct influence on the system computation time. For example, the duration of one complete run will reduce to a value about 1212 hours when M=24M=24. This is an expected situation as there will be a reduced number of the summation terms in the likelihood function. The reduction in the duration of the optimal design algorithm is only based on the reduction in the number of trials which is just equal to the sample size MM. So based on these findings, one will need a speed-up in order to adapt this work to an experiment. There are several ways to speed up the computation:

  1. 1.

    Using a large time step: In this work we have integrated the equations using a time step of 11-millisecond or 0.0010.001-seconds. This value may be increased to levels as high as 0.010.01-seconds. This modification will have a little contribution to the speed of computation. The benefits will most likely be from the optimal design part due to the manipulation of a single interval (no consideration of past/historical spike trains). However, the higher the time step the lower the accuracy of the estimates and the optimality of the stimuli. This main contributor to this fact is the spike generation algorithm where the accuracy of the locations of some spikes are lost when a coarse integration interval is applied.

  2. 2.

    Using and/or developing streamlined optimization algorithms: This can be a subject of a new project on the same field. This development is expected to have a considerable contribution to the computation time without any trade-offs over performance.

  3. 3.

    Generating the stimuli as a block rather than one by one: This is also a potential topic for a new project. This is expected to reduce the optimal design time without losing performance.

  4. 4.

    Employment of larger cluster computing systems (or high performance computing systems (HPC)) having more than 100100 CPU cores: Though the most sophisticated and expensive solution, it is the best approach to cure the overall computational burdens and transform the theoretical only study to an experiment adaptable one.

  5. 5.

    Porting the algorithms to a lower level programming language such as C/C++ or FORTRAN may help in speeding up the computation. If an efficient and stable numerical differentiation algorithm can be employed in this set-up, another optimality criterion such as D or E optimality can be used in the computation of the Fisher Information Metric which might help in reducing the number of steps (i.e MM and/or NitrN_{itr} values).

  6. 6.

    Knowing the fact that the optimal stimulus amplitudes AnA_{n} tend to the upper boundary AmaxA_{max} (remember from Figure 4B). All the amplitudes can be set to same value as An=AA_{n}=A and only AA is computed from optimization. This setting can be helpful for speeding up the optimization time during an actual experiment. However, it should be verified by simulation whether this choice is meaningful for an actual experiment. In this case, one may not need many repeats as only the statistics of the stimuli is required. This occurrence is quite common in the simulation results thus we do not expect a performance degradation when this change is applied to the stimuli characterization.

With the above adaptations, it is expected that we are within reach to reduce the computation time for each 3-second stimuli to less than 3 seconds. In addition one can has the following options which are related to tuning of the algorithms used in this research:

  1. 1.

    The parameter related to the sample size (M)(M) might be reduced. That is an examined situation in this study. Optimal design seems to yield a better estimation performance with a reduced MM compared to random stimuli case with same MM. However, the former will lead to a slightly increased computational time. However, the trade-off is not very direct. For example, a random stimuli based simulation with M=120M=120 samples requires a longer run than a simulation based on an optimally designed stimulus with 2424 samples. This is fairly a good trade-off.

  2. 2.

    Another option to increase the computational performance might be the reduction of the cut-off points in the optimization algorithm such as the first order optimality measure (tolerance of the gradient) and the step-size. This will result in a faster computation but this approach may bring out questions on the accurate detection of the local minimums among which the best one is chosen (both in OED and likelihood optimization). For the fmincon algorithm in MATLAB the first order optimality tolerance and step-size might both be shifted from 1×10091\times 10^{-09} to 1×10061\times 10^{-06}. This tuning brings improvement in the computational duration about 10% without a considerable performance loss. However, if one has a HPC supported computational environment it is strongly recommended not to modify these settings.

Future Issues:

The above network is rather a simpler example to demonstrate the optimal design approach and its computational challenges. However more features can be brought to this research concerning efficiency and applicability to an actual experiment.

  1. 1.

    Speed up issues: The tasks related to speed of computation discussed in Section 4.2 may be a separate project to be developed on top this research.

  2. 2.

    Large number of neurons may be considered together with multiple stimulus inputs and response data collection from multiple neurons (both excitatory and inhibitory groups of neurons).

  3. 3.

    More complex stimulus structures may be utilized. This can be achieved by increasing NN in (15) or considering different stimulus representations other than phase cosines.

  4. 4.

    In this research, the primary goal was the estimation of the network weights ww_{**} and time constants τ\tau_{*}. However, it will be interesting to test the methodology for its performance in estimation of firing thresholds and slopes. (i.e. the parameters aja_{j} and hjh_{j} in (2))

  5. 5.

    Some more realistic details like plasticity can be included to obtain a model describing the synaptic adaptation. Although it is expected to be a harder problem, the method takes fewer stimuli and should be faster.

  6. 6.

    In this research, the optimal design process is performed by maximization of the Fisher Information Metric with respect to a single parameter θk\theta_{k} which are individual elements of the main diagonal of the Fisher Information Matrix. This is at best close to A-Optimality measure of Optimal Design. However, it is stated in [41] that, D-Optimality brings an advantage that the optimization will be immune to the scales of the variables. On the contrary, it is also stated in the same source that this mentioned fact is not true for A- and E-Optimality criteria in general. The sensitivity to scaling of the variables lead to another issue that the confounding of the parameters brings certain problems about the bias and efficiency of the estimates. So it will be quite beneficial to see the results obtained from the same research with the optimal designs performed by D-Optimal and other measures of Fisher Information Metric such as E- and F-Optimality. As these will require a new set of computations it will be better to include them in a future study.

  7. 7.

    Similar to the discussion in Article 6 above, the methodology of optimization in optimal designs and likelihood optimization should be considered. Current work involves evaluation of gradients for the sake of faster computation. However, this requires larger efforts in the preparation as one should develop a specific algorithm to compute the evolution of the gradients satisfactorily. This is also required for a speed-up. With the availability of a high performance computing system, other optimization methods such as simulated-annealing, genetic algorithms and pattern search might be employed instead of the local minimizers such as fmincon of MATLAB. These algorithms may help in searching for a better optimal stimulus.

  8. 8.

    After a sufficiently fast simulation is obtained an experiment can be performed involving a living experimental subject. The mapping of the actual sound heard by the animal during the course of experiment to the optimally designed stimulus is a critical issue here and will also be a part of the future related research.

References

  • [1] V. V. Fedorov, S. L. Leonov, Optimal Design for Nonlinear Response Models, Vol. 53, CRC Press Llc, 2013.
  • [2] D. Telen, F. Logist, E. Van Derlinden, J. Van Impe, Approximate robust optimal experiment design in dynamic bioprocess models, in: Control & Automation (MED), 2012 20th Mediterranean Conference on, IEEE, 2012, pp. 157–162.
  • [3] J. Benda, T. Gollisch, C. K. Machens, A. V. Herz, From response to stimulus: adaptive sampling in sensory physiology, Current Opinion in Neurobiology 17 (4) (2007) 430–436.
  • [4] J. P. Newman, R. Zeller-Townson, M.-F. Fong, S. A. Desai, R. E. Gross, S. M. Potter, Closed-loop, multichannel experimentation using the open-source neurorighter electrophysiology platform, Frontiers in neural circuits 6.
  • [5] L. Paninski, B. Lau, A. Reyes, Noise-driven adaptation: in vitro and mathematical analysis, Neurocomputing 52 (2003) 877–883.
  • [6] P. Z. Marmarelis, V. Z. Marmarelis, Analysis of physiological systems: The white-noise approach, Plenum Press New York, 1978.
  • [7] E. Chichilnisky, A simple white noise analysis of neuronal light responses, Network: Computation in Neural Systems 12 (2) (2001) 199–213.
  • [8] L. Paninski, Maximum likelihood estimation of cascade point-process neural encoding models, Network: Computation in Neural Systems 15 (4) (2004) 243–262.
  • [9] M. C.-K. Wu, S. V. David, J. L. Gallant, Complete functional characterization of sensory neurons by system identification, Annu. Rev. Neurosci. 29 (2006) 477–505.
  • [10] C. DiMattina, K. Zhang, How to modify a neural network gradually without changing its input-output functionality, Neural computation 22 (1) (2010) 1–47.
  • [11] C. DiMattina, K. Zhang, Adaptive stimulus optimization for sensory systems neuroscience, Frontiers in neural circuits 7.
  • [12] C. DiMattina, K. Zhang, Active data collection for efficient estimation and comparison of nonlinear neural models, Neural computation 23 (9) (2011) 2242–2288.
  • [13] F. Pukelsheim, Optimal design of experiments, Vol. 50 of SIAM Classics In Applied Mathematics, SIAM, 1993.
  • [14] H. DeGroot Morris, Optimal statistical decisions, McGraw-Hill Company, New York, 1970.
  • [15] I. J. Myung, Tutorial on maximum likelihood estimation, Journal of Mathematical Psychology 47 (1) (2003) 90–100.
  • [16] D. Eyal, Adaptive on-line modeling in the auditory system: in vivo implementation of the optimal experimental design paradigm, Master’s thesis, Department of Biomedical Engineering, Johns Hopkins University, Baltimore, MD, USA (2012).
  • [17] W. Tam, Adaptive modeling of marmoset inferior colliculus neurons in vivo, Ph.D. thesis, Department of Biomedical Engineering, Johns Hopkins University, Baltimore, MD, USA (2012).
  • [18] R. D. Beer, On the dynamics of small continuous-time recurrent neural networks, Adaptive Behavior 3 (4) (1995) 469–509.
  • [19] E. Ledoux, N. Brunel, Dynamics of networks of excitatory and inhibitory neurons in response to time-dependent inputs, Frontiers in computational neuroscience 5.
  • [20] A. L. Hodgkin, A. F. Huxley, A quantitative description of membrane current and its application to conduction and excitation in nerve, The Journal of physiology 117 (4) (1952) 500.
  • [21] K. E. Hancock, K. A. Davis, H. F. Voigt, Modeling inhibition of type ii units in the dorsal cochlear nucleus, Biological cybernetics 76 (6) (1997) 419–428.
  • [22] K. E. Hancock, H. F. Voigt, Wideband inhibition of dorsal cochlear nucleus type iv units in cat: a computational model, Annals of biomedical engineering 27 (1) (1999) 73–87.
  • [23] J. de la Rocha, C. Marchetti, M. Schiff, A. D. Reyes, Linking the response properties of cells in auditory cortex with network architecture: cotuning versus lateral inhibition, The Journal of Neuroscience 28 (37) (2008) 9151–9163.
  • [24] M. N. Shadlen, W. T. Newsome, Noise, neural codes and cortical organization, Current opinion in neurobiology 4 (4) (1994) 569–579.
  • [25] P. A. Lewis, G. S. Shedler, Simulation of nonhomogeneous poisson processes by thinning, Naval Research Logistics Quarterly 26 (3) (1979) 403–413.
  • [26] U. T. Eden, Point process models for neural spike trains, Neural Signal Processing: Quantitative Analysis of Neural Activity (2008) 45–51.
  • [27] E. N. Brown, R. Barbieri, V. Ventura, R. E. Kass, L. M. Frank, The time-rescaling theorem and its application to neural spike train data analysis, Neural computation 14 (2) (2002) 325–346.
  • [28] M. Nawrot, A. Aertsen, S. Rotter, Single-trial estimation of neuronal firing rates: from single-neuron spike trains to population activity, Journal of neuroscience methods 94 (1) (1999) 81–92.
  • [29] H. Shimazaki, S. Shinomoto, Kernel bandwidth optimization in spike rate estimation, Journal of computational neuroscience 29 (1-2) (2010) 171–182.
  • [30] H. Shimazaki, S. Shinomoto, A method for selecting the bin size of a time histogram, Neural Computation 19 (6) (2007) 1503–1527.
  • [31] S. Koyama, S. Shinomoto, Histogram bin width selection for time-dependent poisson processes, Journal of Physics A: Mathematical and General 37 (29) (2004) 7255.
  • [32] K. D. Miller, F. Fumarola, Mathematical equivalence of two common forms of firing rate models of neural networks, Neural computation 24 (1) (2012) 25–31.
  • [33] S. Flila, P. Dufour, H. Hammouri, M. Nadri, A combined closed loop optimal design of experiments and online identification control approach, in: Control Conference (CCC), 2010 29th Chinese, IEEE, 2010, pp. 1178–1183.
  • [34] D. Telen, F. Logist, E. Van Derlinden, J. F. Van Impe, Robust optimal experiment design: A multi-objective approach, in: Mathematical Modelling, Vol. 7, 2012, pp. 689–694.
  • [35] C. DiMattina, K. Zhang, How optimal stimuli for sensory neurons are constrained by network architecture, Neural computation 20 (3) (2008) 668–708.
  • [36] B. Mendelson, Introduction to topology, Courier Corporation, 1990.
  • [37] H. B. Mann, D. R. Whitney, et al., On a test of whether one of two random variables is stochastically larger than the other, The annals of mathematical statistics 18 (1) (1947) 50–60.
  • [38] J. D. Gibbons, S. Chakraborti, Nonparametric statistical inference, Springer, 2011.
  • [39] M. Hollander, D. A. Wolfe, E. Chicken, Nonparametric statistical methods, Vol. 751, John Wiley & Sons, 2013.
  • [40] A. A. Prinz, D. Bucher, E. Marder, Similar network activity from disparate circuit parameters, Nature neuroscience 7 (12) (2004) 1345–1352.
  • [41] L. A. Khinkis, L. Levasseur, H. Faessel, W. R. Greco, Optimal design for estimating parameters of the 4-parameter hill model, Nonlinearity in biology, toxicology, medicine 1 (3) (2003) 15401420390249925.