Discrete time rescaling theorem: determining goodness of fit for discrete time statistical models of neural spiking.

Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Charlestown, MA 02129, USA.
Neural Computation (Impact Factor: 1.69). 10/2010; 22(10):2477-506. DOI: 10.1162/NECO_a_00015
Source: PubMed

ABSTRACT One approach for understanding the encoding of information by spike trains is to fit statistical models and then test their goodness of fit. The time-rescaling theorem provides a goodness-of-fit test consistent with the point process nature of spike trains. The interspike intervals (ISIs) are rescaled (as a function of the model's spike probability) to be independent and exponentially distributed if the model is accurate. A Kolmogorov-Smirnov (KS) test between the rescaled ISIs and the exponential distribution is then used to check goodness of fit. This rescaling relies on assumptions of continuously defined time and instantaneous events. However, spikes have finite width, and statistical models of spike trains almost always discretize time into bins. Here we demonstrate that finite temporal resolution of discrete time models prevents their rescaled ISIs from being exponentially distributed. Poor goodness of fit may be erroneously indicated even if the model is exactly correct. We present two adaptations of the time-rescaling theorem to discrete time models. In the first we propose that instead of assuming the rescaled times to be exponential, the reference distribution be estimated through direct simulation by the fitted model. In the second, we prove a discrete time version of the time-rescaling theorem that analytically corrects for the effects of finite resolution. This allows us to define a rescaled time that is exponentially distributed, even at arbitrary temporal discretizations. We demonstrate the efficacy of both techniques by fitting generalized linear models to both simulated spike trains and spike trains recorded experimentally in monkey V1 cortex. Both techniques give nearly identical results, reducing the false-positive rate of the KS test and greatly increasing the reliability of model evaluation based on the time-rescaling theorem.

  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: Likelihood-based encoding models founded on point processes have received significant attention in the literature because of their ability to reveal the information encoded by spiking neural populations. We propose an approximation to the likelihood of a point-process model of neurons that holds under assumptions about the continuous time process that are physiologically reasonable for neural spike trains: the presence of a refractory period, the predictability of the conditional intensity function, and its integrability. These are properties that apply to a large class of point processes arising in applications other than neuroscience. The proposed approach has several advantages over conventional ones. In particular, one can use standard fitting procedures for generalized linear models based on iteratively reweighted least squares while improving the accuracy of the approximation to the likelihood and reducing bias in the estimation of the parameters of the underlying continuous-time model. As a result, the proposed approach can use a larger bin size to achieve the same accuracy as conventional approaches would with a smaller bin size. This is particularly important when analyzing neural data with high mean and instantaneous firing rates. We demonstrate these claims on simulated and real neural spiking activity. By allowing a substantive increase in the required bin size, our algorithm has the potential to lower the barrier to the use of point-process methods in an increasing number of applications.
    Neural Computation 11/2013; DOI:10.1162/NECO_a_00548 · 1.69 Impact Factor
  • [Show abstract] [Hide abstract]
    ABSTRACT: This paper studies the multiscale analysis of neural spike trains, through both graphical and Poisson process approaches. We introduce the interspike interval plot, which simultaneously visualizes characteristics of neural spiking activity at different time scales. Using an inhomogeneous Poisson process framework, we discuss multiscale estimates of the intensity functions of spike trains. We also introduce the windowing effect for two multiscale methods. Using quasi-likelihood, we develop bootstrap confidence intervals for the multiscale intensity function. We provide a cross-validation scheme, to choose the tuning parameters, and study its unbiasedness. Studying the relationship between the spike rate and the stimulus signal, we observe that adjusting for the first spike latency is important in cross-validation. We show, through examples, that the correlation between spike trains and spike count variability can be multiscale phenomena. Furthermore, we address the modeling of the periodicity of the spike trains caused by a stimulus signal or by brain rhythms. Within the multiscale framework, we introduce intensity functions for spike trains with multiplicative and additive periodic components. Analyzing a dataset from the retinogeniculate synapse, we compare the fit of these models with the Bayesian adaptive regression splines method and discuss the limitations of the methodology. Computational efficiency, which is usually a challenge in the analysis of spike trains, is one of the highlights of these new models. In an example, we show that the reconstruction quality of a complex intensity function demonstrates the ability of the multiscale methodology to crack the neural code. Copyright © 2013 John Wiley & Sons, Ltd.
    Statistics in Medicine 01/2014; 33(2). DOI:10.1002/sim.5923 · 2.04 Impact Factor
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: To build a cognitive prosthesis that can replace the memory function of the hippocampus, it is essential to model the input-output function of the damaged hippocampal region, so the prosthetic device can stimulate the downstream hippocampal region, e.g., CA1, with the output signal, e.g., CA1 spike trains, predicted from the ongoing input signal, e.g., CA3 spike trains, and the identified input-output function, e.g., CA3-CA1 model. In order for the downstream region to form appropriate long-term memories based on the restored output signal, furthermore, the output signal should contain sufficient information about the memories that the animal has formed. In this study, we verify this premise by applying regression and classification modelings of the spatio-temporal patterns of spike trains to the hippocampal CA3 and CA1 data recorded from rats performing a memory-dependent delayed non-match-to-sample (DNMS) task. The regression model is essentially the multiple-input, multiple-output (MIMO) non-linear dynamical model of spike train transformation. It predicts the output spike trains based on the input spike trains and thus restores the output signal. In addition, the classification model interprets the signal by relating the spatio-temporal patterns to the memory events. We have found that: (1) both hippocampal CA3 and CA1 spike trains contain sufficient information for predicting the locations of the sample responses (i.e., left and right memories) during the DNMS task; and more importantly (2) the CA1 spike trains predicted from the CA3 spike trains by the MIMO model also are sufficient for predicting the locations on a single-trial basis. These results show quantitatively that, with a moderate number of unitary recordings from the hippocampus, the MIMO non-linear dynamical model is able to extract and restore spatial memory information for the formation of long-term memories and thus can serve as the computational basis of the hippocampal memory prosthesis.
    Frontiers in Systems Neuroscience 05/2014; 8:97. DOI:10.3389/fnsys.2014.00097

Full-text (2 Sources)

Available from
May 31, 2014