~/projects/neuro/past/

Neuro-plausible Machine Learning — Past Projects

View current projects

Learning stimuli via dictionary learning

The recording of neural activity in response to repeated presentations of an external stimulus is an established experimental paradigm in the field of neuroscience. Generalized Linear Models are commonly used to describe how neurons encode external stimuli, by statistically characterizing the relationship between the covariates (the stimuli or their derived features) and neural activity. An important question becomes: how to choose appropriate covariates? We propose a data-driven answer to this question that learns the covariates from the neural spiking data, i.e. in an unsupervised manner, and requires minimal user intervention. We use neural spiking data recorded from the Barrel cortex of mice in response to periodic whisker deflections to obtain a data-driven estimate of the covariates that modulate whisker motion.

Relevant papers
[1] Tolooshams B., Song A., Temereanca S., Ba D. "Auto-encoders for natural exponential-family dictionary learning", International Conference on Machine Learning, 2020.
[2] Song A. , Tolooshams B., Temereanca S., Ba D. "Convolutional dictionary learning of stimulus from spiking data", Computational and Systems Neuroscience, 2020.

Main researchers: Bahareh Tolooshams, Andrew Song


Spike sorting using dictionary learning

Decades of research in experimental neuroscience suggest that neurons communicate through spikes, which emphasizes the importance of accurate spike sorting from raw electrophysiology data. In these projects, we cast the spike sorting problem as a convolutional dictionary learning problem and suggest 1) optimization-based 2) autoencoder-based approaches. We demonstrate that not only these frameworks yield more accurate results compared to the baselines, but also are much more computationally efficient.

Relevant papers
[1] Song A., Flores F., Ba D. "Convolutional dictionary learning with grid refinement", IEEE Transactions on Signal Processing, 2020.
[2] Tolooshams B., Dey S., Ba D. "Deep residual auto-encoders for expectation maximization-inspired dictionary learning", IEEE Transactions on Neural Networks and Learning Systems, 2020.

Main researchers: Andrew Song, Bahareh Tolooshams


Clustering neural spiking data

Neural spiking data are an important source of information for understanding the brain. Analyzing the collective behavior of large assemblies of neurons brings further insight on how knowledge is acquired by humans and animals alike. In response to an external stimulus, sets of neurons often exhibit grouped response patterns. The identity of these neural groups -- as well as the highly nonlinear dynamics of their firing trajectories -- may inform us about how complicated emotions such as fear are encoded at the inter-neuronal level. From a machine learning perspective, we tackle this problem by viewing it as an instantiation of (nonlinear) time series clustering. We develop a general statistical framework and apply its methodology to answering a host of inferential questions in the neuroscience domain.

Relevant papers
[1]Lin A., Zhang Y., Heng J., Allsop S., Tye K., Jacob P., Ba D. "Clustering time series with nonlinear dynamics: a bayesian non-parametric and particle-based approach", International Conference on Artificial Intelligence and Statistics, 2019.

Main researchers: Alex Lin


State-space multitaper time-frequency analysis

Rapid growth in sensor and recording technologies is spurring rapid growth in time series data. Nonstationary and oscillatory structure in time series is commonly analyzed using time-varying spectral methods. These widely used techniques lack a statistical inference framework applicable to the entire time series. We develop a state-space multitaper (SS-MT) framework for time-varying spectral analysis of nonstationary time series. We efficiently implement the SS-MT spectrogram estimation algorithm in the frequency domain as parallel 1D complex Kalman filters. In analyses of human EEGs recorded under general anesthesia, the SS-MT paradigm provides enhanced denoising (>10 dB) and spectral resolution relative to standard multitaper methods, a flexible time-domain decomposition of the time series, and a broadly applicable, empirical Bayes' framework for statistical inference.

Relevant papers
[1] Kim S-E., Behr M., Ba D., Brown E. "State-space multitaper time-frequency analysis", Proceedings of the National Academy of Sciences, 2018.

Main researchers: Demba Ba


Teaching new tricks

The study of learning in populations of subjects can provide insights into the changes that occur in the brain with aging, drug intervention, and psychiatric disease. Obtaining objective measures of behavioral changes is challenging since learning is dynamic, varies between individuals, and observations are frequently binary. We introduce a method for analyzing binary response data acquired during the learning of object-reward associations across multiple days. The method can quantify the variability of performance within a day and across days, and can capture abrupt changes in learning that would occur, for instance, due to reversal in the reward. The backbone of the method is a separable two-dimensional random field model for binary response data from multi-day behavioral experiments. We use data from young and aged female macaque monkeys performing a reversal-learning task that the model is well-suited for discriminating between-group differences and identifying subgroups.

Relevant papers
[1] Malem-Shinitski N., Zhang Y., Gray D., Burke S., Smith A., Barnes C., Ba D. "A separable two-dimensional random field model of binary response data from multi-day behavioral experiments", Journal of Neuroscience Methods, 2018.
[2] Malem-Shinitski N., Zhang Y., Gray D., Burke S., Smith A., Barnes C., Ba D. "Can you teach an old monkey a new trick?", Computational and Systems Neuroscience, 2017.

Main researchers: Diana Zhang


Estimating a separably-Markov random field

A fundamental problem in neuroscience is to characterize the dynamics of spiking from the neurons in a circuit that is involved in learning about a stimulus or a contingency. Classical methods to analyze neural spiking data collapse neural activity overtime or trials, which may cause the loss of information pertinent to understanding the function of a neuron or circuit. We introduce a new method that can determine not only the trial-to-trial dynamics that accompany the learning of a contingency by a neuron, but also the latency of this learning with respect to the onset of a conditioned stimulus. The backbone of the method is a separable two-dimensional random field model of neural spike raster. We develop efficient statistical and computational tools to estimate the parameters of the separable 2D RF model. We apply these to data collected from neurons in the prefrontal cortex in an experiment designed to characterize the neural underpinnings of the associativelearning of fear in mice.

Relevant papers
[1] Zhang Y., Malem-Shinitski N., Allsop S., Tye K., Ba D. "Estimating a separably-Markov random field (SMuRF) from binary observations", Neural Computation, 2018.
[2] Zhang Y., Malem-Shinitski N., Allsop S., Tye K., Ba D. "A two-dimensional seperable random field model of within and cross-trial neural spiking dynamics", Computational and Systems Neuroscience, 2017.

Main researchers: Diana Zhang