##Systems neuroscience beyond the tuning curve
John Pearson
[pearsonlab.github.io/beyond-tuning-curve](https://pearsonlab.github.io/beyond-tuning-curve)
##The tuning curve paradigm
- Subject repeats same behavior many times
- Average neural firing across repeats
- Neuronal populations are characterized by tuning
##The prefrontal reality
- Complex behaviors never quite repeated
- Dynamics not quite regular enough to average
- Where are the tuning curves?
Why machine learning?
Models that actually fit
is the model really a good model?
structured black box
The ANN as model system
algorithm as homology
what details matter?
takes distributed, asynchronous seriously
What we do
Organization of complex foraging, social behavior.
Neural mechanisms from data-driven models
Whole-brain dynamics
### Today's plan:
Three exercises in removing constraints:
- Learning stimulus space without labels ([link](http://journals.plos.org/ploscompbiol/article?id=10.1371/journal.pcbi.1005645))
- Complex behavior without trial averaging ([arXiv](https://arxiv.org/abs/1702.07319))
- Single-trial analysis of neural spiking (in prep)
###We are not the first
- Gallant lab (fMRI) ([Huth 2012](http://www.sciencedirect.com/science/article/pii/S0896627312009348), [Stansbury 2013](http://www.sciencedirect.com/science/article/pii/S0896627313005503))
- Continuous latent states ([Park 2014](http://www.nature.com/neuro/journal/v17/n10/abs/nn.3800.html), [Buesing 2014](http://papers.nips.cc/paper/5339-clustered-factor-analysis-of-multineuronal-spike-data), [Archer 2015](https://arxiv.org/abs/1511.07367), [Park 2015](http://papers.nips.cc/paper/5790-unlocking-neural-population-non-stationarities-using-hierarchical-dynamics-models))
- Discrete latent states ([Escola 2011](http://www.mitpressjournals.org/doi/abs/10.1162/NECO_a_00118#.WNqSexLythE), [Putzky 2014](http://papers.nips.cc/paper/5338-a-bayesian-model-for-identifying-hierarchically-organised-states-in-neural-population-activity))
- ...and many more
###So what's different?
- Previous models: latents capture *internal* dynamics
- latents can be driven by stimuli
- but vary for presentations of the same stimulus
- Our model: latents capture *stimulus* dynamics
- each stimulus frame has a set of binary tags
- tags follow a Hidden (semi-)Markov Model
- latents are *the same* for repeated stim presentations
###Model fitting
We have $p(N|Z, \Theta)$
$Z$: latent variables, $\Theta$: model parameters
We want
$$
p(Z, \Theta | N) \propto p(N|Z, \Theta)\, p(Z) \, p(\Theta)
$$
But too hard to do Bayes' Rule exactly!
> Do you want the wrong answer to the right question or the right answer to the wrong question? I think you want the former.
>
> — David Blei
###Variational Bayesian (VB) Inference
- Replace true posterior $p$ with *approximate* posterior $q$
- Minimize "distance" $KL(q \Vert p)$ between actual and approximate posteriors
- Same as maximizing the evidence lower bound (ELBO): $\log p(N)$
Experiment I: Synthetic data
Experiment II: Temporal Cortex
McMahon et al. (PNAS, 2014)
Face, monkey, and body part cells!
Experiment II: Temporal Cortex
Viewpoint selectivity!
###What did we do?
- Given spike counts, *what features drive firing?*
- Multiply "tag" each stimulus frame
- Model recovers features from even modest data sizes when signal is strong
- Goal is to look for patterns that **suggest new experiments.**
From movement to strategy
Shariq Iqbal
Caroline Drucker
Jean-Francois Gariépy
Michael Platt
Penalty Shot
Iqbal and Pearson (arXiv)
Complexity tax
each trial a different length
how to average, align?
need to "reduce" dynamics
Real trials
### What we want
- ~~Joystick censoring~~
- ~~Details of motor execution~~
- Model of latent cognitive state
- Capture interaction between players
### Our approach
- Borrow from control theory, time series
- Structured black box models (pieces make sense)
- Neural networks for flexible fitting
Modeling I
Observed positions at each time ($y_t$):
$$
y_t = \begin{bmatrix}
y_{goalie} &
x_{puck} &
y_{puck}
\end{bmatrix}^\top
$$
Control inputs ($u_t$) drive changes in observed positions:
$$y_{t + 1} = y_t + v_{max} \sigma(u_t)$$
Goal: predict control inputs from trial history:
$$u_t = F(y_{1:t})$$
Modeling II
Assumption: PID control
$$
u_t = u_{t-1} + L * (g_{t} - y_{t}) + \epsilon_t
$$
###What did we do?
- Dynamic control tasks let us leverage motor behavior to study cognitive and social decisions.
- Structured black-box models allow us to carve behavior into interpretable pieces.
- We inferred a value function capable of explaining behavior in terms of goals.
## Conclusions
Modeling is a tool for doing the *right* experiment
- Model single trials instead of averaging across trials
- Data-driven firing patterns, not tuning curves
- Population dynamics, not static single units
Sponsors
A social brain?
Mars et al. (PNAS 2014)
Same behavior, different mechanisms
Adams, Watson, Pearson, and Platt (2012)
Foraging, for instance
Pearson, Watson, and Platt (2014)
What I cannot create, I do not understand. — Richard Feynman
###A reverse engineering approach
- Work "outside-in"
- Focus on computational constraints
- "Structured black box" modeling