Skip to main content

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • Article
  • Published:

The geometry of cortical representations of touch in rodents

Abstract

Neurons often encode highly heterogeneous non-linear functions of multiple task variables, a signature of a high-dimensional geometry. We studied the representational geometry in the somatosensory cortex of mice trained to report the curvature of objects touched by their whiskers. High-speed videos of the whiskers revealed that the task can be solved by linearly integrating multiple whisker contacts over time. However, the neural activity in somatosensory cortex reflects non-linear integration of spatio-temporal features of the sensory inputs. Although the responses at first appeared disorganized, we identified an interesting structure in the representational geometry: different whisker contacts are disentangled variables represented in approximately, but not fully, orthogonal subspaces of the neural activity space. This geometry allows linear readouts to perform a broad class of tasks of different complexities without compromising the ability to generalize to novel situations.

This is a preview of subscription content, access via your institution

Access options

Buy this article

Prices may be subject to local taxes which are calculated during checkout

Fig. 1: Different geometries of the neuronal representations and the whisker-based shape discrimination task.
Fig. 2: The whisker-based shape discrimination task can be solved by linearly integrating whisker contacts across time.
Fig. 3: Non-linear classifiers outperform linear classifiers on more complex tasks but not on more complex shapes.
Fig. 4: Populations of neurons in the mouse somatosensory cortex (S1) exhibit non-linear mixed selectivity for task variables.
Fig. 5: Contacts and whisker angular position are the variables that contribute the most to the prediction of S1 population activity.
Fig. 6: Neurons show non-linear mixed selectivity, mostly originating from non-linear interactions between contacts and whisker angular position.
Fig. 7: The geometry of representations observed in S1.
Fig. 8: The geometry of representations in RNNs is modulated by the difficulty of the task.

Similar content being viewed by others

Data availability

The datasets analyzed in this study have been deposited to Zenodo: https://doi.org/10.5281/zenodo.4743837.

Code availability

The code used for data acquisition and pre-processing is available at https://github.com/cxrodgers/Rodgers2021. The code used for all analyses in this study is available at https://github.com/ramonnogueira/TheGeometryOfS1.

References

  1. Johansson, R. S. & Flanagan, J. R. Coding and use of tactile signals from the fingertips in object manipulation tasks. Nat. Rev. Neurosci. 10, 345–359 (2009).

    Article  CAS  Google Scholar 

  2. Bensmaia, S. J., Tyler, D. J. & Micera, S. Restoration of sensory information via bionic hands. Nature Biomedical Engineering1-13 (2020).

  3. Davidson, P. W. Haptic judgments of curvature by blind and sighted humans. J. Exp. Psychol. 93, 43 (1972).

    Article  CAS  Google Scholar 

  4. Lederman, S. J. & Klatzky, R. L. Hand movements: A window into haptic object recognition. Cogn. Psychol. 19, 342–368 (1987).

    Article  CAS  Google Scholar 

  5. Rodgers, C. C. et al. Sensorimotor strategies and neuronal representations for shape discrimination. Neuron 109, 2308–2325 (2021).

    Article  CAS  Google Scholar 

  6. Rigotti, M. et al. The importance of mixed selectivity in complex cognitive tasks. Nature 497, 585–590 (2013).

    Article  CAS  Google Scholar 

  7. Bernardi, S. et al. The geometry of abstraction in the hippocampus and prefrontal cortex. Cell 183, 954–967 (2020).

    Article  CAS  Google Scholar 

  8. Haxby, J. V. et al. A common, high-dimensional model of the representational space in human ventral temporal cortex. Neuron 72, 404–416 (2011).

    Article  CAS  Google Scholar 

  9. Guntupalli, J. S. et al. A model of representational spaces in human cortex. Cereb. cortex 26, 2919–2934 (2016).

    Article  Google Scholar 

  10. Chung, S. & Abbott, L. Neural population geometry: An approach for understanding biological and artificial neural networks. Curr. Opin. Neurobiol. 70, 137–144 (2021).

    Article  CAS  Google Scholar 

  11. Saxena, S. & Cunningham, J. P. Towards the neural population doctrine. Curr. Opin. Neurobiol. 55, 103–111 (2019).

    Article  CAS  Google Scholar 

  12. Fusi, S., Miller, E. K. & Rigotti, M. Why neurons mix: high dimensionality for higher cognition. Curr. Opin. Neurobiol. 37, 66–74 (2016).

    Article  CAS  Google Scholar 

  13. Bengio, Y., Courville, A. & Vincent, P. Representation learning: A review and new perspectives. IEEE Trans. pattern Anal. Mach. Intell. 35, 1798–1828 (2013).

    Article  Google Scholar 

  14. Higgins, I. et al. β-VAE: Learning basic visual concepts with a constrained variational framework. International Conference on Learning Representations (ICLR) (2017).

  15. Higgins, I., Racanière, S. & Rezende, D. Symmetry-based representations for artificial and biological general intelligence. Frontiers in Computational Neuroscience28 (2022).

  16. Higgins, I. et al. Unsupervised deep learning identifies semantic disentanglement in single inferotemporal neurons. arXiv preprint arXiv:2006.14304 (2020).

  17. Boyle, L., Posani, L., Irfan, S., Siegelbaum, S. A. & Fusi, S. The geometry of hippocampal ca2 representations enables abstract coding of social familiarity and identity. bioRxiv (2022).

  18. Raposo, D., Kaufman, M. T. & Churchland, A. K. A category-free neural population supports evolving demands during decision-making. Nat. Neurosci. 17, 1784 (2014).

    Article  CAS  Google Scholar 

  19. Chang, L. & Tsao, D. Y. The code for facial identity in the primate brain. Cell 169, 1013–1028 (2017).

    Article  CAS  Google Scholar 

  20. Insafutdinov, E., Pishchulin, L., Andres, B., Andriluka, M. & Schiele, B. Deepercut: A deeper, stronger, and faster multi-person pose estimation model. In European Conference on Computer Vision, 34-50 (Springer, 2016).

  21. Pishchulin, L. et al. Deepcut: Joint subset partition and labeling for multi person pose estimation. In Proceedings of the IEEE conference on computer vision and pattern recognition, 4929-4937 (2016).

  22. Mathis, A. et al. Deeplabcut: markerless pose estimation of user-defined body parts with deep learning. Nat. Neurosci. 21, 1281–1289 (2018).

    Article  CAS  Google Scholar 

  23. Alain, G. & Bengio, Y. Understanding intermediate layers using linear classifier probes. arXiv preprint arXiv:1610.01644 (2016).

  24. Buonomano, D. V. & Maass, W. State-dependent computations: spatiotemporal processing in cortical networks. Nat. Rev. Neurosci. 10, 113–125 (2009).

    Article  CAS  Google Scholar 

  25. Yamins, D. L. et al. Performance-optimized hierarchical models predict neural responses in higher visual cortex. Proc. Natl Acad. Sci. 111, 8619–8624 (2014).

    Article  CAS  Google Scholar 

  26. Yamins, D. L. & DiCarlo, J. J. Using goal-driven deep learning models to understand sensory cortex. Nat. Neurosci. 19, 356–365 (2016).

    Article  CAS  Google Scholar 

  27. Panichello, M. F. & Buschman, T. J. Shared mechanisms underlie the control of working memory and attention. Nature 592, 601–605 (2021).

    Article  CAS  Google Scholar 

  28. She, L., Benna, M. K., Shi, Y., Fusi, S. & Tsao, D. Y. The neural code for face memory. bioRxiv (2021).

  29. Elsayed, G. F., Lara, A. H., Kaufman, M. T., Churchland, M. M. & Cunningham, J. P. Reorganization between preparatory and movement population responses in motor cortex. Nat. Commun. 7, 1–15 (2016).

    Article  CAS  Google Scholar 

  30. Stringer, C., Pachitariu, M., Steinmetz, N., Carandini, M. & Harris, K. D. High-dimensional geometry of population responses in visual cortex. Nature 571, 361–365 (2019).

    Article  CAS  Google Scholar 

  31. Gao, P. & Ganguli, S. On simplicity and complexity in the brave new world of large-scale neuroscience. Curr. Opin. Neurobiol. 32, 148–155 (2015).

    Article  CAS  Google Scholar 

  32. Lindsay, G. W., Rigotti, M., Warden, M. R., Miller, E. K. & Fusi, S. Hebbian learning in a random network captures selectivity properties of the prefrontal cortex. J. Neurosci. 37, 11021–11036 (2017).

    Article  CAS  Google Scholar 

  33. Dang, W., Jaffe, R. J., Qi, X.-L. & Constantinidis, C. Emergence of non-linear mixed selectivity in prefrontal cortex after training. Journal of Neuroscience (2021).

  34. Meshulam, L., Gauthier, J. L., Brody, C. D., Tank, D. W. & Bialek, W. Collective behavior of place and non-place neurons in the hippocampal network. Neuron 96, 1178–1191 (2017).

    Article  CAS  Google Scholar 

  35. Nogueira, R. et al. The effects of population tuning and trial-by-trial variability on information encoding and behavior. J. Neurosci. 40, 1066–1083 (2020).

    Article  CAS  Google Scholar 

  36. Stefanini, F. et al. A distributed neural code in the dentate gyrus and in ca1. Neuron 107, 703–716 (2020).

    Article  CAS  Google Scholar 

  37. Valente, M. et al. Correlations enhance the behavioral readout of neural population activity in association cortex. Nat. Neurosci. 24, 975–986 (2021).

    Article  CAS  Google Scholar 

  38. Frost, N. A., Haggart, A. & Sohal, V. S. Dynamic patterns of correlated activity in the prefrontal cortex encode information about social behavior. PLoS Biol. 19, e3001235 (2021).

    Article  CAS  Google Scholar 

  39. Hirokawa, J., Vaughan, A., Masset, P., Ott, T. & Kepecs, A. Frontal cortex neuron types categorically encode single decision variables. Nature 576, 446–451 (2019).

    Article  CAS  Google Scholar 

  40. Zhang, W. & Bruno, R. M. High-order thalamic inputs to primary somatosensory cortex are stronger and longer lasting than cortical inputs. Elife 8, e44158 (2019).

    Article  Google Scholar 

  41. Manita, S. et al. A top-down cortical circuit for accurate sensory perception. Neuron 86, 1304–1316 (2015).

    Article  CAS  Google Scholar 

  42. Banerjee, A. et al. Value-guided remapping of sensory cortex by lateral orbitofrontal cortex. Nature 585, 245–250 (2020).

    Article  CAS  Google Scholar 

  43. Moore, J. D., Mercer Lindsay, N., Deschênes, M. & Kleinfeld, D. Vibrissa self-motion and touch are reliably encoded along the same somatosensory pathway from brainstem through thalamus. PLoS Biol. 13, e1002253 (2015).

    Article  Google Scholar 

  44. Ranganathan, G. N. et al. Active dendritic integration and mixed neocortical network representations during an adaptive sensing behavior. Nat. Neurosci. 21, 1583–1590 (2018).

    Article  CAS  Google Scholar 

  45. Gulli, R. A. et al. Context-dependent representations of objects and space in the primate hippocampus during virtual navigation. Nat. Neurosci. 23, 103–112 (2020).

    Article  CAS  Google Scholar 

  46. Roussy, M. et al. Ketamine disrupts naturalistic coding of working memory in primate lateral prefrontal cortex networks. Molecular Psychiatry1-16 (2021).

  47. Nelson, M. E. & MacIver, M. A. Sensory acquisition in active sensing systems. J. Comp. Physiol. A 192, 573–586 (2006).

    Article  CAS  Google Scholar 

  48. Krakauer, J. W., Ghazanfar, A. A., Gomez-Marin, A., MacIver, M. A. & Poeppel, D. Neuroscience needs behavior: correcting a reductionist bias. Neuron 93, 480–490 (2017).

    Article  CAS  Google Scholar 

  49. Nogueira, R. et al. Lateral orbitofrontal cortex anticipates choices and integrates prior with current information. Nat. Commun. 8, 1–13 (2017).

    Article  Google Scholar 

  50. Pachitariu, M., Steinmetz, N. A., Kadir, S. N., Carandini, M. & Harris, K. D. Fast and accurate spike sorting of high-channel count probes with kilosort. Adv. neural Inf. Process. Syst. 29, 4448–4456 (2016).

    Google Scholar 

  51. Ashwood, Z. C. et al. Mice alternate between discrete strategies during perceptual decision-making. Nat. Neurosci. 25, 201–212 (2022).

    Article  CAS  Google Scholar 

  52. Calhoun, A. J., Pillow, J. W. & Murthy, M. Unsupervised identification of the internal states that shape natural behavior. Nat. Neurosci. 22, 2040–2049 (2019).

    Article  CAS  Google Scholar 

Download references

Acknowledgements

We would like to thank the members of the Center for Theoretical Neuroscience, Marcus K. Benna and Mattia Rigotti for all their insightful comments and suggestions. Support was provided by NINDS/NIH (R01NS094659, R01NS069679, F32NS096819, and U01NS099726); NSF 1707398 (Neuronex); the Gatsby Charitable Foundation (GAT3708); the Simons Foundation; the Swartz Foundation; Northrop Grumman; a Kavli Institute for Brain Science postdoctoral fellowship (to CR); and a Brain and Behavior Research Foundation Young Investigator Award (to CR).

Author information

Authors and Affiliations

Authors

Contributions

RN and SF conceived the project and the analytic approach. CR developed the behavior, videography and performed the electrophysiological recordings. RN analyzed the data. RN, CR, RB, and SF decided how to interpret the results. RN and SF wrote, and CR and RB edited the manuscript.

Corresponding authors

Correspondence to Ramon Nogueira or Stefano Fusi.

Ethics declarations

Competing interests

The authors declare no competing interests.

Peer review

Peer review information

Nature Neuroscience thanks the anonymous reviewers for their contribution to the peer review of this work.

Additional information

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Extended data

Extended Data Fig. 1 Contact rate for convex and concave shapes and for correct and error trials for all mice.

Contact rate (y-axis) as a function of time throughout the trial (x-axis) separately for convex and concave shapes (left), and for correct and error trials (right). Contacts were higher for convex than concave shapes for whisker C1 and C2, whereas whisker C3 showed the opposite trend. Contacts were higher for correct than error trials for all whiskers and animals. (a-j) Results for all mice. (k) Mean results across animals (n = 10 mice).

Extended Data Fig. 2 Linear and non-linear decoders performed similarly for correct and error trials, different stopping locations, and flatter shapes.

(a) A multi-layer feedforward network model is trained to use the full spatio-temporal pattern of contacts and angle of contacts to predict the stimulus and the choice of the animal on a trial-by-trial basis (see Fig. 2). Models were trained using all trials and tested on correct (black) and incorrect (red) trials. Only linear and non-linear models with one hidden layer are shown. Stimulus decoding (left) produced higher decoding performance for correct trials than errors, probably due to the higher number of contacts made by mice on correct trials. On the contrary, correct trials conveyed much more information about animals’ choice than incorrect trials (right). One possible explanation of these effects is that in approximately 60% of the trials animals make very accurate choices that are based on properly sampled sensory cues. In the other 40% of the trials, animals still sample information properly but their choice is inaccurate and based on a hidden variable we do not have access to51,52. (b) Decoding performance (y-axis) for the different decoders (x-axis) for the easy (linearly separable; left panel) and very complex (non-linearly separable, 3D-parity; right panel) tasks (see Methods). Non-linear cue integration is only advantageous when the task itself requires complex sensory integration across time and whiskers. (c) Linear and non-linear classifiers performed equally well on the shape discrimination task from the real spatio-temporal pattern of whisker contacts when conditioned on trials corresponding to far, medium and close stopping locations. Error bars in panels (a-c) correspond to s.e.m. across mice (n = 10). (d) Similar to behavior (see5), flatter shapes were more difficult to discriminate than the standard ones from the real spatio-temporal pattern of whisker contacts. Errorbars correspond to s.e.m. across mice that were presented both standard and flatter shapes (n = 3).

Extended Data Fig. 3 Different feedforward and recurrent architectures were used to decode behavior and fit the S1 encoding models.

(a) Stimulus (green) and choice (blue) were decoded on a trial-by-trial basis from the spatio-temporal pattern of whisking contacts using RNNs with different number of hidden units. On each time step and trial the input to the RNN decoder was the number of contacts and angle of contact for all the whiskers. Errorbars correspond to s.e.m. across mice (n = 10). Feed-forward networks performed in general better than RNNs for decoding behavior (see Fig. 2). (b) S1 population activity was regressed against whisker and task variables (encoding model) using RNNs with different number of hidden units. Errorbars correspond to s.e.m. across neurons (n = 584). See Extended Data Fig. S5a for the distribution of CV R2 across neurons for the linear and the best non-linear (feed-forward with one hidden layer of hundred units) encoding models. Feed-forward networks performed in general better than RNNs on explaining S1 population activity (see Fig. 4). (c) Linear and non-linear feed-forward networks with different number of units (artificial neurons) in the hidden layers are equally good at predicting the presented shape (stimulus; green) and animal’s choice (choice; blue) on a trial-by-trial basis from the spatio-temporal pattern of whisker contacts and angle of contacts (see Fig. 2). Errorbars correspond to s.e.m. across mice (n = 10). (d) The profile of explanatory power for S1 population activity across models is qualitatively equivalent for 20, 100 (see Fig. 4), and 200 units in the hidden layers of the feed-forward encoding models. Errorbars correspond to s.e.m. across neurons (n = 584). See Extended Data Fig. S5a for the distribution of CV R2 across neurons for the linear and the best non-linear (feed-forward with one hidden layer of hundred units) encoding models.

Extended Data Fig. 4 Linear and non-linear classifiers perform equally well on other shapes used in the simulated whisker discrimination task.

(a-e) Linear and non-linear classifiers performed equally well from the spatio-temporal pattern of simulated whisker contacts and angles of contact when different general parameters of the shapes were used in the simulations: more distant stopping locations (a); smaller shapes (b); further shapes (c); closer shapes (d); and flatter shapes (e) (see Methods). (f) Flatter shapes (larger radius) were more difficult to discriminate than more curved shapes on a simulation of the whisker-based discrimination task. (g) Curvature discrimination task: discriminate between curved (green) or flat (orange) shapes (first panel). Non-linear classifiers substantially outperform linear ones on the simulated curvature discrimination task when the full spatio-temporal pattern of contacts and angle of contacts is used to discriminate between flatter (orange) and more curved (green) shapes (second panel). Total number of contacts for the pair (C1, C2) (third panel), (C1, C3) (fourth panel) and (C2, C3) (fifth panel). The boundary between the two categories is non-linear. (h) Example of another simulated non-linear discrimination task (green vs orange bars). Non-linear classifiers substantially outperform linear ones on this task when the full spatio-temporal pattern of contacts and angle of contacts is used. The boundary between the two categories is non-linear, especially for C1 vs C2. Errorbars in all panels correspond to s.e.m. across independent simulations (n = 5).

Extended Data Fig. 5 Mean goodness-of-fit across neurons, correct and incorrect trials, neuronal types, and S1 layers.

(a) Mean goodness-of-fit across neurons (y-axis; R2) for the different encoding models (x-axis) on held-out data (left), and the distribution of R2 across neurons for the non-linear (one hidden layer) (y-axis) and the linear encoding models (x-axis) (right). All models were fit on simultaneously recorded populations. For held-out data, the best model is a feed-forward fully connected network with only one hidden layer (NonLin-1). (b) Mean goodness-of-fit across neurons (y-axis; R2) for the different encoding models (x-axis) on the data used for training. For the train data, the more complex the model (more parameters), the better the prediction. (c) Mean CV R2 (y-axis) across S1 neurons for the different encoding models (x-axis) when evaluated in correct (black) and incorrect trials (red). Encoding models explained S1 activity better on correct than error trials. Errorbars in panels (a-c) correspond to s.e.m. across neurons (n = 584). (d) Mean CV R2 across neurons for the different encoding models (x-axis) on held-out data for all neurons (left), only excitatory (middle) and only inhibitory neurons (right). (e) Mean CV R2 across neurons for the different encoding models on held-out data for neurons across layers for all (top), excitatory (middle) and inhibitory neurons (bottom). (f-g) Encoding fits when using Poisson loss instead of mean squared error (MSE). The y-axis shows the Poisson-loss equivalent of the R2, the Pseudo-R2. The Pseudo-R2 is calculated as 1 - PLoss/Variance, where PLoss is the negative Log-likelihood of the Poisson model. All models were fit on simultaneously recorded populations. Errorbars in all panels correspond to s.e.m. across neurons (all n = 584; excitatory n = 491; inhibitory n = 93; layer 2/3 n = 68; layer 4 n = 157; layer 5 n = 249; layer 6 n = 96).

Extended Data Fig. 6 Contribution of the different regressors to S1 activity across neuronal types, layers, and columns in S1.

(a) Distribution of R2 for the full model (x-axis; R2 Full Model) vs R2 for the different ablations (y-axis; R2 Reduced Model) across all recorded neurons. The R2 for the different ablations is calculated by testing the model on held-out trials where those features (or groups of features) have been set to zero. (b) Mean ΔR2 across all neurons (y-axis; \(\Delta {{{{\rm{R}}}}}^{2}={{{{{\rm{R}}}}}^{2}}_{{{{\rm{Full}}}}}-{{{{{\rm{R}}}}}^{2}}_{{{{\rm{Reduced}}}}}\)) for the different groups of variables. Whisker contacts (Contacts) and continuous angle (Angle) are the most important groups of regressors in predicting S1 activity. Errorbars correspond to s.e.m. across neurons (n = 584). (c) ΔR2 separately for excitatory and inhibitory neurons. Inhibitory populations show a higher ΔR2 because their R2 is overall higher (see Extended Data Fig. S5e). (d) ΔR2 across layers of the somatosensory cortex. Whisker contacts have a stronger effect on superficial layers, while whisker angle has a stronger effect on deeper layers. (e) The metric ΔR2 reveals that neurons in S1 do not strictly obey somatotopy during the whisker-based discrimination task (see also5). The contribution to the encoding model’s performance (y-axis; ΔR2) for the groups of variables associated with whiskers’ contacts (C0, C1, C2 and C3 for all time steps) and angular position (θC0, θC1, θC2, and θC3 for all time steps) for all neurons. (f-h) Whisker and angular position encoding strength for C1 column (b), C2 column (c) and C3 column (d). While C1 contacts is the strongest driver in C1 column, C2 and C3 columns are not dominated by C2 and C3 contacts, respectively. Errorbars in all panels correspond to s.e.m. across neurons (all n = 584; excitatory n = 491; inhibitory n = 93; layer 2/3 n = 68; layer 4 n = 157; layer 5 n = 249; layer 6 n = 96; C1 column n = 117; C2 column n = 208; C3 column n = 204).

Extended Data Fig. 7 Non-linear mixed selectivity across neuronal types and cortical layers for the different groups of task variables.

(a) Different coding scenarios would produce different values for \(\Delta {{{{\rm{R}}}}}^{2}-{\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}}\). Here we show different tuning schemes with respect to C1 and C3 for a fictional neuron (r1). The metric \(\Delta {{{{\rm{R}}}}}^{2}-{\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}}\) was used to evaluate to what extend the pure non-linear terms were important to predict the population’s firing rate. (Left) If the relationship between neuronal activity and encoding variables is linear, \({\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}}\ne 0,\Delta {{{{\rm{R}}}}}^{2}={\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}}\) and therefore \(\Delta {{{{\rm{R}}}}}^{2}-{\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}}=0\). (Middle) If the relationship between neuronal activity and encoding variables is purely non-linear, \({\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}}=0\), ΔR2 ≠ 0, and \(\Delta {{{{\rm{R}}}}}^{2}-{\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}} > 0\). (Right) If the encoding model is composed of both linear and non-linear components, \({\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}}\ne 0\), \(\Delta {{{{\rm{R}}}}}^{2}\ne {\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}}\), and \(\Delta {{{{\rm{R}}}}}^{2}-{\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}} > 0\). (b-c) Pure non-linear mixed selectivity contribution (\(\Delta {{{{\rm{R}}}}}^{2}-{\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}}\)) for the interaction between the different blocks of variables across neuronal types and S1 layers. Results were qualitatively equivalent for the excitatory (a) and the inhibitory (b) populations.

Extended Data Fig. 8 Non-linear mixed selectivity for whisker contacts and angular position for the different time steps and cortical columns in S1.

(a) Pure non-linear mixed selectivity (\(\Delta {{{{\rm{R}}}}}^{2}-{\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}}\)) contribution for the interaction between contacts for the different time steps (time lags) and whiskers separately by columnar location (mean across neurons). The strongest interactions occur at the time lags of 100 ms (1 time step). Even though C1 column shows that C1 terms have the strongest interaction, C2 and C3 columns present a more heterogeneous interaction pattern. (b) Equivalent plot for the interaction between angular position for the different time steps (time lags) and whiskers. The strongest interactions occurs at time lags of 0ms. All columns present strong interactions terms with the rest of whiskers. (c) Pure non-linear mixed selectivity contribution (\(\Delta {{{{\rm{R}}}}}^{2}-{\Delta {{{{\rm{R}}}}}^{2}}_{{{{\rm{Linear}}}}}\)) for the interaction between contacts and angular position for the different time steps (time lags) and whiskers (single regressors in the encoding models) (mean across neurons). The strongest non-linear contribution in whisker angular position occurs on the current time step for all whiskers. The strongest non-linear contribution for the interaction between contacts and angular position occurs on time lags between angular position (and neuronal activity) and contacts of 100 ms (1 time step). (d) Equivalent plot for the interaction between contacts for the different time steps (time lags) and whiskers. The strongest non-linear contribution in whisker contacts occurs on time lags between neuronal activity and contacts of 100 ms (1 time step) for all whiskers.

Extended Data Fig. 9 Different whiskers are represented across S1 columns in approximately orthogonal sub-spaces.

(a) Top: Decoding performance on whether the sum of whisker contacts across the current and previous four time steps corresponded to a high or a low number of contacts with respect to the median for each different whisker (decode high vs low number of contacts for each whisker) (C1 blue; C2 green; C3 red) and columns (see Methods). Information about whisker contacts for all whiskers was present in all columns. Decoding Performance was evaluated on surrogate activity generated by the best encoding model (NonLin-1) for each recording session. In each recording session, activity from only one column was recorded. In order to compare information across columns, surrogate activity was generated for 10 neurons, which corresponded to the smallest number of simultaneously recorded neurons across recording sessions (columns). Bottom: Qualitatively equivalent results were found when the CCGP was evaluated (see Methods). All columns encode information about all whiskers in approximately orthogonal spaces. Errorbars in all panels correspond to s.e.m. across recording sessions (C1 column n = 6; C2 column n = 9; C3 column n = 6). (b) Top: Decoding performance on whether the total number of contacts in a trial (2 sec.) corresponded to a high or a low number of contacts with respect to the median for each different whisker (decode high vs low number of contacts for each whisker) (C1 blue; C2 green; C3 red) and columns using pseudopopulations of neurons (see Methods). All bars are significantly above chance. Bottom: CCGP with respect to the total number of contacts for the different whiskers and columns. Even though there seems to be some somatotopic structure on CCGP for pseudopopulations, the differences are small and all bars are significantly above chance. Errorbars correspond to the standard deviation across cross-validation iterations (n = 10).

Extended Data Fig. 10 The geometry of representations and encoding properties in RNNs change for the easy, the complex, and the very complex tasks.

(a) Probability of correct response (y-axis) as a function of time (x-axis) when the RNN was trained on the easy task 1 (left panel), the complex task 1 (central panel) and the very complex task (right panel). RNNs were trained on low input noise (λlow = 0.23 and λhigh = 0.77) (see Methods). For each network, additional readout weights on the activity of the artificial units were trained to perform the rest of the tasks (solid lines). While an RNN trained on easy task 1 produced the best performance for easy task 1, the neuronal representations were not well suited for the rest of tasks (left). When an RNN was trained on the complex 1 (central) and the very complex (right) tasks, it produced representations that allowed the performance of many different tasks. This came at the expense of losing performance for the easy task 1. Generalization performance, defined as cross-condition generalization performance (CCGP), was also tested for the three easy tasks (dashed lines) when the RNN was trained on the easy task 1 (left panel), the complex task 1 (central panel) and the very complex task (right panel) (see Methods). While abstraction (CCGP) is high for easy task 1 when the RNN was trained on the easy task 1, it is low for easy tasks 2 and 3. Since complex task 1 is defined as the 2D-XOR between C1 and C2, CCGP was higher for easy task 1 and 2 (C1 and C2) than for easy task 3 (C3). For the RNN trained on the very complex task, CCGP was significantly above chance for all easy-task variables. (b-c) The same qualitative results were obtained when medium (b; λlow = 0.3 and λhigh = 0.7) and high (c; λlow = 0.35 and λhigh = 0.65) noise levels were used instead. For all panels in (a-c) the performance curves correspond to the mean across random realizations of input patterns and tasks (n = 50) (see Methods). (d) Similar to Fig. 2d, linear and non-linear classification models that read out the input (x-axis), were trained to perform the easy (left panel; easy task 1), the complex (central panel; complex task 1) and the very complex tasks (right panel). On the easy task, both linear and non-linear classifiers performed equally well, as shown by decoding performances (y-axis) of the different models. On the contrary, only non-linear classifiers that allow for complex cue combination, performed above chance on the complex (central) and very complex (right) tasks. The behavioral results obtained on the whisker-based discrimination task (see Fig. 2) are aligned with the easy task (left panel). In all panels errorbars correspond to the s.e.m. across different network realizations (n = 4). (e) Similar to Fig. 4d, CV R2 on explaining artificial units activity (y-axis), is plotted against the different encoding models (x-axis). RNNs trained to perform tasks that require non-linear integration of sensory cues are better explained by an encoding model that allows for non-linear mixed selectivity (central and right panels), while a non-linear encoding scheme provide marginal additional explanatory power for the easy task RNN (left panel). The higher the complexity of the trained task, the higher the advantage of non-linear encoding models on explaining the activity of the artificial units. In contrast to (d), the encoding properties of S1 are aligned with those of RNNs trained to perform tasks that require non-linear combination of sensory evidence (see Fig. 4). Errorbars correspond to the s.e.m. across artificial units (n = 240; 4 networks × 60 units).

Supplementary information

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Nogueira, R., Rodgers, C.C., Bruno, R.M. et al. The geometry of cortical representations of touch in rodents. Nat Neurosci 26, 239–250 (2023). https://doi.org/10.1038/s41593-022-01237-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1038/s41593-022-01237-9

This article is cited by

Search

Quick links

Nature Briefing

Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.

Get the most important science stories of the day, free in your inbox. Sign up for Nature Briefing