Abstract
The connection patterns of neural circuits form a complex network. How signaling in these circuits manifests as complex cognition and adaptive behaviour remains the central question in neuroscience. Concomitant advances in connectomics and artificial intelligence open fundamentally new opportunities to understand how connection patterns shape computational capacity in biological brain networks. Reservoir computing is a versatile paradigm that uses high-dimensional, nonlinear dynamical systems to perform computations and approximate cognitive functions. Here we present conn2res: an open-source Python toolbox for implementing biological neural networks as artificial neural networks. conn2res is modular, allowing arbitrary network architecture and dynamics to be imposed. The toolbox allows researchers to input connectomes reconstructed using multiple techniques, from tract tracing to noninvasive diffusion imaging, and to impose multiple dynamical systems, from spiking neurons to memristive dynamics. The versatility of the conn2res toolbox allows us to ask new questions at the confluence of neuroscience and artificial intelligence. By reconceptualizing function as computation, conn2res sets the stage for a more mechanistic understanding of structure-function relationships in brain networks.
Similar content being viewed by others
Introduction
Brains are complex networks of anatomically connected and functionally interacting neurons that have the ability to seamlessly assimilate and interact with a perpetually changing external environment1. Sensory stimuli elicit signaling events within structural connectivity networks and manifest as patterned neural activity. These emergent neural dynamics are thought to support the computations that underlie cognition and adaptive behavior. However, a computational framework that describes how information processing and functional specialization occur in brain networks remains elusive. Developing such a framework would require understanding the multiple levels of the information-processing hierarchy, from how the brain’s network architecture shapes the complex activity patterns elicited by external stimuli, to how neural circuits extract from these evoked activity patterns the necessary information to compute with time-varying inputs.
How does network structure shape spatiotemporal patterns of neural activity, and how do neural dynamics support computations that underlie cognitive functions and behaviors? An important piece of the puzzle is the study of connectomics2. Technological and analytic advances in neuroimaging methods have made it possible to reconstruct the wiring patterns of nervous systems, yielding high-resolution connectomes of brains in multiple species3,4,5,6. The availability of connectomes has led to the formulation of a variety of models that aim to map network architecture to various functional aspects of the brain7, such as emergent neural dynamics8,9, functional co-activation patterns10, and inter-individual differences in behavior11,12,13. Multiple network features are correlated with emergent functional phenomena14,15,16,17,18,19, but there is no clear mechanistic link between the static network architecture and cognition.
Furthermore, descriptive studies of the connectome across different species provide evidence that structural connectivity networks display topological features that are thought to shape the segregation and integration of information20. For instance, the simultaneous presence of a highly clustered architecture of segregated modules promotes specialized information processing21,22,23,24,25,26,27, while a densely interconnected core of high-degree hubs shortens communication pathways and promotes the integration of information from distributed specialized domains28,29. How these ubiquitous organizational principles of the architecture of the brain confer computational capacity remains unknown.
Artificial intelligence offers alternative ways to approach the link between structure and function in brain networks that take into account computation30,31. Within the expanding spectrum of artificial neural network models, reservoir computing makes it possible to describe how recurrent neural circuits extract information from a continuous stream of external stimuli and how they approximate complex time-varying functions32,33. In reservoir networks learning occurs exclusively at the readout connections, and hence the main network architecture of the reservoir does not require specific weight calibration, remaining fixed throughout training. This eliminates a confounder while avoiding biologically implausible credit assignment problems such as the use of backpropagation training34. These reasons make reservoir computing an ideal paradigm to study the effects of connectome architecture on computation and learning. In this regard, machine-learning and artificial intelligence algorithms offer new ways to study structure-function relationships in brain networks by conceptualizing function as a computational property31,35.
Here we review the fundamentals of reservoir computing and how it can be applied to gain mechanistic insight about the information processing of biological neural circuits. We then present conn2res (https://github.com/netneurolab/conn2res), an open-source Python toolbox that implements connectomes as reservoirs to perform cognitive tasks. In the spirit of open-science, conn2res builds on top of and is interoperable with other third-party resources and research initiatives to offer an exhaustive set of experimental configurations/settings that researchers can experiment with. These include a comprehensive corpus of cognitive tasks spanning a wide spectrum of computational and behavioral paradigms, multiple local intrinsic dynamics, and various linear optimization algorithms for task learning. All of this combined with the possibility of implementing connectomes reconstructed at different scales and obtained from any imaging modality. We have added a tutorial section with several use-case examples to illustrate different types of inferences that the conn2res toolbox supports, as well as to showcase its flexibility in terms of network architecture, network dynamics, and task paradigm. While being inclusive of different modeling traditions, from microcircuits to whole-brain network models, conn2res contributes a novel way for researchers to explore the link between structure and function in biological brain networks.
Building a reservoir computer
Reservoir computing (RC) is an umbrella term that unifies two computational paradigms, liquid state machines32 and echo-state networks36. The two originated independently in the fields of computational neuroscience and machine-learning, respectively, with a common goal: exploiting the computational properties of complex, nonlinear dynamical systems37. However, the ideas encompassed by the RC paradigm had been around in different forms for more than two decades prior38,39,40. The conventional reservoir computing (RC) architecture consists of an input layer, followed by the reservoir and a readout module (Fig. 1a)32,36,37. Typically, the reservoir is a recurrent neural network (RNN) of nonlinear units, while the readout module is a simple linear model. The readout module is trained to read the activation states of the reservoir — elicited by an external input signal — and map them to the desired target output in a supervised manner. In contrast to traditional artificial RNNs, the recurrent connections within the reservoir are fixed and randomly assigned; only the connections between the reservoir and the readout module are learned (Fig. 1a)41.
So, how does RC work? RC capitalizes on the nonlinear response of high-dimensional dynamical systems, referred to as reservoirs. The reservoir performs a nonlinear projection of the input into a high-dimensional space. This transformation of the input converts nonlinearly separable signals into linearly separable ones such that a linear model in the readout module can be trained to map the transformed input to the desired output41. In other words, the reservoir converts inputs into rich dynamic patterns that contain integrated information about the history of inputs and are read out linearly to solve complex tasks. As long as the reservoir has sufficient built-in dynamical complexity and rich dynamics, a large variety of input-output mappings can be realized, including the approximation of complex time-varying functions, such as forecasting chaotic time series, considered to be a problem of high computational complexity. Under certain conditions, such as the presence of fading memory and separation properties, reservoirs can act as universal function approximators32,42,43,44.
The computational capabilities of the reservoir are thus determined by its dynamics, which arise from the interaction between the fixed network architecture of the reservoir and the equations or rules governing the time evolution of its internal units. Importantly, unlike traditional artificial neural networks (Fig. 2a), in RC the experimenter can specify the connectivity of the reservoir and the equations governing its local dynamics (Fig. 2b). Likewise, by tuning the parameters of the system, the experimenter can transition global network dynamics through qualitatively different dynamical regimes such as stability or chaos45. The RC paradigm thus offers the advantages that arbitrary network architectures and dynamics can be superimposed on the reservoir, providing a tool for neuroscientists to investigate how connectome organization and neural dynamics interact to support learning in biologically-informed reservoirs (Fig. 2c).
The architectural flexibility of RC is multi-scale: the network architecture of reservoirs can be informed by connectomes reconstructed at different spatial scales, from microcircuits to meso- and macro-scale networks (Fig. 1a). Depending on the context, the units of the reservoir represent either populations of neurons or entire brain regions. The choice of local dynamics is mainly determined by the spatial scale of the reservoir’s network, but the nature of the research question at hand should also be considered (Fig. 1a). In contrast to traditional RNNs, in which global network dynamics are determined by connectivity changes due to learning, RC allows us to impose not only different types of local dynamics, but global dynamics governing the population-level behavior can also be controlled. This means that the dynamical regime — or qualitative dynamical behavior — of the reservoir can be tuned to progressively transition from stable to chaotic, thus passing through a critical phase transition, or criticality46,47. By parametrically tuning the dynamics to be closer or further from criticality, RC allows us to investigate the effects of qualitatively different neural trajectories near this critical point on the computational performance of the reservoir47,48,49. An additional advantage of the dynamical and structural flexibility of RC is the possibility to enforce computational priors in the form of either functional or structural inductive biases50. Therefore, RC allows us to explore the functional consequences of information-processing strategies, such as critical dynamics or the presence of computational priors, thought to be exploited by biological brains45,46,47,51,52.
The fact that RC can be used with arbitrary network architectures and dynamics, plus the possibility of performing a variety of tasks spanning multiple cognitive domains — from perceptual-motor functions, memory and learning, to complex attention and executive functions — makes it ideal to investigate how specific network attributes and dynamics influence the neural computations that support cognition53. Specifically, by implementing various tasks along these multiple cognitive domains, connectome-informed reservoirs allow us to systematically map network structure and dynamics to a unique set of identifiable computational properties exclusive to the task at hand (Fig. 1b). In this way, the RC framework allows us to build a comprehensive structural-functional ontology, relating network structure and dynamics to fundamental blocks of computation and, ultimately, to cognitive function.
The application of this hybrid approach between artificial intelligence and neuroscience goes beyond exploring the link between structure and function in the healthy brain. For instance, it can be applied in the clinical setting to study how neurological diseases affect learning in the network. By comparing the performance of reservoirs informed by clinical populations against those informed by healthy controls, this framework allows us to investigate whether cognitive decline, measured as variations in computational capacity, can be explained by measurable changes in network architecture due to neurodegeneration (Fig. 1c). Another relevant application of the RC framework is the exploration of how the link between structure and function changes throughout adaptive processes such as development or evolution (Fig. 1c). For example, by implementing connectomes obtained throughout the lifespan or from different species, as reservoirs, this framework allows us to investigate how variations in network architecture translate into differences in computational capacity across ontogeny and phylogeny, respectively. In all cases, the RC framework allows for statistical significance testing by benchmarking empirical neural network architectures against random or null network models54. Altogether, this hybrid framework proposes a shift in the way structure-function relationships are studied in brain networks: from understanding function as a phenomenon (i.e., inter-regional functional interactions or functional activation maps), to a concept of function that is closer to the computational and information-processing properties of brain networks, thus contributing to a more mechanistic understanding of how computations and functional specialization emerge from the interaction between network architecture and dynamics in neural circuits.
Apart from the advantages that RC offers to the neuroscience community, this paradigm is also promising from an engineering point of view. Reservoirs can be realized using physical systems, substrates or devices, as opposed to – generally time- and energy-consuming — simulated RNNs55,56. In this regard, the architecture of these neuromorphic chips could benefit from the emerging understanding of connection patterns in biological networks57,58,59. For instance, systematically mapping combinations of network attributes and dynamical regimes to a range of computational functions could assist the design of ad hoc or problem-specific tailored architectures. Due to their physical nature, neuromorphic systems are limited by spatial, material and energetic constraints, akin to biological neural networks. Because of this, insights gained about the economical organization of brain networks could contribute to the cost-effective design of these information-processing systems35. Furthermore, the fact that training only occurs at the readout stage makes RC an extraordinarily computationally efficient learning approach. In addition to this, parallel information processing can be achieved by simultaneously training multiple readout modules to perform various parallel tasks. Therefore, physical RC and RC in general, provide a powerful method for faster and simpler multi-task learning, compared to other RNNs. Thanks to the dynamical and versatile nature of the reservoir, the RC paradigm is perfectly suited for a wide range of supervised tasks involving the processing of temporal and sequential data. These include: time series prediction, dynamical pattern generation, classification and segmentation, control, signal processing, and monitoring of rare events, among others60. Because of all these reasons, physical RC systems have become ideal candidates for the development of novel brain-inspired computing architectures61.
RESULTS
The conn2res toolbox
In a nutshell, conn2res is an open-source Python toolbox that allows users to implement biological neural networks as reservoirs to perform cognitive tasks (https://github.com/netneurolab/conn2res). The toolbox is built on top of the following well established, documented and maintained Python package dependencies: NumPy (https://numpy.org;62,63,64), SciPy (https://scipy.org;65), pandas (https://pandas.pydata.org;66), Scikit-Learn (https://scikit-learn.org;67), Gym (https://www.gymlibrary.dev;68), NeuroGym (https://neurogym.github.io;53), ReservoirPy (https://github.com/reservoirpy/reservoirpy;69), bctpy (https://github.com/aestrivex/bctpy;70), Seaborn (https://seaborn.pydata.org;71) and Matplotlib (https://matplotlib.org;72). The toolbox is also interoperable with other relevant Python packages including The Virtual Brain (https://www.thevirtualbrain.org;73), bio2art (https://github.com/AlGoulas/bio2art;74), NeuroSynth (https://neurosynth.org;75), Neuromaps (https://neuromaps-main.readthedocs.io;76) and the Enigma Toolbox (https://enigma-toolbox.readthedocs.io;77). Besides its extensive interoperability with other Python packages, a major strength of the conn2res toolbox is its flexibility in the choice of the different components that make part of the main RC workflow. The conn2res toolbox was expressly conceived as a tool for neuroscientists to explore a variety of hypotheses about structure-function coupling in brain networks. Therefore, compared to other RC-related packages, it offers higher flexibility in terms of network architecture, local dynamics, learning algorithms, task paradigms and performance metrics. To our knowledge, some of these are usually fixed or limited in other RC packages. Table S1 compares the conn2res toolbox against other well-known RC Python packages69,78, based on these criteria.
The baseline conn2res workflow requires the following input arguments (Fig. 3a): (i) task name or dataset: the name of the task to be performed, or a labeled dataset of input-target pairs for supervised learning can also be provided. conn2res is a wrapper of NeuroGym53, a curated collection of behavioral paradigms that were designed to facilitate the training of neural network models, and that are relevant for the neuroscience community. All of the 20+ tasks available in NeuroGym are also available in conn2res — some of these include perceptual decision making, context-dependent decision making, delayed comparison, delayed-paired association and delayed match category —; (ii) connectome: the connectivity matrix, which serves as the reservoir’s network architecture. The toolbox supports binary and weighted connectivity matrices of both directed and undirected networks; (iii) input nodes: the set of nodes that receive the external input signals concerning the task at hand; (iv) readout nodes: the set of nodes or modules from which information will be retrieved to train the linear model in the readout module; (v) reservoir local dynamics: the type of dynamics governing the activation of the reservoir’s units. Local dynamics can be split into two categories: discrete-time, governed by difference equations and continuous-time, based on differential equations. The former category includes both linear and nonlinear artificial neuron models with activation functions such as ReLU, leaky ReLU, sigmoid, and hyperbolic tangent, whereas the latter category includes a nonlinear spiking neuron model. The underlying leaky-integrate-and-fire neuron model is based on the framework proposed in79, with model parameters from80 and implements a double-exponential synaptic filter for the presynaptic spike trains. This flexible model provides higher biological plausibility, notably allowing for user-specified heterogeneous synaptic time constants, artificial stimulations/inhibitions, and interneuron-specific connectivity constraints79,81; (vi) linear model: specified as an instance of a linear model estimator from the Scikit-Learn library to be implemented for learning by the readout module67.
The typical conn2res workflow is depicted in Fig. 3b. In the first stage, fetch task dataset, a supervised dataset consisting of input-label pairs is either fetched from the conn2res repository, if the name of the task is provided by the user, or directly loaded if an external path is provided instead. In the second stage, set connectivity matrix, an instance of a reservoir object is created, and its network architecture and dynamics are set based on the connectivity matrix and the type of local nonlinearity specified by the user, respectively. In the third stage, simulate reservoir dynamics, the task inputs from the previous stage are introduced as external signals to the reservoir through the set of input nodes specified by the user. The dynamical models in the conn2res toolbox simulate the time evolution of the reservoir’s internal units (which are activated thanks to the propagation of the external input signals), and generate time series activity for each node in the reservoir. In the fourth stage, learning, the time series activity of the readout nodes or modules specified by the user are retrieved and passed on to the readout module, together with the task labels from the first stage. Both of these are used to train the linear model in the readout module. Finally, during the fifth and last stage, assess performance, depending on the nature of the reservoir, the final output can be either a single performance score, or a performance curve that displays performance as a function of the parameter that controls for the qualitative behavior of the reservoir’s dynamics (i.e., stable, critical or chaotic). Various performance metrics are currently available depending on whether the task requires a classification or a regression model. To facilitate the user’s experience, the toolbox provides several example scripts that illustrate use-case driven workflows.
The conn2res toolbox has a modular design. It consists of six modules, each one containing functions that support a specific step along the mainstream conn2res pipeline (Fig. 3c). The wrapper functions and classes used to generate the task datasets can be found in the tasks.py module. All types of manipulations on the connectivity matrix, such as binarization, weight scaling, normalization and rewiring, are handled by the Conn class in the connectivity.py module. Reservoir’s features including its network architecture, local dynamics and the retrieval of the reservoir’s activation states, are handled by the Reservoir class in the reservoir.py module. The functions in charge of the training and test of the linear model in the readout module are contained in the readout.py and performance.py modules, respectively. Finally, the plotting.py module offers a set of plotting tools that assist with the visualization of the different types of data generated along the pipeline, including the task input-output data, the 2D connectivity matrix of the reservoir’s network architecture, the simulated reservoir states, the decision function of the readout module, and the performance curve.
Tutorial
This section provides a broader overview of the multiple experimental settings and inferences that the conn2res toolbox supports. The first part consists of a detailed step-by-step example to illustrate the main conn2res workflow in action. The second part presents three applied cases in which specific hypotheses are proposed and tested using the toolbox. In each case, we evaluate the effect of global network architecture and dynamics on the computational capacity of reservoirs informed by connectomes of different animal species, reconstructed at different scales and obtained from different imaging modalities. Annotated notebooks and scripts to reproduce these results are included in the toolbox documentation https://github.com/netneurolab/conn2res/tree/master/examples.
Example 1: toolbox components
In this first example we quantify the effect of different types of local and global dynamics on the performance of a connectome-informed reservoir across two cognitive tasks: perceptual decision making82 and context-dependent decision making83. To do so, we implement an echo-state network36 whose connections are constrained by a human consensus connectome reconstructed from diffusion-weighted MRI data (n = 66 subjects. Data source: https://doi.org/10.5281/zenodo.2872624)84. To select the set of input and readout nodes, we use a functional connectivity-based partition of the connectome into intrinsic networks85. We define input nodes as a set of randomly selected brain regions from the visual system, and for the readout nodes we select all brain regions in the somatomotor system. Local dynamics are determined by the activation function of the reservoir’s units. Here we use sigmoid and hyperbolic tangent activation functions. Global network dynamics are set by parametrically tuning α, which corresponds to the spectral radius of the connectivity matrix86. The dynamics of the reservoir are considered to be stable if α < 1, and chaotic if α > 1. When α ≈ 1, the dynamics are said to be critical46. Because both tasks can be treated as supervised classification problems, we use a Ridge classifier model to train the readout module. We generate 1000 trials per task (70% training, 30% test), and we perform each task using 50 different realizations of the task labeled dataset. The distribution of performance scores is reported across the 50 instances of the task dataset.
Next we walk the reader through each of the steps along the main conn2res pipeline, and use the visualization tools included in the plotting.py module to depict the main output at each stage, facilitating the conceptual understanding of the workflow. Details about the practical implementation can be found in the examples folder of the conn2res toolbox. Results for the perceptual and context-dependent decision-making tasks are shown on the left and right columns of Fig. 4, respectively. Top panel in Fig. 4 consists of a single plot that displays the time series of the input (xi) and target labels (y) obtained during the task dataset fetching process. The perceptual decision-making task is a two-alternative forced choice task in which the reservoir must be able to integrate two stimuli to decide which one is higher on average (left column in Fig. 4). In the context-dependent decision-making task the reservoir has to perform one of two different perceptual discriminations, indicated by a contextual cue in every trial (right column in Fig. 4). Trials are delimited by vertical black dotted lines.
The second panel in Fig. 4 from top to bottom is a toy representation of the assignment of the connectome-based connectivity matrix to the reservoir’s network (center); it also shows the input nodes (blue nodes on the left) used for the introduction of the external input signal into the reservoir during the simulation of the reservoir’s dynamics, and the readout nodes (purple nodes on the right) used for the retrieval of information from the reservoir during the learning phase. The third panel in Fig. 4 depicts the simulation of the reservoir’s dynamics and it consists of two plots: the top plot presents the time series of the input signals (x), while the bottom plot shows the simultaneous reservoir’s activation state at every time step. These plots help the user visualize how reservoir states evolve as a function of the external inputs. The fourth panel in Fig. 4 makes reference to the learning process of the readout module during training. This panel contains three plots: the time series of the input signal (top), the decision function of the Ridge classifier (middle), and the predicted versus the ground truth target signals (bottom). Finally, the fifth panel in Fig. 4 shows the performance of the reservoir as a function of both local and global network dynamics. This panel presents two plots, each one corresponding to a different classification performance metrics: balanced accuracy (top) and F1 score (bottom). Each plot displays two curves that indicate how performance varies as a function of α, and each curve corresponds to a different activation function: hyperbolic tangent (pink) and sigmoid (green).
Results in Fig. 4 suggest that both local and global network dynamics have an impact on task performance. At the local level, both tasks benefit from having a hyperbolic tangent activation function, compared to the sigmoid. However, dependence of task performance on global network dynamics varies from one task to the other. In the perceptual decision-making task, a choice must be made based on the time integration of two past stimuli, which means that a temporal memory is required. Because stability enforces memory in the reservoir, computations required in the perceptual decision-making task should take advantage of stable network dynamics32,42,43,49. This is indeed the case: if the local nonlinearity is hyperbolic tangent, a decrease in performance from stable (α < 1) to chaotic (α > 1) dynamics is observed (pink lines in bottom panel on the left column of Fig. 4). If the local nonlinearity is a sigmoid, however, the reservoir does not show a strong dependence with respect to global network dynamics (green lines in bottom panel on the left column of Fig. 4). In contrast, in the context-dependent decision-making task, a binary perceptual discrimination must be made, and hence the reservoir must learn to differentiate between two temporal patterns. Because chaotic dynamics contribute to the separability property of a reservoir32,42,43,49, performance in this task should be enhanced by the presence of chaos. This is observed by an increase in performance as global network dynamics transition from stable to chaotic (pink and green lines in bottom panel on the right column of Fig. 4). Even though this is observed for both types of local nonlinearities - i.e., hyperbolic tangent and sigmoid - the effects are stronger for the hyperbolic tangent type. As expected, the effect of local and global network dynamics on task performance depends on the type of computations required by the task at hand.
This toy example helps us illustrate the flexibility of the conn2res toolbox in terms of choice of network architecture, local and global network dynamics, computational property, and performance metrics. Even though the type of experiments that the conn2res toolbox has been designed for have more of an exploratory character, we expect that as imaging technologies improve together with our understanding of the anatomical structure of biological brains, more hypothesis-driven of experiments can be carried out with conn2res.
Example 2: applications
In the second part of the tutorial we show how the toolbox can be applied to address three specific biological questions. First, we quantify the memory capacity of human white matter connectomes reconstructed using diffusion-weighted MRI (n = 66;84). Here we use subcortical regions as input nodes and cortical regions as readout nodes. We then ask whether memory capacity specifically depends on the topological organization of the brain, rather than low-level features such as density or degree sequence. To address this question, we compare memory capacity in a group-level, empirical connectome against memory capacity in a population of 500 randomly rewired null connectomes with preserved density and degree sequence54,87. Figure 5a shows that at criticality (α = 1), the memory capacity of empirical brain networks is significantly greater than in rewired nulls, suggesting that the topology of the human brain confers computational function35.
In the previous example we focused on global computational capacity and showed that it relates to global network topology. For the second example, we demonstrate how the toolbox can be used to make inferences about regional heterogeneity or specificity for computational capacity. To address this question we implement the perceptual decision-making task on a single subject-level, connectome-informed reservoir. We stratify cortical nodes according to their affiliation with the canonical intrinsic networks85. Specifically, we use brain regions in the visual network as input nodes, and the remaining networks separately as a readout module each to quantify task performance. Figure 5b shows prominent differentiation in performance depending on which network is chosen as the readout module. Interestingly, the two modules with the greatest performance are the default mode and somatomotor networks, consistent with the notion that perceptual decision making involves integration of sensory inputs, comparison with internal goals, and formulation of an appropriate motor response88. Collectively, these results demonstrate how connectome-based reservoir computing can be used to make inferences about the computational capacity of anatomically circumscribed neural circuits.
For the final example, we show how the toolbox can be applied to comparative questions in which a researcher seeks to compare networks. In this example, we implement connectomes reconstructed from four different species: fruit fly89, mouse90, rat91 and macaque92. As in the first example, we compare memory capacity in each empirical connectome with a population of 500 rewired null networks. Figure 5c shows that, despite differences in brain size, connectome resolution and reconstruction technique, the four model organism connectomes show a similar dependence on dynamics. Importantly, as with the human connectome, peak memory capacity is significantly greater in the empirical connectomes compared to the rewired nulls, except for the fruit fly, suggesting that this principle is potentially ubiquitous across nervous systems.
Discussion
Despite common roots, modern neuroscience and artificial intelligence have followed diverging paths. The advent of high-resolution connectomics and the incredible progress of artificial neural networks in recent years present fundamentally new and exciting opportunities for the convergence of these vibrant and fast-paced fields. Here we briefly summarized the principles of the RC paradigm and introduced conn2res, an open-source code initiative designed to promote cross-pollination of ideas and bridge multiple disciplines, including neuroscience, psychology, engineering, artificial intelligence, physics and dynamical systems. Below we look outward and propose how the conn2res toolbox can address emerging questions in these fields.
The conn2res toolbox embodies the versatility of the RC paradigm itself. By allowing arbitrary network architecture and dynamics to be superimposed on the reservoir, conn2res can be applied to investigate a wide range of neuroscience problems: from understanding the link between structure and function, studying individual differences in behavior, to exploring the functional consequences of network perturbations, such as disease or stimulation, or the computational benefits of specific architectural features, such as hierarchies and modules. The conn2res toolbox can readily accommodate network reconstructions at different spatial scales, from microcircuits to large-scale brain networks, and obtained using different imaging modalities, such as tract-tracing or diffusion MRI. Networks reconstructed at different points in either development and evolution can also be implemented in the toolbox to study, for instance, how structural adaptations across ontogeny and phylogeny shape computational capacity in brain networks. Collectively, conn2res offers new and numerous possibilities to discover how computation and functional specialization emerge from the brain’s anatomical network structure.
The RC paradigm can also be adapted to jointly study the influence of network wiring and spatial embedding on computation. Namely, the placement of connections in the brain is subject to numerous material, energetic and spatial constraints, a factor that is often overlooked in classical paradigms that focus exclusively on network topology93. Right now the models included in the conn2res toolbox do not explicitly take into account spatial embedding but they can be readily adapted to do so. One way is to introduce conduction delays that are proportional to inter-regional connection lengths or geodesic distances over the cortical surface94,95. Another interesting and slightly different approach that incorporates geometric constraints is the recently introduced concept of spatially-embedded recurrent neural networks (seRNNs)96. These are recurrent networks with adaptive weights, confined within a 3D Euclidean space, whose learning is constrained by biological optimization processes, like the minimization of wiring costs or the optimization of inter-regional communicability, in addition to the maximization of computational performance. When the pruning of the network is guided by these biological optimization principles, the resulting network architecture displays characteristic features of biological brain networks, such as modular structure with a small-world topology, and the emergence of functionally specialized regions that are spatially co-localized and implement an energetically-efficient, mixed-selective code96,97. More broadly the cortex, which is typically studied in these models, is part of a wider network of the central nervous system that is embedded in a perpetually changing environment. The RC paradigm can accommodate this “embodied” view of the brain. Specifically, RC models can include adaptive homeostatic mechanisms that regulate brain-environmental feedback loops to ensure that reservoirs are maintained in a desired dynamical state such as criticality98,99,100.
RC is often presented as a unified framework to train RNNs, but in a broader sense, it is a general framework to compute with high-dimensional, nonlinear dynamical systems, regardless of the choice of reservoir! Since any high-dimensional physical system with nonlinear dynamics could serve as reservoir — and these are abundant in both natural and man-made systems — a new field of applied research has emerged: physical reservoir computing. Here the goal is to exploit the rich dynamics of complex physical systems as information-processing devices. Physical substrates used for reservoirs are quite diverse: from analog circuits101,102,103,104, field programmable gate arrays105,106,107,108, photonic/opto-electronic devices109,110,111,112,113,114, spintronics115,116,117, quantum dynamics118,119, nanomaterials120,121,122,123,124,125,126, biological materials and organoids127,128,129,130,131,132,133, mechanics and robotics134,135,136, up to liquids or fluids137,138, and most recently, origami structures139. The development of physical reservoir systems has been accompanied by advances in more efficient and effective RC frameworks, for instance by including time delays140,141,142. As physical reservoir computing becomes more popular, we envision the use of conn2res as a workbench to explore the effect of network interactions on the computational properties of physical reservoirs. Anticipating this, conn2res is currently equipped with a dedicated class for physical reservoirs, which allows memristive networks — a promising alternative for neuromorphic computing122 — to be implemented as reservoirs. In this sense, the paradigm and the conn2res toolbox can be applicable to a wide variety of problems in adjacent scientific disciplines. From the neuro-connectomics perspective, conn2res offers new and numerous possibilities to discover how structure and function are linked in biological brain networks. From the artificial intelligence perspective, reverse-engineering biological networks will provide insights and novel design principles for re-engineering artificial, brain-inspired RC architectures and systems.
Altogether, conn2res is an easy-to-use toolbox that allows biological neural networks to be implemented as artificial neural networks. By combining connectomics and AI, the RC paradigm allows us to address new questions in a variety of scales of description and many adjacent fields. We hope that by reconceptualizing function as computation, conn2res allows us to take the next step towards understanding structure-function relationships in brain networks.
Data availability
The structural human connectome data used in the first and second parts of the tutorial section of the present report are publicly available at https://doi.org/10.5281/zenodo.287262484. The structural connectomes of the four model organisms used in the second part of the tutorial section are publicly available at: fruit fly (https://www.flycircuit.tw;89), mouse (http://connectivity.brain-map.org;90), rat (http://brancusi1.usc.edu/connections/grid/168;91) and macaque (supporting information for https://www.pnas.org/doi/epdf/10.1073/pnas.1008054107;92,143). To facilitate the reproduction of the results, all processed connectivity data used for the Tutorial section can be directly downloaded at https://doi.org/10.5281/zenodo.10205004144.
Code availability
Source code for conn2res is available on GitHub (https://github.com/netneurolab/conn2res) and is provided under the BSD 3-Clause “New" or “Revised" License. We have integrated conn2res with Zenodo (https://zenodo.org/doi/10.5281/zenodo.10437157145), which generates unique digital object identifiers (DOIs) for each new release of the toolbox. Researchers can access comprehensive online documentation via readthedocs (https://conn2res.readthedocs.io). Finally, as an open-source toolbox, conn2res is open to user suggestions and improvements, ensuring that it remains a continuously evolving resource. All code used for data processing, simulation, analysis, and figure generation relies on the following open-source Python packages: NumPy62,63,64, Scipy65, Pandas66, Scikit-learn67, bctpy https://github.com/aestrivex/bctpy70, Gym https://github.com/openai/gym68, NeuroGym https://github.com/neurogym/neurogym53, Matplotlib72, and Seaborn71.
References
Sporns, O. Structure and function of complex brain networks. Dialogues Clin. Neurosci. 15, 247–262 (2013).
Sporns, O., Tononi, G. & Kötter, R. The human connectome: a structural description of the human brain. PLoS Comput. Biol. 1, e42 (2005).
Insel, T. R., Landis, S. C. & Collins, F. S. The NIH brain initiative. Science 340, 687–688 (2013).
Van den Heuvel, M. P., Bullmore, E. T. & Sporns, O. Comparative connectomics. Trends Cogn. Sci. 20, 345–361 (2016).
Assaf, Y., Bouznach, A., Zomet, O., Marom, A. & Yovel, Y. Conservation of brain connectivity and wiring across the mammalian class. Nat. Neurosci. 23, 805–808 (2020).
Suárez, L. E. et al. A connectomics-based taxonomy of mammals. Elife 11, e78635 (2022).
Suárez, L. E., Markello, R. D., Betzel, R. F. & Misic, B. Linking structure and function in macroscale brain networks. Trends Cogn. Sci. 24, 302–315 (2020).
Breakspear, M. Dynamic models of large-scale brain activity. Nat. Neurosci. 20, 340–352 (2017).
Uddin, L. Q. Bring the noise: reconceptualizing spontaneous neural activity. Trends Cogn. Sci. 24, 734–746 (2020).
Cabral, J., Kringelbach, M. L. & Deco, G. Functional connectivity dynamically evolves on multiple time-scales over a static structural connectome: models and mechanisms. NeuroImage 160, 84–96 (2017).
Mišić, B. & Sporns, O. From regions to connections and networks: new bridges between brain and behavior. Curr. Opin. Neurobiol. 40, 1–7 (2016).
Seguin, C., Tian, Y. & Zalesky, A. Network communication models improve the behavioral and functional predictive utility of the human structural connectome. Netw. Neurosci. 4, 980–1006 (2020).
Melozzi, F. et al. Individual structural features constrain the functional connectome. Proc. Natl Acad. Sci. USA 116, 26961–26969 (2019).
Bettinardi, R. G. et al. How structure sculpts function: unveiling the contribution of anatomical connectivity to the brain’s spontaneous correlation structure. Chaos 27, 047409 (2017).
Goñi, J. et al. Resting-brain functional connectivity predicted by analytic measures of network communication. Proc. Natl Acad. Sci. USA 111, 833–838 (2014).
Zhang, M. & Saggar, M. Complexity of intrinsic brain dynamics shaped by multiscale structural constraints. Preprint at bioRxiv https://doi.org/10.1101/2020.05.14.097196 (2020).
Sporns, O. Network attributes for segregation and integration in the human brain. Curr. Opin. Neurobiol. 23, 162–171 (2013).
Fletcher, J. M. & Wennekers, T. From structure to activity: using centrality measures to predict neuronal activity. Int. J. Neural Syst. 28, 1750013 (2018).
Sethi, S. S., Zerbi, V., Wenderoth, N., Fornito, A. & Fulcher, B. D. Structural connectome topology relates to regional bold signal dynamics in the mouse brain. Chaos. 27, 047405 (2017).
Bertolero, M. A., Yeo, B. T., Bassett, D. S. & D’Esposito, M. A mechanistic model of connector hubs, modularity and cognition. Nat. Hum. Behav. 2, 765–777 (2018).
Watts, D. J. & Strogatz, S. H. Collective dynamics of ‘small-world’ networks. Nature 393, 440–442 (1998).
Sporns, O. & Zwi, J. D. The small world of the cerebral cortex. Neuroinformatics 2, 145–162 (2004).
Bassett, D. S. & Bullmore, E. Small-world brain networks. Neuroscientist 12, 512–523 (2006).
Chen, Z. J., He, Y., Rosa-Neto, P., Germann, J. & Evans, A. C. Revealing modular architecture of human brain structural networks by using cortical thickness from MRI. Cereb. Cortex 18, 2374–2381 (2008).
Betzel, R. F. et al. The modular organization of human anatomical brain networks: accounting for the cost of wiring. Netw. Neurosci. 1, 42–68 (2017).
Bertolero, M. A., Yeo, B. T. & D’Esposito, M. The modular and integrative functional architecture of the human brain. Proc. Natl Acad. Sci. USA 112, E6798–E6807 (2015).
Hilgetag, C. C. & Kaiser, M. Clustered organization of cortical connectivity. Neuroinformatics 2, 353–360 (2004).
Zamora-López, G., Zhou, C. & Kurths, J. Cortical hubs form a module for multisensory integration on top of the hierarchy of cortical networks. Front. Neuroinform. 4, 1 (2010).
van den Heuvel, M. P., Kahn, R. S., Goñi, J. & Sporns, O. High-cost, high-capacity backbone for global brain communication. Proc. Natl Acad. Sci. USA 109, 11372–11377 (2012).
Marblestone, A. H., Wayne, G. & Kording, K. P. Toward an integration of deep learning and neuroscience. Front. Comput. Neurosci. 10, 94 (2016).
Richards, B. A. et al. A deep learning framework for neuroscience. Nat. Neurosci. 22, 1761–1770 (2019).
Maass, W., Natschläger, T. & Markram, H. Real-time computing without stable states: a new framework for neural computation based on perturbations. Neural Comput. 14, 2531–2560 (2002).
Buonomano, D. V. & Maass, W. State-dependent computations: spatiotemporal processing in cortical networks. Nat. Rev. Neurosci. 10, 113 (2009).
Prince, L. Y. et al. Current state and future directions for learning in biological recurrent neural networks: a perspective piece. Preprint at arXiv https://doi.org/10.48550/arXiv.2105.05382 (2021).
Suárez, L. E., Richards, B. A., Lajoie, G. & Misic, B. Learning function from structure in neuromorphic networks. Nat. Mach. Intell. 3, 771–786 (2021).
Jaeger, H. The “echo state” approach to analysing and training recurrent neural networks-with an erratum note. Bonn Ger. Ger. Natl Res. Cent. Inf. Technol. GMD Tech. Rep. 148, 13 (2001).
Verstraeten, D., Schrauwen, B., d’Haene, M. & Stroobandt, D. An experimental unification of reservoir computing methods. Neural Netw. 20, 391–403 (2007).
Dominey, P. F. & Arbib, M. A. A cortico-subcortical model for generation of spatially accurate sequential saccades. Cereb. Cortex 2, 153–175 (1992).
Dominey, P. F. Complex sensory-motor sequence learning based on recurrent state representation and reinforcement learning. Biol Cybern. 73, 265–274 (1995).
Dominey, P., Arbib, M. & Joseph, J.-P. A model of corticostriatal plasticity for learning oculomotor associations and sequences. J. Cogn. Neurosci. 7, 311–336 (1995).
Lukoševičius, M. & Jaeger, H. Reservoir computing approaches to recurrent neural network training. Comput. Sci. Rev. 3, 127–149 (2009).
Legenstein, R. & Maass, W. Edge of chaos and prediction of computational performance for neural circuit models. Neural Netw. 20, 323–334 (2007).
Legenstein, R. & Maass, W. What makes a dynamical system computationally powerful. In New Directions in Statistical Signal Processing: From Systems to Brain 127–154 (The MIT Press, 2006).
Maass, W. & Markram, H. On the computational power of circuits of spiking neurons. J. Comput. Syst. Sci. 69, 593–616 (2004).
Deco, G. & Jirsa, V. K. Ongoing cortical activity at rest: criticality, multistability, and ghost attractors. J. Neurosci. 32, 3366–3375 (2012).
O’Byrne, J. & Jerbi, K. How critical is brain criticality? Trends Neurosci. 45, 820–837 (2022).
Cocchi, L., Gollo, L. L., Zalesky, A. & Breakspear, M. Criticality in the brain: a synthesis of neurobiology, models and cognition. Prog. Neurobiol. 158, 132–152 (2017).
Langton, C. Computation at the edge of chaos: phase transition and emergent computation. Phys. D Nonlinear Phenom. 42, 12–37 (1990).
Bertschinger, N. & Natschläger, T. Real-time computation at the edge of chaos in recurrent neural networks. Neural Comput. 16, 1413–1436 (2004).
Subramoney, A., Scherr, F. & Maass, W. Reservoirs learn to learn. In Reservoir Computing: Theory, Physical Implementations, and Applications 59–76 (Springer, Singapore, 2021).
Tagliazucchi, E., Balenzuela, P., Fraiman, D. & Chialvo, D. R. Criticality in large-scale brain fMRI dynamics unveiled by a novel point process analysis. Front. Physiol. 3, 15 (2012).
Kitzbichler, M. G., Smith, M. L., Christensen, S. R. & Bullmore, E. Broadband criticality of human brain network synchronization. PLoS Comput. Biol. 5, e1000314 (2009).
Molano-Mazon, M. et al. Neurogym: an open resource for developing and sharing neuroscience tasks. Preprint at PsyArXiv https://doi.org/10.31234/osf.io/aqc9n (2022).
Váša, F. & Mišić, B. Null models in network neuroscience. Nat. Rev. Neurosci. 23, 493–504 (2022).
Tanaka, G. et al. Recent advances in physical reservoir computing: a review. Neural Netw. 115, 100–123 (2019).
Nakajima, K. Physical reservoir computing–an introductory perspective. Jpn. J. Appl. Phys. 59, 060501 (2020).
Loeffler, A. et al. Modularity and multitasking in neuro-memristive reservoir networks. Neuromorphic Comput. Eng. 1, 014003 (2021).
Loeffler, A. et al. Neuromorphic learning, working memory, and metaplasticity in nanowire networks. Sci. Adv. 9, eadg3289 (2023).
Fu, K. et al. Reservoir computing with neuromemristive nanowire networks. In 2020 International Joint Conference on Neural Networks (IJCNN) 1–8 (IEEE, 2020).
Lukoševičius, M., Jaeger, H. & Schrauwen, B. Reservoir computing trends. KI-Künstliche Intell. 26, 365–371 (2012).
Dale, M., Miller, J. F., Stepney, S. & Trefzer, M. A. Reservoir computing in material substrates. In Reservoir Computing: Theory, Physical Implementations, and Applications 141–166 (Springer, Singapore, 2021).
Harris, C. R. et al. Array programming with NumPy. Nature 585, 357–362 (2020).
Walt, Svd, Colbert, S. C. & Varoquaux, G. The NumPy array: a structure for efficient numerical computation. Comput. Sci. Eng. 13, 22–30 (2011).
Oliphant, T. E. A guide to NumPy, vol. 1 (Trelgol Publishing USA, 2006).
Virtanen, P. et al. Scipy 1.0: fundamental algorithms for scientific computing in Python. Nat. Methods 17, 261–272 (2020).
McKinney, W. et al. Data structures for statistical computing in Python. In Proceedings of the 9th Python in Science Conference, Vol. 445, 51–56 (Austin, TX, 2010).
Pedregosa, F. et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 12, 2825–2830 (2011).
Brockman, G. et al. OpenAI Gym. Preprint at bioRxiv https://doi.org/10.48550/arXiv.1606.01540 (2016).
Trouvain, N., Pedrelli, L., Dinh, T. T. & Hinaut, X. ReservoirPy: an efficient and user-friendly library to design echo state networks. In Artificial Neural Networks and Machine Learning – ICANN 2020 494–505 (Springer International Publishing, 2020). https://doi.org/10.1007/978-3-030-61616-8_40.
Rubinov, M. & Sporns, O. Complex network measures of brain connectivity: uses and interpretations. Neuroimage 52, 1059–1069 (2010).
Waskom, M. et al. seaborn: v0.7.0. Zenodo. https://doi.org/10.5281/zenodo.45133 (2016).
Hunter, J. D. Matplotlib: a 2d graphics environment. Comput. Sci. Eng. 9, 90–95 (2007).
Ritter, P., Schirner, M., McIntosh, A. R. & Jirsa, V. K. The virtual brain integrates computational modeling and multimodal neuroimaging. Brain Connect. 3, 121–145 (2013).
Goulas, A., Damicelli, F. & Hilgetag, C. C. Bio-instantiated recurrent neural networks: Integrating neurobiology-based network topology in artificial networks. Neural Netw. 142, 608–618 (2021).
Yarkoni, T., Poldrack, R., Nichols, T., Van Essen, D. & Wager, T. NeuroSynth: a new platform for large-scale automated synthesis of human functional neuroimaging data. In Frontiers in Neuroinformatics Conference Abstract: 4th INCF Congress of Neuroinformatics. https://doi.org/10.3389/conf.fninf.2011.08.00058 (2011).
Markello, R. D. et al. Neuromaps: structural and functional interpretation of brain maps. Nat. Methods 19, 1472–1479 (2022).
Larivière, S. et al. The enigma toolbox: multiscale neural contextualization of multisite neuroimaging datasets. Nat. Methods 18, 698–700 (2021).
Damicelli, F. echoes: Echo state networks with python. https://github.com/fabridamicelli/echoes (2019).
Kim, R., Li, Y. & Sejnowski, T. J. Simple framework for constructing functional spiking recurrent neural networks. Proc. Natl Acad. Sci. USA 116, 22811–22820 (2019).
Nicola, W. & Clopath, C. Supervised learning in spiking neural networks with force training. Nat. Commun. 8, 2208 (2017).
Song, H. F., Yang, G. R. & Wang, X.-J. Training excitatory-inhibitory recurrent neural networks for cognitive tasks: a simple and flexible framework. PLoS Comput. Biol. 12, e1004792 (2016).
Britten, K. H., Shadlen, M. N., Newsome, W. T. & Movshon, J. A. The analysis of visual motion: a comparison of neuronal and psychophysical performance. J. Neurosci. 12, 4745–4765 (1992).
Mante, V., Sussillo, D., Shenoy, K. V. & Newsome, W. T. Context-dependent computation by recurrent dynamics in prefrontal cortex. Nature 503, 78–84 (2013).
Griffa, A., Alemán-Gómez, Y., & Hagmann, P. Structural and functional connectome from 70 young healthy adults [Data set]. Zenodo. https://doi.org/10.5281/zenodo.2872624 (2019).
Thomas Yeo, B. et al. The organization of the human cerebral cortex estimated by intrinsic functional connectivity. J. Neurophysiol. 106, 1125–1165 (2011).
Seung, H. S. How the brain keeps the eyes still. Proc. Natl Acad. Sci. USA 93, 13339–13344 (1996).
Maslov, S. & Sneppen, K. Specificity and stability in topology of protein networks. Science 296, 910–913 (2002).
Heekeren, H. R., Marrett, S. & Ungerleider, L. G. The neural systems that mediate human perceptual decision making. Nat. Rev. Neurosci. 9, 467–479 (2008).
Chiang, A.-S. et al. Three-dimensional reconstruction of brain-wide wiring networks in drosophila at single-cell resolution. Curr. Biol. 21, 1–11 (2011).
Rubinov, M., Ypma, R. J., Watson, C. & Bullmore, E. T. Wiring cost and topological participation of the mouse brain connectome. Proc. Natl Acad. Sci. USA 112, 10032–10037 (2015).
Bota, M., Sporns, O. & Swanson, L. W. Architecture of the cerebral cortical association connectome underlying cognition. Proc. Natl Acad. Sci. USA 112, E2093–E2101 (2015).
Modha, D. S. & Singh, R. Network architecture of the long-distance pathways in the macaque brain. Proc. Natl Acad. Sci. USA 107, 13485–13490 (2010).
Pang, J. C. et al. Geometric constraints on human brain function. Nature 618, 566–574 (2023).
Deco, G., Jirsa, V., McIntosh, A. R., Sporns, O. & Kötter, R. Key role of coupling, delay, and noise in resting brain fluctuations. Proc. Natl Acad. Sci. USA 106, 10302–10307 (2009).
Deco, G., Jirsa, V. K. & McIntosh, A. R. Emerging concepts for the dynamical organization of resting-state activity in the brain. Nat. Rev. Neurosci. 12, 43–56 (2011).
Achterberg, J., Akarca, D., Strouse, D., Duncan, J. & Astle, D. E. Spatially embedded recurrent neural networks reveal widespread links between structural and functional neuroscience findings. Nat. Mach. Intell. 5, 1369–1381 (2023).
Rigotti, M. et al. The importance of mixed selectivity in complex cognitive tasks. Nature 497, 585–590 (2013).
Hellyer, P. J., Clopath, C., Kehagia, A. A., Turkheimer, F. E. & Leech, R. From homeostasis to behavior: Balanced activity in an exploration of embodied dynamic environmental-neural interaction. PLoS Comput. Biol. 13, e1005721 (2017).
Hellyer, P. J., Jachs, B., Clopath, C. & Leech, R. Local inhibitory plasticity tunes macroscopic brain dynamics and allows the emergence of functional brain networks. NeuroImage 124, 85–95 (2016).
Falandays, J. B., Yoshimi, J., Warren, W. H. & Spivey, M. J. A potential mechanism for Gibsonian resonance: Behavioral entrainment emerges from local homeostasis in an unsupervised reservoir network. Cogn. Neurodyn. 1–24 (2023).
Appeltant, L. et al. Information processing using a single dynamical node as complex system. Nat. Commun. 2, 1–6 (2011).
Soriano, M. C. et al. Delay-based reservoir computing: noise effects in a combined analog and digital implementation. IEEE Trans. Neural Netw. Learn. Syst. 26, 388–393 (2014).
Li, J., Bai, K., Liu, L. & Yi, Y. A deep learning based approach for analog hardware implementation of delayed feedback reservoir computing system. In 2018 19th International Symposium on Quality Electronic Design (ISQED), 308–313 (IEEE, 2018).
Zhao, C. et al. Novel spike based reservoir node design with high performance spike delay loop. In Proceedings of the 3rd ACM International Conference on Nanoscale Computing and Communication 1–5 (Association for Computing Machinery, New York, NY, United States, 2016).
Antonik, P. Application of FPGA to Real-Time Machine Learning: Hardware Reservoir Computers and Software Image Processing (Springer, 2018).
Alomar, M. L., Canals, V., Martínez-Moll, V. & Rosselló, J. L. Low-cost hardware implementation of reservoir computers. In 2014 24th International Workshop on Power and Timing Modeling, Optimization and Simulation (PATMOS), 1–5 (IEEE, 2014).
Antonik, Piotr. Application of FPGA to Real‐Time Machine Learning: Hardware Reservoir Computers and Software Image Processing. (Springer, 2018).
Wang, Q., Li, Y., Shao, B., Dey, S. & Li, P. Energy efficient parallel neuromorphic architectures with approximate arithmetic on FPGA. Neurocomputing 221, 146–158 (2017).
Vandoorne, K. et al. Toward optical signal processing using photonic reservoir computing. Opt. Express 16, 11182–11192 (2008).
Vandoorne, K. et al. Experimental demonstration of reservoir computing on a silicon photonics chip. Nat. Commun. 5, 1–6 (2014).
Zhang, H. et al. Integrated photonic reservoir computing based on hierarchical time-multiplexing structure. Opt. Express 22, 31356–31370 (2014).
Katumba, A., Freiberger, M., Bienstman, P. & Dambre, J. A multiple-input strategy to efficient integrated photonic reservoir computing. Cogn. Comput. 9, 307–314 (2017).
Katumba, A. et al. Low-loss photonic reservoir computing with multimode photonic integrated circuits. Sci. Rep. 8, 1–10 (2018).
Laporte, F., Katumba, A., Dambre, J. & Bienstman, P. Numerical demonstration of neuromorphic computing with photonic crystal cavities. Opt. Express 26, 7955–7964 (2018).
Taniguchi, T. et al. Reservoir computing based on spintronics technology. In Reservoir Computing: Theory, Physical Implementations, and Applications 331–360 (Springer, Singapore, 2021).
Riou, M. et al. Reservoir computing leveraging the transient non-linear dynamics of spin-torque nano-oscillators. In Reservoir Computing: Theory, Physical Implementations, and Applications 307–329 (Springer, Singapore, 2021).
Nomura, H., Kubota, H. & Suzuki, Y. Reservoir computing with dipole-coupled nanomagnets. In Reservoir Computing: Theory, Physical Implementations, and Applications 361–374 (Springer, Singapore, 2021).
Fujii, K. & Nakajima, K. Quantum reservoir computing: a reservoir approach toward quantum machine learning on near-term quantum devices. In Reservoir Computing: Theory, Physical Implementations, and Applications 423–450 (Springer, Singapore, 2021).
Negoro, M., Mitarai, K., Nakajima, K. & Fujii, K. Toward nmr quantum reservoir computing. In Reservoir Computing: Theory, Physical Implementations, and Applications 451–458 (Springer, Singapore, 2021).
Yang, X., Chen, W. & Wang, F. Z. Investigations of the staircase memristor model and applications of memristor-based local connections. Analog Integr. Circuits Signal Process. 87, 263–273 (2016).
Bennett, C. H., Querlioz, D. & Klein, J.-O. Spatio-temporal learning with arrays of analog nanosynapses. In 2017 IEEE/ACM International Symposium on Nanoscale Architectures (NANOARCH) 125–130 (IEEE, 2017).
Kulkarni, M. S. & Teuscher, C. Memristor-based reservoir computing. In 2012 IEEE/ACM international symposium on nanoscale architectures (NANOARCH) 226–232 (IEEE, 2012).
Du, C. et al. Reservoir computing using dynamic memristors for temporal information processing. Nat. Commun. 8, 1–10 (2017).
Sillin, H. O. et al. A theoretical and experimental study of neuromorphic atomic switch networks for reservoir computing. Nanotechnology 24, 384004 (2013).
Kendall, J. D., Nino, J. C. & Suárez, L. E. Deep learning in bipartite memristive networks. US Patent App. 15/985,212 (2008).
Suárez, L. E., Kendall, J. D. & Nino, J. C. Evaluation of the computational capabilities of a memristive random network (mn3) under the context of reservoir computing. Neural Netw. 106, 223–236 (2018).
Cai, H. et al. Brain organoid reservoir computing for artificial intelligence. Nat. Electron. 6, 1032–1039 (2023).
Nakajima, K. et al. A soft body as a reservoir: case studies in a dynamic model of octopus-inspired soft robotic arm. Front. Comput. Neurosci. 7, 91 (2013).
Obien, M. E. J., Deligkaris, K., Bullmann, T., Bakkum, D. J. & Frey, U. Revealing neuronal function through microelectrode array recordings. Front. Neurosci. 8, 423 (2015).
Ortman, R. L., Venayagamoorthy, K. & Potter, S. M. Input separability in living liquid state machines. In Adaptive and Natural Computing Algorithms: 10th International Conference, ICANNGA 2011, Ljubljana, Slovenia, April 14-16, 2011, Proceedings, Part I 10 220–229 (Springer, Berlin, 2011).
Dockendorf, K. P., Park, I., He, P., Príncipe, J. C. & DeMarse, T. B. Liquid state machines and cultured cortical networks: the separation property. Biosystems 95, 90–97 (2009).
Jones, B., Stekel, D., Rowe, J. & Fernando, C. Is there a liquid state machine in the bacterium Escherichia coli? In 2007 IEEE Symposium on Artificial Life,187–191 (IEEE, 2007).
Didovyk, A. et al. Distributed classifier based on genetically engineered bacterial cell cultures. ACS Synth. Biol. 4, 72–82 (2015).
Hauser, H. Physical reservoir computing in robotics. In Reservoir Computing: Theory, Physical Implementations, and Applications 169–190 (Springer, Singapore, 2021).
Dion, G., Oudrhiri, A. I.-E., Barazani, B., Tessier-Poirier, A. & Sylvestre, J. Reservoir computing in MEMS. In Reservoir Computing: Theory, Physical Implementations, and Applications 191–217 (Springer, Singapore, 2021).
Caluwaerts, K., D’Haene, M., Verstraeten, D. & Schrauwen, B. Locomotion without a brain: physical reservoir computing in tensegrity structures. Artif. Life 19, 35–66 (2013).
Natschläger, T., Maass, W. & Markram, H. The" liquid computer": a novel strategy for real-time computing on time series. Telematik 8, 39–43 (2002).
Nakajima, K. & Aoyagi, T. The memory capacity of a physical liquid state machine. IEICE Tech. Rep. 115, 109–112 (2015).
Bhovad, P. & Li, S. Physical reservoir computing with origami and its application to robotic crawling. Sci. Rep. 11, 13002 (2021).
Duan, X.-Y. et al. Embedding theory of reservoir computing and reducing reservoir network using time delays. Phys. Rev. Res. 5, L022041 (2023).
Sakemi, Y., Morino, K., Leleu, T. & Aihara, K. Model-size reduction for reservoir computing by concatenating internal states through time. Sci. Rep. 10, 21794 (2020).
Larger, L. et al. High-speed photonic reservoir computing using a time-delay-based architecture: Million words per second classification. Phys. Rev. X 7, 011015 (2017).
Bakker, R., Wachtler, T. & Diesmann, M. Cocomac 2.0 and the future of tract-tracing databases. Front. Neuroinform. 6, 30 (2012).
Suárez, L. E. The conn2res toolbox [data set]. Zenodo (2023).
Suárez, L. E. et al. Connectome-based reservoir computing with the conn2res toolbox (2023). https://doi.org/10.5281/zenodo.10437157.
Werbos, P. J. Backpropagation through time: what it does and how to do it. Proc. IEEE 78, 1550–1560 (1990).
Acknowledgements
We thank Bertha Vazquez-Rodriguez, Andrea Luppi, Ross Markello, Golia Shafiei, Vincent Bazinet, Justine Hansen, Zhen-Qi Liu, Eric Ceballos, Moohebat Pourmajidian and Asa Farahani for insightful comments on the manuscript. BM acknowledges support from the Natural Sciences and Engineering Research Council of Canada (NSERC Discovery Grant RGPIN #017-04265), from the Brain Canada Future Leaders Fund and the Canadian Institutes of Health Research (CIHR). GL acknowledges support from NSERC (Discovery Grant: RGPIN-2018-04821), CIFAR (Canada AI Chair), and a Canada Research Chair in Neural Computations and Interfacing (CIHR, tier 2). PEV acknowledges support from MQ: Transforming Mental Health (Grant No. MQF17-24). All research from the Department of Psychiatry at the University of Cambridge is made possible by the National Institute for Health and Care Research Cambridge Biomedical Research Centre and National Institute for Health and Care Research East of England Applied Research Centre. LES acknowledges support from the Fonds de Recherche du Québec - Nature et Technologies (FRQNT) Strategic Clusters Program (2020-RS4-265502 - Centre UNIQUE - Union Neurosciences and Artificial Intelligence - Quebec) and the Fonds de Recherche du Québec - Nature et Technologies (FRQNT).
Author information
Authors and Affiliations
Contributions
L.E.S., A.M., F.M., and B.M. conceived the toolbox. L.E.S. and B.M. wrote the manuscript, with valuable revision from P.E.V., A.M. and G.L. L.E.S. and A.M. developed the toolbox with help from K.M., M.L., and F.M. B.M. was the project administrator.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Peer review
Peer review information
Nature Communications thanks Robert Leech, and the other, anonymous, reviewer(s) for their contribution to the peer review of this work. A peer review file is available.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Suárez, L.E., Mihalik, A., Milisav, F. et al. Connectome-based reservoir computing with the conn2res toolbox. Nat Commun 15, 656 (2024). https://doi.org/10.1038/s41467-024-44900-4
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41467-024-44900-4
This article is cited by
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.