Abstract
The processing of information is an indispensable property of living systems realized by networks of active processes with enormous complexity. They have inspired many variants of modern machine learning, one of them being reservoir computing, in which stimulating a network of nodes with fading memory enables computations and complex predictions. Reservoirs are implemented on computer hardware, but also on unconventional physical substrates such as mechanical oscillators, spins, or bacteria often summarized as physical reservoir computing. Here we demonstrate physical reservoir computing with a synthetic active microparticle system that self-organizes from an active and passive component into inherently noisy nonlinear dynamical units. The self-organization and dynamical response of the unit are the results of a delayed propulsion of the microswimmer to a passive target. A reservoir of such units with a self-coupling via the delayed response can perform predictive tasks despite the strong noise resulting from the Brownian motion of the microswimmers. To achieve efficient noise suppression, we introduce a special architecture that uses historical reservoir states for output. Our results pave the way for the study of information processing in synthetic self-organized active particle systems.
Similar content being viewed by others
Introduction
Storing and processing of information is vital for living systems1. The detection of low amounts of chemicals by a bacterium to navigate environments2,3, the feedback mechanisms controlling and maintaining the function of organisms4, or the highly sophisticated computations in large biological neural networks in the brain5 are intricate examples of this importance created by evolutionary development. All these processes with living matter as the substrate of computation rely on its inherent activity, e.g., the microscopic energy conversion to power the signaling cascades in the presence of strong thermal noise. They have inspired many computational models of machine learning that are not executed on living matter, but on well-designed electronic hardware using completely different information representations than living matter6. Recurrent neural networks are a variant of such mathematical algorithms with a fading memory that allows learning from information sequences as in language or time series7. Reservoir computers employ sparsely and statically connected recurrent nodes8,9,10 or even a single node by using time-multiplexing11,12 to create a high-dimensional space. Information can be injected into this space to spread over the many degrees of freedom. Unlike the training of other neural networks, where the interactions of all components are optimized, training reservoir computers is often only restricted to finding how the desired information can be retrieved from the node states using adjustable readout only13,14.
As one of the main properties of recurrent nodes is the memory of past states, reservoir computers also allow for a physical realization on unconventional computational substrates15,16,17 using optoelectronic oscillators18,19,20, mechanical oscillators21,22, carbon nanotubes23 or passive soft bodies15 as excitable physical systems. It is thus intriguing to close the loop and draw inspiration from active living systems to explore microscopic reservoir computing in synthetic active microsystems, where noise is omnipresent as well but a precise control over the shape and the physics of the active system is possible. Motile synthetic active particles have generated enormous interests as a model for self-propelled systems far from equilibrium and emergent collective effects24 that mimic, for example, the dynamics of swarms25,26,27. Information processing28,29 and machine learning30 in experiment31 and simulation32,33,34,35,36 have also entered the field of synthetic active matter. However, studies that extend the use of synthetic microparticles as computational substrates are still rare or purely numerical37.
We demonstrate that motile self-propelled active microparticles can be used for physical reservoir computing. An active particle self-organizes into a nonlinear dynamical unit based on a retarded propulsion towards an immobile target forming a noisy physical recurrent node. The node is perturbed by a time-multiplexed input signal to form a network of virtual nodes with sparsely connected topology. Multiple of these active units realize a high dimensional space of our reservoir computer. Harnessing the physics and inherent dynamics of the active particles, the reservoir computer is capable of predicting chaotic series despite the strong influence of the Brownian motion of the active particles. In particular, we find that using historical reservoir states for the output derivation effectively suppresses the intrinsic noise of the reservoir opening new routes for reservoir computing in noisy systems.
Results
Active particle recurrent node
A reservoir computer (RC), as a paradigm derived from recurrent neural networks, consists of recurrent nodes that nonlinearly process external signal inputs as well as their previous states8,10. We realize a simple recurrent node with the help of a single synthetic active particle as a microscopic model for motile active matter24. An active particle is a polymer microbead of 2.19 μm diameter with the surface decorated by gold nanoparticles (about 8 nm diameter). It is immersed in a thin layer of water bounded by two coverslips and can move freely in two dimensions. The active particle is propelled with a speed v038 towards an immobile target particle by partially heating the gold nanoparticles using a focused laser in a microscopy setup (see Fig. 1a, b and Methods section). The continuous propulsion in a desired direction is realized by adjusting the focused laser spot position in real time with the help of a feedback loop using a spatial light modulator (SLM). A time delay δt is added to the intrinsic feedback latency to control the active particle27,39 and to introduce a finite reaction time, as is inherent in many processes in living systems. Correspondingly, the attraction \(\hat{{{{{{{{\bf{F}}}}}}}}}(t)\) experienced by the active particle at time t is determined by its position at previous time t − δt,
where r denotes the location of the active particle with respect to the immobile particle center. The consequence of this retardation is an angular displacement θ(t) = ϕ(t) − ϕ(t − δt) during the delay time with ϕ denoting its angular position, thus results in a transient rotational motion of the active particle around the immobile target (Fig. 1c). The dynamics of the angular position can then be described by a nonlinear delay differential equation
assuming that the active and the immobile particle (radius of aact and aimm) are in physical contact, i.e., R0 = aact + aimm, under the delayed attraction. The active particle in water is subject to Brownian motion as represented by the noise term in Eq. (2) with w(t) denoting Gaussian white noise. The diffusion coefficient was determined from the experiment to be D = 0.08 μm2 s−1, giving rise to a Péclet number of Pe = aactv0/D = 38.7. To make the physical recurrent node capable of receiving external inputs, we introduce the u(t) in Eq. (2) representing an angular deviation of the particle propulsion direction from \(\hat{{{{{{{{\bf{F}}}}}}}}}\) (Fig. 1c).
For its function as a recurrent node, the dynamics of the angle θ(t) is important. Described by Eq. (2), the dynamics of θ can be approximated by an overdamped motion of a particle in a self-generated effective quartic potential Ueff(θ)27, which resembles a generic Landau-type description27,40. The shape of the potential is controlled by a dimensionless parameter θ0 = v0δt/R0. With increasing θ0, Ueff(θ) transitions from a near parabolic shape with a minimum at θ = 0 to a symmetric double well shape with minima at θ+ and θ− in a pitchfork bifurcation (see Fig. 2c and Supplementary Note 1).
A perturbation of θ (solid arrow in Fig. 2c) will result in a relaxation with a dynamics determined by the control parameter θ0. We have experimentally determined the response of θ(t) to an impulsive perturbation u(t) = δ(t) for different values of θ0 (Fig. 2a, b) active particle. However, the ensemble average of 500 trajectories of θ(t) (red lines in Fig. 2a) exhibits an asymptotic behavior, which nicely reflects the response evaluated in the deterministic simulation (Fig. 2b dashed lines). The characteristic relaxation time τθ extracted from the deterministic simulations reveals the expected strong increase around the transition point. Tuning the control parameter θ0, i.e., activity (v0) and/or delay (δt) therefore allows to manipulate the fading memory of the active particle as a recurrent node, which is paramount for the coupling and response of the recurrent nodes in our reservoir computer. Note that the nonlinear dynamics of the active particle system is the result of the delayed propulsion towards the target.
Reservoir computer with active particle nodes
The asymptotic relaxation of θ(t) demonstrated in Fig. 2b represents a basic requirement for reservoir computing9,41,42,43. The time delay also realizes the coupling of the recurrent nodes with their past states. In the discrete-time setting of our experiment with the sampling period Δt, Eq. (2) can be rewritten as
where T = t/Δt is an integer number representing the time step. W denotes Gaussian random numbers with zero mean and unit variance. The evolution of θ then follows as
with the discrete time delay δT = δt/Δt. Referring to the concept of virtual nodes and time-multiplexing12, we consider the transient state θ(T) of a physical node at different time steps as virtual nodes constituting the reservoir. Each virtual node state θ(T) is, according to Eq. (4), coupled to its previous states θ(T − 1) and θ(T − δT − 1). The virtual nodes thus reflect a topology with sparse interconnections realized by the delay of the physical node (Fig. 3a). The interconnections are inherently nonlinear due to the sine function in Eq. (4), which originates from the physical interaction between the active and the immobile particle and naturally serves as the activation function in our RC.
The working principle of our RC is now illustrated in Fig. 3. For simplicity of the discussion, we consider a RC with a single physical node, scalar input Xn and output Yn at the n-th computation step (for the general case see Supplementary Note 2). The input layer un is generated via a matrix of input weights \({{\mathbb{W}}}_{{{{{{{{\rm{in}}}}}}}}}\in {{\mathbb{R}}}^{{P}_{{{{{{{{\rm{in}}}}}}}}}\times 2}\),
with the scalar bin as the input bias. un is a one-dimensional array containing Pin elements, which are sequentially input into the physical node as the angle u(T) of the active particle in Eq. (3). This operation is equivalent to a time-multiplexing of the input Xn into Pin virtual nodes with \({{\mathbb{W}}}_{{{{{{{{\rm{in}}}}}}}}}\) as the mask, as commonly applied in continuous-time single node physical RC approaches12,44. The output Yn is derived as a linear combination of a scalar bias bout, the input signal Xn, and the node states θ of the past Pout time steps using an output weight matrix \({{\mathbb{W}}}_{{{{{{{{\rm{out}}}}}}}}}\in {{\mathbb{R}}}^{1\times (2+{P}_{{{{{{{{\rm{out}}}}}}}}})}\),
where T = 0 denotes the time of the current computation step. The output weight matrix \({{\mathbb{W}}}_{{{{{{{{\rm{out}}}}}}}}}\) is the only quantity to be trained in the RC framework to tune the output towards the target signal via a ridge regressions14 for each computation cycle.
As compared to conventional time-multiplexed physical RCs where Pin = Pout, we explicitly allow these two parameters to be independent and freely adjustable (see Fig. 3b). In particular, we set Pout ≫ Pin for our RC. By this means, each output is derived not only from the reservoir states of its corresponding step, but previous nhist = Pout/Pin steps. It will be demonstrated in the next sections that this setting enables us to carry out the RC with a good stability of the output and, most importantly, an effective reduction of the impact of the intrinsic noise.
This single physical node architecture can be further extended to multiple physical nodes operated in parallel. In our experiment, we control Nnode = 10 independent physical nodes in one sample simultaneously and set Pin = 2. The virtual nodes of the 10 physical nodes together constitute the reservoir with a size of NnodePin = 20. Figure 1d and Supplementary Movie 1 show the real time image and video of the sample in the experiment. Exemplary traces of θ(t) of four physical nodes (blue lines) driven by an external signal X (red line) in the experiment are plotted in Fig. 3c.
The configuration of the RC is optimized in the simulation for the best performance and then applied in the experiment. The input weights are selected from a binary distribution \({{\mathbb{W}}}_{{{{{{{{\rm{in}}}}}}}}}^{i}\in \{-2,2\}\), which brings a better noise resistance of the RC than the multi-value weights according to simulation results. The details of the RC configuration are described in Supplementary Note 2.
Chaotic series prediction
We test our RC with the free-running prediction of the chaotic Mackey–Glass series (MGS). The MGS is generated by the delay differential equation
which was introduced to model the complex dynamics of physiological feedback systems45. It has been widely used as a benchmark task for series forecasting46,47,48,49. With parameters α = 0.2, β = 10, γ = 0.1, and a delay parameter τ = 17, the MGS exhibits a chaotic behavior with the Lyapunov exponent of around 0.00646. The performance of the prediction is evaluated by the normalized root-mean-square error (NRMSE, see Supplementary Note 3 for details). Figure 4 shows the results of the MGS predictions by our RC in experiments and simulations.
Simulated prediction
The deterministic simulations have been carried out with a very small reservoir with only 20 virtual nodes but using a large Pout = 400, i.e., nhist = 200 historical reservoir states for each output. The simulation result shows a very good prediction of the target MGS up to around 900 steps (corresponding to 5.4 Lyapunov time) with a NRMSE of 6.7 × 10−2 (Fig. 4a). Similarly, Supplementary Note 4 also describes the prediction of a three-dimensional chaotic Lorenz series by our RC in a deterministic simulation. These results signify the capability of our architecture for chaotic systems predictions.
Experimental prediction
As compared to the simulations, the experimental RC performance is significantly degraded as a result of the Brownian motion of the active particles and the sensitivity of chaotic systems. The noise induced by Brownian motion acts on the particle angular position ϕ(T) (Eq. (3)) and propagates to the virtual node states θ(T) (Eq. (4)). The signal-to-noise ratio (SNR) of the RC can be estimated by comparing the ϕ(T) in simulations with and without noise (see Supplementary Note 5) and reveals an extremely low value (SNR = 1.9 (2.78 dB)). Figure 4b depicts the experimental results from 50 repetitions of the MGS prediction. The outputs Y(n) (gray curves) from different experimental predictions exhibit a notable influence of the noise. The mean of the outputs \(\overline{Y}(n)\) (blue curve) can reproduce the fundamental period of the target MGS for around 200 steps, and the details of the series for about 50 steps, which is, considering the extremely low SNR, still remarkable. These results obtained from the experimental RC are further underscored by the very good prediction of a non-chaotic periodic trigonometric series (see Supplementary Note 4).
The outstanding RC performance under these very noisy conditions becomes possible by introducing the special architecture using a number of historical reservoir states nhist for the output derivation. Figure 4c–g demonstrates the experimental results of the RC with Pout varying from 200 to 400, corresponding to nhist from 100 to 200. The accuracy of the predictions is improved by increasing nhist (Fig. 4c). Changing nhist from 100 to 200 results in a decrease of the difference between the mean of the outputs \(\overline{Y}\) and the target MGS by 22% (evaluated by the root-mean-square (RMS) of 200 steps \(| \overline{Y}-{Y}^{{{{{{{{\rm{target}}}}}}}}}|\) in Fig. 4d). The standard deviation σY of 50 predictions decreases by 32% (200 steps RMS of σY in Fig. 4e), and the resulting NRMSE of 200 step predictions diminishes by 28%. This trend of lower prediction error for the experimental system is also picked up by stochastic simulations with an NRMSE reduction of 11% (Fig. 4f).
This finding is striking as increasing nhist does not add more information to the RC, nor increases the reservoir dimensionality, which is determined by the number of linearly independent variables43,50 of the reservoir. During the computation of the RC, the node states of each step are nonlinearly transformed and mapped into the states of the following steps43, while they attenuate in magnitude due to the memory fading property of the node. Conventional RCs derive the output from the reservoir state of its corresponding step, which implicitly contains the information of the historical states. Whereas in our RC, the historical reservoir states can directly contribute to the output.
The actual contributions of the historical states are determined via the training of the output weights \({{\mathbb{W}}}_{{{{{{{{\rm{out}}}}}}}}}\). The states correlating more to the current output obtain higher weight magnitudes. Figure 4g plots the \({{\mathbb{W}}}_{{{{{{{{\rm{out}}}}}}}}}\) trained in experiments with nhist of 100 and 200. The first 2000 elements of \({{\mathbb{W}}}_{{{{{{{{\rm{out}}}}}}}}}\) corresponding to the near past reservoir states (n from 0 to −99) have similar structures in both cases. As compared to the near past, the far past states (n from −100 to − 199 in the nhist = 200 case) correspond to smaller weights in amplitude, indicating the weaker correlations and contributions to the current output. The highest magnitude of \({{\mathbb{W}}}_{{{{{{{{\rm{out}}}}}}}}}\) appears at around n = − 17, which coincides with the delay parameter τ = 17 of the target MGS (Eq. (7)). The trained \({{\mathbb{W}}}_{{{{{{{{\rm{out}}}}}}}}}\) thereby partly reveals the property of the target signal.
Impact of noise
These results suggest that there might be an optimal relation between the used number of historical states and the error of the RC to stabilize the prediction and to reduce the impact of the inherent noise due to Brownian motion. To investigate this interrelation we refer to deterministic and stochastic simulations. Figure 5a, b display the RC performance as measured by the NRMSE of the predictions as functions of Pin and Pout. Without noise (Fig. 5a), the results indicate a larger error with low Pin or Pout and poor performances for Pin < 2 and Pout < 50. The performance is improving for increasing Pin and Pout simultaneously. The best results for the deterministic system are obtained with nhist = Pout/Pin = 60.
For the noisy active particle nodes (Fig. 5b), the RC outputs are unstable for Pout < 100. The free-running prediction has a high probability to yield fast diverging outputs and a large NRMSE (bottom subplot). With Pout ≥ 100, stable outputs can be obtained. A trend towards lower NRMSE with higher nhist can be observed in agreement with our experimental results (Fig. 4f). Thus, both experiments and simulations confirm that using historical states for the prediction of our RC improves the stability and the quality of the prediction even under extremely noisy conditions.
Impact of the dynamical properties of the nodes
Our active particle recurrent node also provides the opportunity to tune the dynamical response of the physical node across the transition of the pitchfork bifurcation with either particle speed v0 or delay δt. The tuning varies how fast the memory of each node fades, as indicated in Fig. 2d by the relaxation time τθ and also the coupling of the virtual nodes, as given by Eq. (4).
The impact of the node dynamics on the performance of the RC is depicted in Fig. 5c–e with the NRMSE of the MGS predictions as function of v0/R0 and δt resulting from deterministic simulations. For stochastic simulation results see Supplementary Note 6.
As discussed above, θ0 determines the asymptotic behavior of θ after an impulsive perturbation. The correlation between NRMSE and θ0 with the optimum at around θ0 ≈ 0.2 in Fig. 5e indicates that the dynamical properties of the physical node as characterized by the relaxation time τθ (Fig. 2d), are indeed a key factor of the RC performance for this task. The optimal performance of the RC is found below the transition point (yellow dashed line) where the approximate effective potential Ueff(θ) is largely determined by its parabolic term (Fig. 5c, right)27. At the transition point, where the relaxation time τθ is largest and the memory fades slowest, the RC does not exhibit the worst performance as expected9. The maximal NRMSE appears beyond the transition point, where the Ueff is in a double well form. The double well Ueff may cause inconsistent responses of the RC, i.e., similar inputs into the node may result in θ relaxations towards different potential minima. This inconsistency is presumed to induce the degradation of the RC performance37,51,52. If the barrier of the double-well potential is further raised by increasing θ0, the NRMSE appears to decrease again (Fig. 5e) presumably due to the fact that the θ is largely residing in one of the potential wells with a short relaxation time. This interpretation is supported by the comparison of the RC performances with different \(| {{\mathbb{W}}}_{{{{{{{{\rm{in}}}}}}}}}^{i}|\) magnitudes in Fig. 5e. Larger \(| {{\mathbb{W}}}_{{{{{{{{\rm{in}}}}}}}}}^{i}|\) induces stronger inputs as perturbations on θ, thereby a higher probability of θ to jump over the barrier to another potential well. Hence the maximum of NRMSE with larger \(| {{\mathbb{W}}}_{{{{{{{{\rm{in}}}}}}}}}^{i}|\) appears at higher θ0, where the barrier of the potential is also higher.
Discussion
We have demonstrated above the realization of a physical reservoir computer in an experiment using self-propelled active microparticles. A retarded propulsion towards an immobilized target particle creates a self-organized non-linear dynamical system that is, despite the strong Brownian noise, capable of predicting chaotic time series such as Mackey–Glass and Lorenz series when used as a physical node in a reservoir computer. The key element of this physical recurrent node is a time delay realizing a retarded interaction27,53,54 that creates the fading memory as a basic requirement for reservoir computing8. It also provides the coupling of the active particle dynamics to its past allowing to implement virtual nodes living on a single physical active particle system via time-multiplexing12,44. The information processing that is provided by such a single node may therefore extend the rare simulation work on reservoir computing with active particle swarms with interparticle coupling37. Additionally, the nonlinearity that is required for computations is an intrinsic physical property of our active particle system and requires no extra treatment of the output signal 55. Future work could introduce direct physical coupling between the isolated physical nodes in our configuration, e.g., through hydrodynamic or other interactions, to obtain more complex dynamical networks of interacting synthetic active particles.
The dynamics of the physical recurrent node that is the basic unit of our reservoir computer is controlled by a parameter containing the product of activity (active particle speed) and time delay. It can be understood with a simple Landau-like self-induced quartic effective potential, which exhibits a transition from a single-well to double-well form27. The system thereby allows to address the relation of RC performance and node relaxation time, where we found a clear indication for an optimal performance below the transition point. While the memory fading becomes extremely slow at the transition point and one might expect the worst performance of the system accordingly, it is observed that a double-well potential with a small barrier leading to inconsistencies in the relaxation dynamics is more severe. Interestingly, such double-well potentials have recently been discussed as non-linear stochastic-resonance-based activation functions in an attempt to provide better stability of Echo-State-Networks against noise56.
Noise is an inherent property of our information processing units, as Brownian motion causes strong fluctuations of the node state. Such noises are inevitable at the smallest scales also in the context of biological information processing57, for instance in neurons58, with both positive and negative effects59,60. In physical RC approaches, noise is commonly a major limiting factor for the performances43,47,61 although subtle noises are reported to be beneficial as well8,19,46,62,63. Yet, general strategies for the noise suppression are unclear except increasing the reservoir size64, which is normally costly for physical RCs. While the performance of our reservoir computer for the chaotic system prediction is highly degraded by the noise due to the sensitivity of chaotic systems (for non-chaotic series predictions see Supplementary Note 4), we have introduced the architecture that utilizes historical reservoir states for output derivation providing remarkable stability and noise reduction even under low signal-to-noise ratios. This architecture is not increasing the dimensionality of the reservoir nor changing the dynamics of the nodes, and could be potentially useful in future reservoir computing studies.
In summary, simple retarded interactions in synthetic active microparticle systems can give rise to nonlinear self-driven dynamics that form a basis for information processing with active matter. Our reservoir computer highlights this connection between information processing, machine learning, and active matter on the microscale, and also paves the way for new studies on noise in reservoir computing. While we so far referred to isolated active recurrent units, we envision that the high level control of the synthetic active matter will yield new emergent physical collective states, which may leverage the field of active synthetic dynamical systems for information processing.
Methods
Sample preparation
The sample used in experiments contains two kinds of microparticles: polystyrene (PS) particles (microParticles GmbH) of 3 μm diameter and the melamine formaldehyde (MF) (microParticles GmbH) particles of 2.19 μm diameter, suspended in a water solution (Fig. 1b). Gold (Au) nano-particles of around 8 nm diameter are uniformly distributed on the surface of the MF particle, cover about 10% of the total surface area of the latter. Two glass coverslips (20 × 20 mm2 and 24 × 24 mm2) confine a 3 μm thick sample layer in between. Due to the surface tension of water, the PS particles are compressed and immobilized on the coverslips serving as spacers to define the sample thickness. The PS and MF–Au particles are separately added into two 2% Pluronic F-127 solutions. After 30 min, the Pluronic concentration of both solutions is decreased to 0.02% by diluting twice. Each dilution is followed by a centrifugation and then a removal of a part of the solution to keep the particle concentration. A 0.3 μL sample of PS particles is pipetted on one of the coverslips, then a 0.3 μL sample of MF–Au particles is pipetted in the droplet of PS particles. The sample is then covered carefully with a second coverslip. The edges of the sample are sealed by polydimethylsiloxane (PDMS) to prevent leakage and evaporation, also to relieve the liquid flow inside the sample. The experiment starts about 1 h after the sample preparation to wait until residual liquid flows in the sample ceased.
Experimental setup
The experimental setup is illustrated in Fig. 1a. The MF–Au microparticles are heated by a focused, continuous-wave laser with a wavelength of 532 nm. The light from the laser module (CNI, MGL-H-532-1W) is expanded by two tube lenses (35 mm, 150 mm focal lengths) in the beam size, then guided by mirrors to a high-speed reflective Spacial Light Modulator (SLM, Meadowlark Optics, HSP512-532), which modulates the phase of the reflected laser. The reflected laser is then guided through two lenses (500 mm, 300 mm focal lengths) to an inverted microscope (Olympus, IX73). A small opaque dot on a glass window located at the focal point of the 500 mm lens serves as a mask to block the unmodulated laser reflected by the SLM top surface. The laser in the microscope is reflected by a dichroic beam splitter (Omega Optical, 560DRLP), then focused by an objective lens (100x, Olympus, UPlanFL N x 100/1.30, Oil, Iris, NA. 0.6–1.3) on the sample plane with a beam width at half maximum about 0.6 μm.
The sample is illuminated by white light from an LED lamp (Thorlabs, SOLIS-3C) through an oil-immersion dark-field condenser (Olympus, U-DCW, NA 1.2–1.4). The image of the sample is projected by the objective lens and a tube lens (180 mm focal length) inside the microscopy stand as well as two additional lenses (100 mm, 150 mm focal lengths) outside the microscopy stand to a camera (Hamamatsu digital sCMOS, C11440-22CU). The numerical aperture (NA) of the objective is set to a value below the minimal NA of the dark-field condenser. Two filters (EKSMA Optics 246-2506-532, Thorlabs FESH0800) in front of the camera block the back reflections of the laser from the microscope. A desktop PC (Intel(R) Core™ i7-7700K CPU @4 × 4.20 GHz, NVIDIA GeForce GTX 1050Ti) with a LabVIEW program (v. 2019) analyzes the images, records data, and manipulates the active particles by controlling the laser through the phase pattern on the SLM. More details are given in Supplementary Note 7.
Data availability
All data in support of this work is available in the manuscript or the supplementary materials. Further data and materials are available from the corresponding author upon request.
References
Tkačik, G. & Bialek, W. Information processing in living systems. Annu. Rev. Condens. 7, 1–29 (2014).
Wadhams, G. H. & Armitage, J. P. Making sense of it all: bacterial chemotaxis. Nat. Rev. Mol. Cell Biol. 5, 1024–1037 (2004).
Celani, A. & Vergassola, M. Bacterial strategies for chemotaxis response. Proc. Natl Acad. Sci. USA 107, 1391–1396 (2010).
Cosentino, C. & Bates, D. Feedback Control in Systems Biology (CRC Press, 2019).
Knudsen, E. I., Lac, S. & Esterly, S. D. Computational maps in the brain. Annu. Rev. Neurosci. 10, 41–65 (1987).
Markoviç, D., Mizrahi, A., Querlioz, D. & Grollier, J. Physics for neuromorphic computing. Nat. Rev. Phys. 2, 499–510 (2020).
Abiodun, O. I. et al. State-of-the-art in artificial neural network applications: a survey. Heliyon 4, e00938 (2018).
Jaeger, H. The “Echo State" Approach to Analysing and Training Recurrent Neural Networks-With an Erratum Note’. https://www.ai.rug.nl/minds/uploads/EchoStatesTechRep.pdf (2001).
Maass, W., Natschläger, T. & Markram, H. Real-time computing without stable states: a new framework for neural computation based on perturbations. Neur. Comput. 14, 2531–2560 (2002).
Verstraeten, D., Schrauwen, B., D’Haene, M. & Stroobandt, D. An experimental unification of reservoir computing methods. Neur. Netw. 20, 391–403 (2007).
Paquot, Y., Dambre, J., Schrauwen, B., Haelterman, M. & Massar, S. Reservoir computing: a photonic neural network for information processing. Nonlin. Opt. Appl. 12, IV77280B–77280B (2010).
Appeltant, L. et al. Information processing using a single dynamical node as complex system. Nat. Commun. 2, 468 (2011).
Lukoševičius, M. & Jaeger, H. Reservoir computing approaches to recurrent neural network training. Comput Sci. Rev. 3, 127–149 (2009).
Lukosevicius, M. In Neural Networks: Tricks of the Trade 3rd edn, Vol. 7700 (eds. Montavon, G., Orr, G. B. & Muller, K. R.) Ch. 659–686 (Springer Heidelberg, 2012).
Tanaka, G. et al. Recent advances in physical reservoir computing: a review. Neur. Netw. 115, 100–123 (2019).
Nakajima, K. Physical reservoir computing-an introductory perspective. Jpn J. Appl. Phys. 59, 060501 (2020).
Rafayelyan, M., Dong, J., Tan, Y., Krzakala, F. & Gigan, S. Large-scale optical reservoir computing for spatiotemporal chaotic systems prediction. Phys. Rev. X 10, 041037 (2020).
Larger, L. et al. Photonic information processing beyond turing: an optoelectronic implementation of reservoir computing. Opt. Expr. 20, 3241–3249 (2012).
Paquot, Y. et al. Optoelectronic reservoir computing. Sci. Rep. 2, 287 (2012).
Van der Sande, G., Brunner, D. & Soriano, M. C. Advances in photonic reservoir computing. Nano Photon. 6, 561–576 (2017).
Coulombe, J. C., York, M. C. A. & Sylvestre, J. Computing with networks of nonlinear mechanical oscillators. PLoS One 12, e0178663 (2017).
Dion, G., Mejaouri, S. & Sylvestre, J. Reservoir computing with a single delay-coupled non-linear mechanical oscillator. J. Appl. Phys. 124, 152132 (2018).
Wu, S., Zhou, W., Wen, K., Li, C. & Gong, Q. Improved reservoir computing by carbon nanotube network with polyoxometalate decoration. In 2021 IEEE 16th International Conference on Nano/Micro Engineered and Molecular Systems (NEMS) 994–997 (IEEE, China, 2021).
Bechinger, C. et al. Active particles in complex and crowded environments. Rev. Mod. Phys. 88, 045006 (2016).
Lavergne, F. A., Wendehenne, H., Bäuerle, T. & Bechinger, C. Group formation and cohesion of active particles with visual perception-dependent motility. Science 364, 70–74 (2019).
Bäuerle, T., Löffler, R. C. & Bechinger, C. Formation of stable and responsive collective states in suspensions of active colloids. Nat. Comm. 11, 2547 (2020).
Wang, X., Chen, P.-C., Kroy, K., Holubec, V. & Cichos, F. Spontaneous vortex formation by microswimmers with retarded attractions. Nat. Commun. 14, 56 (2023).
Woodhouse, F. G. & Dunkel, J. Active matter logic for autonomous microfluidics. Nat. Comm. 8, 15169 (2017).
Colen, J. et al. Machine learning active-nematic hydrodynamics. Proc. Natl Acad. Sci. USA 118, e2016708118 (2021).
Cichos, F., Gustavsson, K., Mehlig, B. & Volpe, G. Machine learning for active matter. Nat. Mach. Intel. 2, 94–103 (2020).
Muiños-Landin, S., Fischer, A., Holubec, V. & Cichos, F. Reinforcement learning with artificial microswimmers. Sci. Robot. 6, eabd9285 (2021).
Liebchen, B. & Löwen, H. Optimal navigation strategies for active particles. EPL 127, 34003 (2019).
Schneider, E. & Stark, H. Optimal steering of a smart active particle. EPL 127, 64003 (2019).
Colabrese, S., Gustavsson, K., Celani, A. & Biferale, L. Flow navigation by smart microswimmers via reinforcement learning. Phys. Rev. Lett. 118, 158004 (2017).
Colabrese, S., Gustavsson, K., Celani, A. & Biferale, L. Smart inertial particles. Phys. Rev. Fluids 3, 084301 (2018).
Gustavsson, K., Biferale, L., Celani, A. & Colabrese, S. Finding efficient swimming strategies in a three-dimensional chaotic flow by reinforcement learning. Eur. Phys. J. E. 40, 110 (2017).
Lymburn, T., Algar, S. D., Small, M. & Jüngling, T. Reservoir computing with swarms. Chaos 31, 033121 (2021).
Fränzl, M., Muiños-Landin, S., Holubec, V. & Cichos, F. Fully steerable symmetric thermoplasmonic microswimmers. ACS Nano 15, 3434–3440 (2021).
Khadka, U., Holubec, V., Yang, H. & Cichos, F. Active particles bound by information flows. Nat. Commun. 9, 3864 (2018).
Goldenfeld, N. Lectures on phase transitions and the renormalization group. In Anomalous Dimensions (CRC Press, 2018).
Boyd, S. & Chua, L. Fading memory and the problem of approximating nonlinear operators with Volterra series. IEEE Trans. Circuit. Syst. 32, 1150–1161 (1985).
Maass, W., Natschläger, T. & Markram, H. Fading memory and kernel properties of generic cortical microcircuit models. J. Physiol. Paris 98, 315–330 (2004).
Dambre, J., Verstraeten, D., Schrauwen, B. & Massar, S. Information processing capacity of dynamical systems. Sci. Rep. 2, 514 (2012).
Appeltant, L., Van der Sande, G., Danckaert, J. & Fischer, I. Constructing optimized binary masks for reservoir computing with delay systems. Sci. Rep. 4, 3629 (2014).
Mackey, M. C. & Glass, L. Oscillation and chaos in physiological control systems. Science 197, 287–289 (1977).
Jaeger, H. & Haas, H. Harnessing nonlinearity: predicting chaotic systems and saving energy in wireless communication. Science 304, 78–80 (2004).
Antonik, P., Haelterman, M. & Massar, S. Brain-inspired photonic signal processor for generating periodic patterns and emulating chaotic systems. Phys. Rev. Appl. 7, 054014 (2017).
Fujii, K. & Nakajima, K. Harnessing disordered-ensemble quantum dynamics for machine learning. Phys. Rev. Appl. 8, 024030 (2017).
Dong, J., Rafayelyan, M., Krzakala, F. & Gigan, S. Optical reservoir computing using multiple light scattering for chaotic systems prediction. IEEE J. Sel. Top. Quant. Electr. 26, 1–12 (2020).
Carroll, T. L. Dimension of reservoir computers. Chaos 30, 013102 (2020).
Jüngling, T., Soriano, M. C., Oliver, N., Porte, X. & Fischer, I. Consistency properties of chaotic systems driven by time-delayed feedback. Phys. Rev. E 97, 042202 (2018).
Lymburn, T. et al. Consistency in echo-state networks. Chaos 29, 023118 (2019).
Mijalkov, M., McDaniel, A., Wehr, J. & Volpe, G. Engineering sensorial delay to control phototaxis and emergent collective behaviors. Phys. Rev. X 6, 1—16 (2016).
Chen, P.-C., Kroy, K., Cichos, F., Wang, X. & Holubec, V. Active particles with delayed attractions form quaking crystallites (a). EPL 142, 67003 (2023).
Cucchi, M., Abreu, S., Ciccone, G., Brunner, D. & Kleemann, H. Hands-on reservoir computing: a tutorial for practical implementation. Neuromorph. Comput. Eng. 2, 032002 (2022).
Liao, Z., Wang, Z., Yamahara, H. & Tabata, H. Echo state network activation function based on bistable stochastic resonance. Chaos Solit. Fractal. 153, 111503 (2021).
Tsimring, L. S. Noise in biology. Rep. Prog. Phys. 77, 026601 (2014).
Faisal, A. A., Selen, L. P. J. & Wolpert, D. M. Noise in the nervous system. Nat. Rev. Neurosci. 9, 292–303 (2008).
Stein, R. B., Gossen, E. R. & Jones, K. E. Neuronal variability: noise or part of the signal? Nat. Rev. Neurosci. 6, 389–397 (2005).
Guo, D., Perc, M., Liu, T. & Yao, D. Functional importance of noise in neuronal information processing. EPL 124, 50001 (2018).
Soriano, M. C. et al. Optoelectronic reservoir computing: tackling noise-induced performance degradation. Opt. Expr. 21, 12 (2013).
Sussillo, D. & Abbott, L. Generating coherent patterns of activity from chaotic neural networks. Neuron 63, 544–557 (2009).
Estebanez, I., Fischer, I. & Soriano, M. C. Constructive role of noise for high-quality replication of chaotic attractor dynamics using a hardware-based reservoir computer. Phys. Rev. Appl 12, 034058 (2019).
Alata, R., Pauwels, J., Haelterman, M. & Massar, S. Phase noise robustness of a coherent spatially parallel optical reservoir. IEEE J. Sel. Top. Quant. Electr. 26, 1–10 (2020).
Acknowledgements
This work is supported by Center for Scalable Data Analytics and Artificial Intelligence (ScaDS.AI) Dresden/Leipzig. The authors also acknowledge careful proof reading of the manuscript by Andrea Kramer.
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Authors and Affiliations
Contributions
X.W. and F.C. conceived the experiments. X.W. carried out the experiments and simulations. X.W. and F.C. discussed the results and wrote the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Peer review
Peer review information
Nature Communications thanks Giovanni Volpe and the other anonymous reviewer(s) for their contribution to the peer review of this work. A peer review file is available.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Wang, X., Cichos, F. Harnessing synthetic active particles for physical reservoir computing. Nat Commun 15, 774 (2024). https://doi.org/10.1038/s41467-024-44856-5
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41467-024-44856-5
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.