The ability to excite chaotic dynamics in nanomagnets1,2,3,4,5,6,7,8,9 would potentially have practical applications, such as encoding and random number generators10. So far, however, it has proven difficult to achieve; chaos is a non-periodic and sustainable motion of a physical body maintained by energy injected into a nonlinear system, whereas the magnetization dynamics induced by applying magnetic field and/or current are mainly in the form of switching or a periodic oscillation11,12,13,14,15,16.

Here, several techniques have recently been proposed and experimentally demonstrated with the goal of exciting and identifying chaos, including ones using a delayed-feedback circuit5,7,9 and ones that add another ferromagnet2,3,4,5,6,8. Another possible stage on which to demonstrate chaotic dynamics is an input-driven dynamical system17, where a series of time-dependent stimuli drive the nonlinear dynamics of a physical system. An example that utilizes such input-driven dynamics is human-voice recognition by a vortex spin-torque oscillator (STO)18, which is based on the algorithm of physical reservoir computing19,20,21. Wherein a human voice is converted into an electric-voltage input signal that modulates the dynamical amplitude of the vortex core. The input signal can be recognized when a one-to-one correspondence between it and the vortex dynamics is obtained through machine learning. Such a correspondence can be regarded as synchronization between the input signal and the STO response. A critical difference between this sort of synchronization and the conventional synchronizations studied in STOs previously, such as forced synchronization22,23, is that the input signal is generally non-periodic, whereas the input in the conventional cases usually has a certain periodicity. Moreover, while the STO dynamics in the conventional cases become identical to the input signal except a phase difference, the STO dynamics in this case are not generally identical to the input signal but have a certain reproducibility; the STO response is the same when the same input signal is injected. Thus, it is a wider concept of synchronization that encompasses the conventional ideas, which can be called generalized synchronization24, i.e., is a kind of input-driven dynamics first examined in the 1990s25,26,27. Another example of input-driven dynamics is chaos28,29, near which the computational capability of physical reservoir computing is sometimes enhanced30,31. As can be seen from these examples, input-driven dynamics is of great interest to researchers studying brain-inspired computing, because its complexity can be used for information processing32,33. However, there are only a few reports29,34,35,36 on input-driven dynamics, in particular chaos, in spintronics devices.

In this work, we study input-driven magnetization dynamics in a vortex STO by performing a numerical simulation of the Thiele equation. A uniformly distributed random magnetic field is used as a series of input signals; such input signals have often been used to study of input-driven dynamics37. We found that when the input is weak, input-driven synchronization occurs38; the output of the STO is determined by the input signal and is independent of the initial state. On the other hand, in the strong input limit, input-driven chaotic dynamics appear, in the form of an intermittency in the synchronized behavior, and therefore, the temporal dynamics depend on the initial state. We distinguish these ordered and chaotic dynamical phases by evaluating the largest conditional Lyapunov exponent, which is a measure quantifying the sensitivity of the dynamics to the initial state of the STO. We also study the relation between the dynamical phase and computational capability of the STO by relating the Lyapunov exponent to the short-term memory capacity.

Model

Thiele equation

The system we consider is schematically shown in Fig. 1a, where the vortex STO consists of a ferromagnetic/nonmagnetic/ferromagnetic trilayer. The free layer has a magnetic vortex, while the reference layer is uniformly magnetized. The z axis is normal to the film-plane, while the x axis is parallel to the projection of the magnetization in the reference layer onto the film-plane. Throughout this paper, the unit vector pointing in the direction of the k-axis (\(k=x,y,z\)) is denoted as \({\textbf{e}}_{k}\). It has been shown that the experimental and numerical results on the magnetization dynamics in a vortex STO are well described by the Thiele equation,39,40,41,42,43,44,45,46

$$\begin{aligned} \begin{aligned}{}&-G {\textbf{e}}_{z} \times \dot{{\textbf{X}}} - |D| \left( 1 + \xi s^{2} \right) \dot{{\textbf{X}}} - \kappa \left( 1 + \zeta s^{2} \right) {\textbf{X}} + a_{J} J p_{z} {\textbf{e}}_{z} \times {\textbf{X}} + c a_{J} J R_{0} p_{x} {\textbf{e}}_{x} + c \mu ^{*} {\textbf{e}}_{z} \times {\textbf{H}} = {\varvec{0}}, \end{aligned} \end{aligned}$$
(1)

where \({\textbf{X}}=(X,Y,0)\) is the position vector of the vortex core, while \(G=2\pi pML/\gamma\) and \(D=-(2\pi \alpha ML/\gamma )[1-(1/2)\ln (R_{0}/R)]\) consist of the saturation magnetization M, the gyromagnetic ratio \(\gamma\), the Gilbert damping constant \(\alpha\), the thickness L, the disk radius R, and the core radius \(R_{0}\) of the free layer. The polarity p and chirality c are each assumed to be \(+1\) for convenience. The normalized distance of the vortex core’s position from the disk center is \(s=|{\textbf{X}}|/R=\sqrt{(X/R)^{2}+(Y/R)^{2}}\). The dimensionless parameter \(\xi\) quantifies the nonlinear damping in a highly excited state45. The parameters \(\kappa\) and \(\zeta\) relate to the magnetic potential energy W via

$$\begin{aligned} W = \frac{\kappa }{2} \left( 1 + \frac{\zeta }{2} s^{2} \right) |{\textbf{X}}|^{2}, \end{aligned}$$
(2)

and \(\kappa =(10/9)4\pi M^{2}L^{2}/R\)45. The dimensionless parameter \(\zeta\) provides the linear dependence of the oscillation frequency on the current45. As mentioned below, the oscillation amplitude of the vortex core is determined by the current magnitude, and thus, we call \(\zeta\) the amplitude-frequency coupling parameter, for convenience. The spin-transfer torque strength with spin polarization P is \(a_{J}=\pi \hslash P/(2e)\)43,44. The electric current density J is positive when the current flows from the reference to the free layer. The current is denoted as \(I=\pi R^{2}J\). The unit vector pointing in the magnetization direction in the reference layer is \({\textbf{p}}=(p_{x},0,p_{z})\). The external magnetic field is denoted as \({\textbf{H}}\), while \(\mu ^{*}=\pi MLR\). The material parameters are summarized in “Methods” section. Note that the current is an experimentally controllable variable; typically, it is on the order of 1 mA (\(I < 10\) mA47). Moreover, the value of \(\zeta\) relates to the material and is on the order of 0.1–1.046. Therefore, we will study the dynamical behavior of the vortex core as a function of J (or equivalently, I) and \(\zeta\); see also “Methods” section.

In the absence of a current and magnetic field, the vortex core is located near the disk center. By applying a direct current, spin-transfer torque moves the vortex core from the disk center and excites an auto-oscillation with an approximately constant amplitude when the current density is larger than a critical value48,

$$\begin{aligned} J_{\textrm{c}} = \frac{|D|\kappa }{Ga_{J}p_{z}}. \end{aligned}$$
(3)

The oscillation amplitude and frequency are \(s_{0}=\sqrt{[(J/J_{\textrm{c}})-1]/(\xi +\zeta )}\) and \(f_{0}=(\kappa /G)(1+\zeta s_{0}^{2})/(2\pi )\), respectively. The condition that the vortex core is inside the ferromagnetic disk, i.e., \(|{\textbf{X}}|<R\) or equivalently, \(s < 1\), implies that the Thiele equation is applicable when the current density is less than another threshold value,

$$\begin{aligned} J_{\textrm{th}} = J_{\textrm{c}} \left( 1 + \xi + \zeta \right) . \end{aligned}$$
(4)

Figure 1b shows an example of such an auto-oscillation, where \(I=5.0\) mA and \(\zeta =7.52\). The figure indicates that the vortex core moves from the disk center (\(s=0\)) and shows an oscillation around the center with an approximately constant radius \(s_{0}\). The small-amplitude oscillation of s is due to the fifth term in Eq. (1), which originates from the in-plane component of \({\textbf{p}}\) and breaks the axial symmetry of the Thiele equation around the z axis.

The input-driven dynamics appear when a series of time-dependent signals is injected to this auto-oscillation state. The input signal in this study is a random-pulse magnetic field, as in an experiment on physical reservoir computing49,

$$\begin{aligned} {\textbf{H}}(t) = h_{x} r(t) {\textbf{e}}_{x}, \end{aligned}$$
(5)

where \(h_{x}\) is the maximum value of the input magnetic field, and r(t) is a uniformly distributed random number in the range of \(-1 \le r(t) \le 1\) whose corresponding input signal has a pulse width \(t_{\textrm{p}}\). Although the pulse input signal in the experiments has a rise time, it can be shortened by the preprocessing50. For simplicity, we assume a step-function pulse in this work.

Figure 1
figure 1

(a) Schematic illustration of vortex spin torque oscillator and dynamics of the vortex core induced by a direct current (DC) and random magnetic field. The DC induces an oscillation of the vortex core around the disk center whereas the random magnetic field modulates the oscillation amplitude. (b) Evolution of the normalized amplitude s in the presence of the current \(I=5\) mA. The inset shows the dynamics of the x component and y component of the magnetic vortex core normalized by the radius of the magnetic vortex core.

Results

Temporal dynamics

The input magnetic field given by Eq. (5) causes two kinds of dynamics; synchronization and chaotic. Figure 2a shows two solutions of Eq. (1) with slightly different initial conditions before injection of the input signal, where \(I=5.0\) mA and \(\zeta =7.52\). For convention, let us denote the two solutions as \({\textbf{X}}_{1}\) and \({\textbf{X}}_{2}\). The solutions oscillate with the same frequency but different phases. When a weak input signal is injected from \(t=200\) ns, they overlap in the limit \(t\rightarrow \infty\); i.e., synchronization occurs; see Fig. 2b, where \(h_{x}=4.0\) Oe and \(t_{\textrm{p}}=3.0\) ns. The difference between the two solutions, \(|{\textbf{X}}_{1}-{\textbf{X}}_{2}|/R\), is illustrated in Fig. 2c. As shown, the difference decreases over time, which also indicates that phase synchronization is achieved. Recall that the input signal does not have periodicity; therefore, the synchronization of the two solutions is different from the forced synchronization found in, for example, Refs.22,23. The synchronization here is input-driven51,52,53,54,55, where the temporal dynamics are determined by the input signal, as was found in a vortex STO recently38. It occurs when the strength of the input signal is relatively weak.

Chaotic dynamics also appear as the input-signal strength increases. Figure 2d shows an example with \(h_{x}=16.0\) Oe, where the two solutions of Eq. (1) were arrived at with slightly different initial conditions. As shown, the solutions are never completely overlapped even though a long time passes. The difference between the two solutions often becomes small; i.e., synchronized behavior appears. However, before the difference completely becomes zero, intermittency appears and the two solutions separate from each other; see also Fig. 2e. This result indicates that the dynamical behavior of the vortex core is not solely determined by the input signal, but is also sensitive to its initial state. The sensitivity of the dynamical behavior implies chaos. We emphasize that the chaotic behavior is caused by the input signal. In fact, without the input signal, the system described by the Thiele equation is a two-dimensional autonomous system, and in such a system, chaos is prohibited by Poincaré–Bendixson theorem56. Therefore, the behavior of the vortex core must be input-driven chaos.

The chaotic dynamics found here derives from the temporal dynamics. To identify it more systematically, we can evaluate the Lyapunov exponent.

Figure 2
figure 2

(a) Temporal dynamics of two X/Rs, denoted as \(X_{1}/R\) and \(X_{2}/R\), with slightly different initial condition. Here, the random magnetic field is not injected yet. (b) Temporal dynamics of \(X_{1}/R\) and \(X_{2}/R\) after injection of the input signal. The input signal is relatively weak (\(h_{x}=4\) Oe). (c) Difference between \({\textbf{X}}/R\)s, \(|{\textbf{X}}_{1}-{\textbf{X}}_{2}|/R\), for a small input strength. (d) Temporal dynamics of \(X_{1}/R\) and \(X_{2}/R\) after injection of the input signal. The input signal is relatively strong (\(h_{x}=16\) Oe). (e) Difference between \({\textbf{X}}/R\)s, \(|{\textbf{X}}_{1}-{\textbf{X}}_{2}|/R\), for a large input strength.

Lyapunov exponent

The Lyapunov exponent is the expansion rate of the difference between two solutions of an equation of motion with slightly different initial conditions. In the present case, we will define the difference between two solutions, \({\textbf{X}}_{1}\) and \({\textbf{X}}_{2}\), as \(\delta x = |{\textbf{X}}_{1}-{\textbf{X}}_{2}|/R\). When \({\textbf{X}}_{1}\simeq {\textbf{X}}_{2}\) at a certain time \(t_{0}\), the time development of \(\delta x\) can be described by a linear equation, \(\delta {\dot{x}}\simeq \lambda \delta x\). The solution is \(\delta x(t)\simeq \delta x(t_{0}) e^{\lambda t}\), and \(\lambda\) is the Lyapunov exponent. Accordingly, the two solutions become identical when the Lyapunov exponent is negative. The dynamics corresponding to a negative Lyapunov exponent are, for example, saturation to a fixed point and phase synchronization. On the other hand, when the Lyapunov exponent is positive, the difference between the two solutions increases, indicating sensitivity of the dynamics to the initial state. A positive exponent implies the appearance of chaos. Moreover, when the exponent is zero, the initial difference is kept constant. An example of such dynamics is a limit-cycle oscillation. The method of evaluating the Lyapunov exponent is summarized in “Methods” section.

Figure 3 summarizes the dependence of the Lyapunov exponent on the input magnetic field strength \(h_{x}\) and amplitude-frequency coupling parameter \(\zeta\) for various current values I; (a) 2.5, (b) 4.0, and (c) 5.0 mA. The pulse width of the input signal is \(t_{\textrm{p}}=3.0\) ns. The figure provides a broad perspective on the relation between the parameter values and the dynamical phases, although the detailed structure depends on the current value (see also Supplementary Information, which has the other data supporting the discussion below, obtained for various currents and pulse widths).

Figure 3
figure 3

Dependence of Lyapunov exponent (\(\lambda\)) on the strength of the input signal (\(h_{x}\)) and amplitude–frequency coupling parameter (\(\zeta\)) for currents of (a) \(I=2.5\) mA, (b) \(I=4.0\) mA, and (c) \(I=5.0\) mA. The pulse width is 3.0 ns. The white dots represent the zero-exponent regions.

First, let us focus on the parameter regions where the values of \(h_{x}\) and \(\zeta\) are relatively small. The Lyapunov exponent takes on large negative values when \(h_{x}\) or \(\zeta\) increases. This tendency is clear in, for example, Fig. 3b, from \(h_{x}\simeq 0\) to \(h_{x}\simeq 4\) Oe. On the other hand, it seems somewhat unclear for \(I=2.5\) mA in Fig. 3a; but a closer look to small \(h_{x}\) and \(\zeta\) regions shows the negative enhancement of the exponent; see also Supplementary Information. This result implies that the input-induced synchronization is rapidly realized because the negative Lyapunov exponent corresponds to the inverse of the time scale decreasing the difference in the initial state; i.e., a large negative Lyapunov exponent means rapid synchronization. When the input strength \(h_{x}\) is large, the input signal substantially modifies the vortex-core position and contributes to a fast alignment of \({\textbf{X}}_{1}\) and \({\textbf{X}}_{2}\). Similarly, when \(\zeta\) is large, the oscillation frequency of the vortex core strongly depends on its oscillation amplitude s, and therefore, the vortex cores having different amplitudes oscillate with different frequencies and decrease their phase difference. Accordingly, synchronization easily occurs as \(h_{x}\) and \(\zeta\) increase, resulting in a large negative Lyapunov exponent.

Increasing the values of these parameters any further, however, makes it more difficult to fine tune the vortex-core position, because even a single input signal causes a large change in position. In such case, it is hard to align the vortex core positions by injection of an input signal; rather, a small difference between the initial states leads to a big difference in position. As a result, chaotic dynamics appears, which is confirmed by the positive Lyapunov exponent. We should also notice that the dependence of the Lyapunov exponent on the parameters is sometimes non-monotonic. For example, for \(I=5.0\) mA in Fig. 3c, at \(h_{x}\gtrsim 5.0\) Oe, the Lyapunov exponent becomes negative or positive depending on the value of \(\zeta\) around \(\zeta =3.0\). Such a parameter dependence has been frequently observed in nonlinear dynamical systems showing chaos10; see also Supplemental Information showing the dependence of the Lyapunov exponent on \(\zeta\) and the bifurcation diagram.

In summary, increasing the values of the system parameters leads to an enhancement in the efficiency of achieving input-driven synchronization, which can be quantified by an increasingly large negative Lyapunov exponent. Increasing the parameters further, however, results in a transition of the dynamical phase from synchronized to chaotic, because a large modulation of the vortex-core position expands the difference between the initial states. Chaotic dynamics appears when the Lyapunov exponent is positive.

At the end of this section, let us give a brief comment on the verification of chaos experimentally. In the experiments, the initial state of the vortex core is uncontrollable. Therefore, it is difficult to evaluate the Lyapunov exponent experimentally by a method similar to that used here, although there are statistical methods estimating the Lyapunov exponent from the experimentally observed data57,58,59. In addition, the stochastic torque due to thermal fluctuation60 makes the vortex-core dynamics complex, which also makes it difficult to evaluate the Lyapunov exponent. Note that there are other methods identifying chaos from the experimentally observed data. For example, the noise limit61,62 can be a figure of merit for the identification of chaos; in fact, it has been used in the recent experiments using the vortex STOs2,9. The noise limit has been, however, mainly evaluated in autonomous systems, while the present system is nonautonomous. An extension making it applicable to the nonautonomous system might be necessary, which could be a future work.

Discussion

We have found that chaotic dynamics in a vortex STO can be driven by injection of a random magnetic-field pulse. Moreover, Ref.29 found that a pulse-current driven bifurcation from ordered to chaotic dynamics occurs in a macrospin STO. While both results were obtained through numerical simulations, our findings, which are based on the use of a magnetic field, might be more suitable for an experimental verification. This is because excitation of chaos often requires a large input signal, and a large current can be used for generating a magnetic field, whereas there is a restriction on the magnitude of the current that can be directly applied to an STO because of its high resistance; see also “Methods” section.

As mentioned in the Introduction, an interesting application of input-driven dynamics is information processing. Let us, therefore, evaluate the computational capability of the STO and investigate the relation between this capability and the dynamical phase. As a figure of merit, we will use the short-term memory capacity50 of physical reservoir computing, which corresponds to, roughly speaking, the number of random input signals a physical reservoir can recognize; see also “Methods” section describing the evaluation of the memory capacity. A large memory capacity corresponds to high computational capability. Figure 4 plots the value of the short-term memory capacity for various currents I for the pulse width of 3.0 ns. In particular, Fig. 4a–c show the dependence of the short-term memory capacity, as well as that of the Lyapunov exponent, on the input strength \(h_{x}\) for a fixed \(\zeta\), while Fig. 4d–f show the dependence of the short-term memory capacity on \(h_{x}\) and \(\zeta\); see also Supplemental Information for the short-term memory capacity obtained for different pulse widths. The short-term memory capacity is large when the current magnitude is large. It is attributed to the current dependence of the relaxation time. The short-term memory capacity becomes large when the reservoir recognizes the input data well. The STO recognizes the input data by the change of the oscillation amplitude. When the current is small, the relaxation time, or the response of the vortex core, is slow48, and therefore, the input data is not well recognized. Therefore, the input data is not well recognized by the STO, leading to a small short-term memory capacity. On the other hand, when the current is large, the vortex core responds to the input data immediately. Therefore, the STO recognizes the injection of the input data leading to a large short-term memory capacity. We also note that the magnitude relationship between the relaxation time and the pulse width affects the computational capability, which was studied in Ref.63.

Figure 4
figure 4

Dependence of short-term memory capacity (\(C_{\textrm{STM}}\)) and the Lyapunov exponent \(\lambda\) on the input strength \(h_{x}\) for currents I of (a) 2.5, (b) 4.0, and (c) 5.0 mA; here \(\zeta =7.52\). Dependence of \(C_{\textrm{STM}}\) and \(h_{x}\) on \(h_{x}\) and \(\zeta\) for currents of (d) 2.5, (e) 4.0, and (f) 5.0 mA.

Comparing the results shown in Fig. 4 with those in Fig. 3 leads us to the following conclusions. The short-term memory capacity is large in the parameter regions where the Lyapunov exponent is negative. On the other hand, it becomes almost zero when the exponent is positive. These results can be interpreted as follows. A recognition task of an input signal is done by learning the response from the STO. Imagine that we inject identical input signals into the STO at different times. In general, the initial states of the STO for these different trials are different. Therefore, for STO to be able to recognize that input signals are the same, the output signal from it should saturate to a certain state that is independent of the initial ones. In other words, the temporal response of the STO should be solely determined by the input signal. Note that a negative Lyapunov exponent corresponds to the dynamics saturating to a fixed point and thus becoming independent of the initial state. Therefore, the dynamical phase corresponding to a negative Lyapunov exponent is applicable to a recognition task. On the other hand, when the dynamics are sensitive to the initial state, the STO will output different signals for the injection of a common input signal. In such a case, the success rate of a recognition task will be low. Chaos is a typical dynamics having sensitivity to the initial state. Therefore, the short-term memory capacity becomes zero when the Lyapunov exponent is positive. At the same time, we should note that, although chaotic dynamics are not useful for computing, its dynamical neighborhood can be used for computing. In some cases33, the boundary of the dynamical phase is useful for particular tasks. For example, the short-term memory capacity is maximized near the edge of chaos in some physical reservoirs30,31, although this is not the present case; see also Supplementary Information, where the Lyapunov exponent and the short-term memory capacity for various parameters are summarized. Therefore, performing computation even in a chaotic region would help to deepen our understanding of the computational capability of STOs.

In summary, we showed that input-driven chaotic magnetization dynamics can be excited in a vortex STO in a numerical simulation of the Thiele equation. Chaos was identified from the temporal dynamics of the solutions of the equation with slightly different conditions, as well as the analysis of the Lyapunov exponent. The correspondence between the dynamical phase of the input-driven STO and the short-term memory capacity of physical reservoir computing was also investigated. The results showed that the performance of brain-inspired computing relates to the dynamical phase of the input-driven dynamical systems, and therefore, manipulation of the phase is necessary for practical applications. The present work indicates that the dynamical phase of STO can be tuned between input-driven synchronized and chaotic states by system parameters. While the present work chose physical reservoir computing as an example, chaotic dynamics might have different applications in brain-inspired computing because chaotic dynamics has been found in artificial neural networks emulating brain activities64. The investigation of such applications will be a future work.

Methods

Material parameters

The material parameters were taken from typical experiments and simulations45,46,47 as \(M=1300\) emu/cm\({}^{3}\), \(\gamma =1.764\times 10^{7}\) rad/(Oe s), \(\alpha =0.01\), \(L=5\) nm, \(R=187.5\) nm, \(R_{0}=10\) nm, \(P=0.7\), \(\xi =2\), and \({\textbf{p}}=(\sin 60^{\circ },0,\cos 60^{\circ })\). A current I of 1 mA corresponds to a current density J of 0.9 MA/cm\({}^{2}\). The value of the current used in, for example, the experiment in Ref.47 is 8 mA at maximum. The value of \(\zeta\) in Ref.46 in the absence of a current is defined as \(\zeta =\kappa _{\textrm{ms}}^{\prime }/\kappa _{\textrm{ms}}\), where \(\kappa _{\textrm{ms}}\) and \(\kappa _{\textrm{ms}}^{\prime }\) relate to the magnetostatic energy and their values are determined by fitting experiments and/or numerical analysis of the Thiele equation. The values of these parameters in Ref.46 are widely distributed between the experimental (about \(\zeta \sim 5\)) and theoretical (about 0.2) analyses.

A typical magnitude of the input magnetic field used in the present work is on the order of 1–10 Oe. In experiments, a pulse magnetic field can be generated by applying a pulse electric current to a metal line. For example, assuming a thin metal line, a current of 1 mA flowing above an STO within a distance on the order of 10 nm generates a magnetic field on the order of 1–10 Oe, where we have used the Ampére law, \(H=I/(2\pi {\mathscr {R}})\) with current I and distance \({\mathscr {R}}\). In reality, the metal line has a finite width and the magnitude of the magnetic field will be small; for example, in Ref.65, a metal line is placed on a nanomagnet within a distance of 20 nm, and current of 20 mA generates a magnetic field of 80 Oe. Note that the magnitude of the magnetic field could be further enhanced by applying a large current, which is possible because the current is in a metal line. On the other hand, Ref.29 reported a bifurcation from order to chaos by injection of electric current pulses into a macrospin STO, where the current magnitude was on the order of 1–10 mA. However, it should be noted that a large current, such as 10 mA, leads to electrostatic breakdown of the STO, which includes an oxide barrier such as MgO.

Method of evaluating the Lyapunov exponent

The Shimada–Nagashima method66 was used to evaluate the Lyapunov exponent of STO6,7,29,36,67. Note that the Lyapunov exponent is the expansion rate of the difference between two solutions of an equation of motion with slightly different initial conditions. In this method, at a certain time \(t_{0}\), we add a small perturbation to a solution of the Thiele equation \({\textbf{X}}(t_{0})\) as \({\textbf{X}}^{(1)}(t_{0})={\textbf{X}}(t_{0})+\varepsilon R {\textbf{n}}_{0}\), where a unit vector \({\textbf{n}}_{0}\) in the xy plane points in an arbitrary direction and the meaning of the superscript “(1)” will be clarified below. The dimensionless parameter \(\varepsilon\) corresponds to the magnitude of the perturbation.

We solve the Thiele equations of \({\textbf{X}}(t_{0})\) and \({\textbf{X}}^{(1)}(t_{0})\) and obtain \({\textbf{X}}(t_{0}+\Delta t)\) and \({\textbf{X}}^{(1)}(t_{0}+\Delta t)\), where \(\Delta t\) is the time increment of the Thiele equation. Then, from the difference between \({\textbf{X}}(t_{0}+\Delta t)\) and \({\textbf{X}}^{(1)}(t_{0}+\Delta t)\), we evaluate the temporal Lyapunov exponent at \(t_{1}=t_{0}+\Delta t\) as \(\lambda ^{(1)}=(1/\Delta t)\ln [\varepsilon ^{(1)}/\varepsilon ]\), where \(\varepsilon ^{(1)}= |{\textbf{X}}^{(1)}(t_{1})-{\textbf{X}}(t_{1})|/R\).

Next, we introduce \({\textbf{X}}^{(2)}(t_{0}+\Delta t)\), which can be obtained by moving \({\textbf{X}}(t_{0}+\Delta t)\) in the direction of \({\textbf{X}}^{(1)}(t_{0}+\Delta t)\) over a distance \(\varepsilon\). Solving the Thiele equation of \({\textbf{X}}(t_{0}+\Delta t)\) and \({\textbf{X}}^{(2)}(t_{0}+\Delta t)\) yields \({\textbf{X}}(t_{0}+2 \Delta )\) and \({\textbf{X}}^{(2)}(t_{0}+2\Delta t)\). Then, the temporal Lyapunov exponent at \(t_{2}=t_{0}+2\Delta t\) is obtained as \(\lambda ^{(2)}=(1/\Delta t)\ln [\varepsilon ^{(2)}/\varepsilon ]\) with \(\varepsilon ^{(2)}=|{\textbf{X}}^{(2)}(t_{2})-{\textbf{X}}(t_{2})|/R\).

Now we repeat the above procedure: i.e., first, we introduce \({\textbf{X}}^{(n+1)}(t_{0}+n \Delta t)\) by moving \({\textbf{X}}(t_{0}+n \Delta t)\) in the direction of \({\textbf{X}}^{(n)}(t_{0}+\Delta t)\) over a fixed distance \(\varepsilon\). Second, by solving the Thiele equation, we obtain \({\textbf{X}}^{(n+1)}[t_{0}+(n+1)\Delta t]\) and \({\textbf{X}}[t_{0}+(n+1)\Delta t]\). Third, the temporal Lyapunov exponent at \(t_{n+1}=t_{0}+(n+1)\Delta t\) is obtained as

$$\begin{aligned} \lambda ^{(n+1)} = \frac{1}{\Delta t} \ln \frac{\varepsilon ^{(n+1)}}{\varepsilon }, \end{aligned}$$
(6)

where

$$\begin{aligned} \varepsilon ^{(n+1)} = \frac{|{\textbf{X}}^{(n+1)}(t_{n+1})-{\textbf{X}}(t_{n+1})|}{R}. \end{aligned}$$
(7)

The Lyapunov exponent is the average of the temporal Lyapunov exponents defined as

$$\begin{aligned} \lambda = \lim _{N \rightarrow \infty } \frac{1}{N \Delta t} \sum _{i=1}^{N} \ln \frac{\varepsilon ^{(i)}}{\varepsilon }. \end{aligned}$$
(8)

Strictly speaking, the Lyapunov exponent obtained by this procedure corresponds to the maximum (or largest) Lyapunov exponent, i.e., the largest expansion rate of the difference between two solutions with slightly different initial conditions. The method assumes that, even though an initial perturbation is given in an arbitrary direction (\({\textbf{n}}_{0}\) mentioned above), the perturbed solution will naturally move in the most expandable direction. In addition, the Lyapunov exponent evaluated here is a conditional one, where, although the system is non-autonomous because time-dependent input signals are injected, only the expansion rate of \({\textbf{X}}\) is evaluated because the input signal is constant during a pulse width longer than the time increment29.

Method of evaluating short-term memory capacity

The memory capacity is a figure of merit quantifying the computational capability of physical reservoir computing68. As can be seen below, roughly speaking, the memory capacity is the number of the previous input signal which can be reproduced by the output signal from physical system. Let us assume that the input signals are pulse-shaped random number \(q_{k}\), where the suffix \(k=1,2,\ldots\) corresponds to the order of the input signals. In the present work, a binary input \(b_{k}=0,1\) is used to evaluate the short-term memory capacity in accordance with Refs.48,49,50. Note that a uniformly distributed random input \(r_{k}\) (\(-1 \le r_{k} \le 1\)) has also been used in some reports29. An example of the target data is

$$\begin{aligned} y_{k,D}^{\textrm{STM}} = q_{k-D}, \end{aligned}$$
(9)

where an integer D, called a delay, characterizes the order of the past input data. There are other kinds of target data, which, in general, include nonlinear transformation of input data37. A memory capacity is defined for each kind of target data. The short-term memory capacity is for target data given by Eq. (9). The following procedure for evaluating of the short-term memory capacity is applicable to general target data. Therefore, for generality, let us denote the target data as \(y_{k,D}\), for a while.

The procedure is as follows. In physical reservoir computing, the physical system consists of many bodies, corresponding to neurons. The output signal from the ith neuron with respect to the kth input signal \(q_{k}\) is denoted as \(u_{k,i}\). Note that, since the present system consists of a single STO, we apply the time-multiplexing method33 to introduce virtual neurons. The output signal from a vortex STO is proportional to the y component of the vortex-core position through the magnetoresistance effect. In experiments18,47,49, the output signal is decomposed into the amplitude s and the phase through the Hilbert transformation. The amplitude is used as a dynamical response in Refs.18,47, while the phase is used in Ref.49. In this work, we use the amplitude as an output signal. The virtual neurons \(u_{k,i}\) are then defined as \(u_{k,i}= s[t_{0} + (k-1 + i/N_{\textrm{node}})t_{\textrm{p}}]\), where \(t_{0}\) is the time at which the first input signal is injected. The number of virtual neurons is denoted as \(N_{\textrm{node}}\). We inject the input signal \(q_{k}\) (\(k=1,2,\ldots ,N\)) to the STO and evaluate \(u_{k,i}\), where N is the number of input signals. After that, we introduce a weight \(w_{D,i}\) and evaluate its value to minimize the error,

$$\begin{aligned} \sum _{k=1}^{N} \left( \sum _{i=1}^{N_{\textrm{node}}+1} w_{D,i} u_{k,i} - y_{k,D} \right) ^{2}, \end{aligned}$$
(10)

where \(u_{k,N_{\textrm{node}}+1}=1\) is a bias term. The process determining the weight is called learning. Note that a weight is introduced for each target data and delay.

After the weight is determined, we inject a different series of input data \(q_{n}^{\prime }\) (\(n=1,2,\ldots ,N^{\prime }\)), where \(N^{\prime }\) is the number of input data \(q_{n}^{\prime }\) and is, in general, not necessarily the same as N. The response from the ith neuron with respect to the nth input signal \(q_{n}^{\prime }\) is denoted as \(u_{n,i}^{\prime }\). Next, we introduce the system output defined as

$$\begin{aligned} v_{n,D}^{\prime } = \sum _{i=1}^{N_{\textrm{node}}+1} w_{D,i} u_{n,i}^{\prime }. \end{aligned}$$
(11)

The target data defined from \(q_{n}^{\prime }\) is denoted as \(y_{n,D}^{\prime }\).

If the learning is done well, the system output \(v_{n,D}^{\prime }\) closely reproduces the target data \(y_{n,D}^{\prime }\). The reproducibility is quantified by the correlation coefficient,

$$\begin{aligned} \textrm{Cor}(D) = \frac{\sum _{n=1}^{N^{\prime }}\left( y_{n,D}^{\prime } - \langle y_{n,D}^{\prime } \rangle \right) \left( v_{n,D}^{\prime } - \langle v_{n,D}^{\prime } \rangle \right) }{ \sqrt{ \sum _{n=1}^{N^{\prime }}\left( y_{n,D}^{\prime } - \langle y_{n,D}^{\prime } \rangle \right) ^{2} \sum _{n=1}^{N^{\prime }} \left( v_{n,D}^{\prime } - \langle v_{n,D}^{\prime } \rangle \right) ^{2}}}, \end{aligned}$$
(12)

where \(\langle \cdots \rangle\) represents the averaged value. When the system output \(v_{n,D}^{\prime }\) completely reproduces the target data \(y_{n,D}^{\prime }\), \(|\textrm{Cor}(D)|=1\). On the other hand, when \(v_{n,D}^{\prime }\) is completely different from \(y_{n,D}^{\prime }\), \(|\textrm{Cor}(D)|=0\). Therefore, the correlation coefficient \(\textrm{Cor}(D)\) quantifies the reproducibility of the input signal D times before the present data.

As mentioned, the short-term memory capacity is the memory capacity when the target data \(y_{k,D}\) is given by Eq. (9). The short-term memory capacity is defined as

$$\begin{aligned} C_{\textrm{STM}} = \sum _{D=1}^{D_{\max }} \left[ \textrm{Cor}(D) \right] ^{2}. \end{aligned}$$
(13)

The maximum delay \(D_{\textrm{max}}\) is 20; see also Supplementary Information for the dependence of \(\textrm{Cor}(D)\) on D. Note that, in this work, the short-term memory capacity is defined from the correlation coefficient \(\textrm{Cor}(D)\) from \(D=1\), as can be seen in Eq. (13), whereas, in other work29, the correlation coefficient for \(D=0\) is included in the definition.