Introduction

Substantial mathematically minded economic research has dealt with the question whether financial data contain in a substantial manner low-dimensional chaotic features, or whether their nature requires the stochastic approach predominantly used by the practitioners dealing with the analysis of financial markets. A comprehensive survey of the various methods that have been used, and the conclusions they have arrived at, has recently been published1. In that survey and more generally, the approaches advocating a chaotic nature of the data have almost exclusively relied on indirect statistical quantifiers, such as evaluation of Lyapunov exponents and fractal dimensions2,3, data recurrence quantification analysis4, or BDS-testing5. While such methods may suggest the presence of a chaotic process in the data, an ultimate test by comparing directly financial time series with their stochastic and with their low-dimensional deterministic modeling, is still missing. We will provide here such a comparison.

In theoretical studies, financial time series have predominantly been seen as pure jump processes6,7, drift-diffusion models8,9,10, jump-diffusion models that combine the former two9,10, or as stochastic volatility models11,12. Many of these models are built on the assumption of \(\alpha\)-stable distributions having finite moments to the order needed, or on distributions truncated to have such properties. In physics, the connection between diffusion and maps and their generated distributions has been a prominent theme13,14,15,16,17,18. Recently, wavelet approaches were used as well19,20, as wavelets are a convenient tool for, e.g., distinguishing activity jumps from noise21,22.

A major goal in practical econometrics is the prediction of the mean and the volatility of markets. A standard stochastics-based tool for this is the ARIMA-GARCH procedure23, where the ARIMA part of the approach models and forecasts the first moment (the mean), and the GARCH part the second moment (the volatility) of the process. Recent dynamics of financial markets have shown spikes both in volatility and in performance (emerging from the impact of COVID-19 in a black swan event24), effects that are difficult to deal with. Such phenomena, however, exhibit a strong affinity with the kind of data commonly recorded in neuroscience from cortical neurons, where the neuron’s activity is the result of many single (excitatory or inhibitory) synaptic stimulations25. The neuron’s response can be seen as the result from impinging input that can, in a generalized sense, seen to be coupled by the neuron’s strongly nonlinear internal machinery. Generally, the produced nonlinear effects would be difficult to predict; in the case of a neuron they can, however, be described by a simple two-dimensional deterministic Rulkov-map26. This map reproduces the whole spectrum from simple to very complex behavior of neurons27, including effects similar to what we see in financial data. For applying this picture to the finance markets, relaying buying-selling activities towards the observed collective financial market behavior, the eco-societal environment would act as the coupling element.

Using this perspective, we will directly compare financial market data with the results from a modified Rulkov-map approach, and, as a competitor, with the commonly used optimized ARIMA-GARCH model, demonstrating that the modeling results only differ by what a small variation of white Gaussian noise would trigger. In addition to this, the map-based description offers advantages in terms of conceptual simplicity and computational speed (cf. “Rulkov-map modeling of the Financial stress index” and “Conclusions” sections). In particular, an obvious asset of the deterministic model is that it operates in a low-dimensional subspace of models defined by means of the form of the deterministic equation, whereas the stochastic approach lacks such a limiting element. Whereas the ARIMA-GARCH approach deals separately with mean and volatility, in our two-dimensional deterministic approach, the first vector component represents an instantaneous activity that is functionally linked to an underlying trend represented by the second vector component. With this, the Rulkov-map equation offers insight into how the driving sub-processes are entwined and to what effects, e.g., a change of one of its few parameters may lead. By this, our explicit deterministic approach opens a new perspective towards understanding financial markets. In what follows, we present a description of our novel modeling approach, followed by review of the classical modeling tools that we have to compare with. Then, we will recall the most popular measures for model comparison, before we display and finally discuss the results obtained, including the performance obtained from two more recent models. Our main focus will, however, not be that of a competition between the stochastic vs. deterministic approaches, but to show that there is a substantial deterministic part in the data that could be exploited for the explanation and prediction of financial market data.

Modeling tools

Generalized Rulkov map

During the recent turbulent years, ten Asian emerging stock markets showed non-linear serial dependence, and in real exchange rates of developing and emerging market economies28, in four USA stock market indices29, and in USA short rate indices30, mean reversion was observed; jumps are generally common in financial time series31. Similarly, biological neurons exhibit a two-phase response: an activation phase, characterized by a transient and presence of fast oscillations, and a subsequent plateau phase characterized by more regular oscillations. The two-dimensional Rulkov map family of biological neurons26 combines a one-dimensional fast subsystem (x) and a one-dimensional slow subsystem (y), where the latency to a stimulating event depends on the fast-slow dynamics. While this interplay between fast-slow dynamics is a general mechanism unterlying pattern formation, the Rulkov map offers through the involved parameters the possibility to adapt the general mechanism to specific features exhibited by a process modeled. In particular by its recursive nonlinear and mean reverting properties, Rulkov maps may be expected to be highly suitable for the modeling of financial  time series, in particular regarding the occurrence of data clusters, heteroskedasticity, mutual synchronization, chaos and regularization of bursts of activity across the markets, after-shock asset classes, and more32. In our application, we will optimize the Rulkov family model to best reproduce the processes and events observed in our data.

To provide a first general insight into the nature of Rulkov maps, we start from a two-dimensional non-linear recurrence of the form

$$\begin{aligned} x_{t+1} = f_\alpha (x_t) + \gamma \, y_t + \delta , \, \text{ with } \, f_\alpha (x_t) = \dfrac{\alpha }{1+x_t^{n}}, \qquad n \in {\mathbb {N}}, \, \alpha , \gamma , \delta , x\in {{\mathbb {R}}}. \end{aligned}$$
(1)

Such a setting is widely used for the modelling short-time effects like pulses and shocks in time series. Vector component y captures a moving average trend implemented as

$$\begin{aligned} y_{t+1} = \beta \, y_t - \mu \, x_{t} + \eta ,\text{ with } \, \beta , \mu , \eta , y \in {{\mathbb {R}}}. \end{aligned}$$
(2)

Combining Eq. (1) with Eq. (2) leads to

$$\begin{aligned} x_{t+1}= & {} f_\alpha (x_t) + \gamma \, y_t + \delta , \nonumber \\ y_{t+1}= & {} \beta \, y_t - \mu \, x_{t} + \eta . \end{aligned}$$
(3)

This version, that differs slightly from the original Rulkov version, is the map that we will use for our financial time series modeling; its two-dimensional form takes equally care of the short-term (x) and the average prediction (y) horizon (the two main subjects of interest in financial modeling), which provides the basis for an optimal modeling framework. Of importance are the following characteristics hosted by function \(f_{\alpha }\).

Proposition 2.1

For any (even) integer n, the function \(f_{\alpha }(x)\), defined in (1), satisfies:

  1. (i)

    If \(\alpha >0\) then \(f_{\alpha }(x)\) assumes its global maximum value (\(\alpha\)) at \(x=0\); otherwise, \(f_{\alpha }(x)\) assumes its global minimum value (\(\alpha\)) at \(x=0\)

  2. (ii)

    For \(x\rightarrow \pm \infty\), \(f_{\alpha }(x)\) approaches 0

  3. (iii)

    \(\lim _{n\rightarrow +\infty } f_{\alpha }(x)=\alpha \Pi \biggl (\frac{x}{2}\biggr )\), where \(\Pi \biggl (\frac{x}{2}\biggr )=0\,\,\,\, \text{ for }\,\,\,|x| > 1, \, \,\, \text{ and } \,\,\, \Pi \biggl (\frac{x}{2}\biggr )=1\,\,\, \text{ for }\;\,\,\,|x|\le 1\) (the rectangle function)

Proof

Without loss of generality we set \(\alpha >0\)

  1. (i)

    \(f'_{\alpha }(x)=-\frac{n \alpha x^{n-1}}{(1+x^n)^2}\), for any \(x\in {{\mathbb {R}}}\). This equals zero only for \(x=0\), is larger than zero for \(x<0\) and smaller than zero for \(x>0\). Hence, \(x=0\) is a global maximum for \(f_{\alpha }\)

  2. (ii)

    is immediate

  3. (iii)

    is the consequence of \(\lim _{n\rightarrow +\infty } \frac{1}{1+(2x)^{n}}=\Pi (x)\), for any even integer n, \(x\in {{\mathbb {R}}}\)

\(\square\)

Figure 1 exhibits that the choice of \(n=2\) provides the x-sensitivity that according to our tests is desirable; this will be our choice in the following, unless specified otherwise.

Figure 1
figure 1

\(f_\alpha (x)\) for different values of n, using \(\alpha =1\).

For our demonstrations, we will mainly use single Rulkov maps; to simulate the interaction between markets, asset classes, and more, Rulkov maps labelled by indices i and j were coupled according to \(\{x_{t+1}^{(i,j)} = f_{\alpha ^{(i)}}(x_t^{(i)}) + \gamma ^{(i)} \, y_t^{(i)} + \delta _{t-\tau ^{(i)}}^{(i,j)}, \,\, y_{t+1}^{(i)} = \beta ^{(i)} \, y_t^{(i)} - \mu ^{(i)} \, x_{t}^{(i)} + \eta ^{(i)} \}\), where the coupling \(\delta _{t-\tau ^{(i)}}^{(i,j)} = \dfrac{1}{1+\exp (-\kappa ^{(i)} x_{t-\tau ^{(i)}}^{(j)} - \theta ^{(i)})}\) describes a time-lagged interaction between the time series i and j (and likewise for interchanged indices). The idea underlying this specific coupling is that such an interaction will mostly occur on a short time-scale. \(\alpha , \beta , \gamma , \mu , \eta , \delta , \kappa , \theta\) are system parameters, whereas \(\tau\) coincides with the time-series specific embedding dimension used for the reconstruction of the phase-space from the one-dimensional time series2,33. If a single time series is our object of study, we will use the term specification “single”; if we consider the coupling among two time series, we will use the term specification “coupled” to indicate this.

For our applications of the Rulkov map, the involved parameters must be chosen appropriately. For this, we extract the x and y initial conditions from the beginning of the data (for details see “Forecasting” section), then calibrate the parameters by running over the rest of the data a robust nonlinear least squares regression, using an iteratively re-weighted least squares algorithm that, at each iteration, recomputes the weights based on the residual from the previous iteration. This process progressively continues until the weights converge. For details see Holland et al.34.

ARIMA-GARCH

To assure validity, we will mainly benchmark our optimized deterministic Rulkov-map approach against an optimized ARIMA-GARCH model (consisting of optimized Auto-Regressive Integrated Moving Average model (ARIMA) combined with an optimized Autoregressive Conditional Heteroskedasticity model (GARCH)). Probably the currently most used and best-documented stochastic approach23. A less detailed comparison to the performance of more advanced versions of the stochastic approach is furnished in the context of the conclusions from this comparison.

In financial data, potential serial correlations deserve special attention. Their effects can be eliminated by differencing the original non-stationary time series once, or if required, a higher number of times (the procedure must, however, not be pushed too far, as higher \((> 2)\) orders tend to distort the desired frequency components35). In our tests up to order four, only the first order yielded improvements. The resulting stationary time series can then be treated by the Box-Jenkins methodology (ARMA or ARIMA)23,36.

The AR part of ARIMA combines a linear extrapolation based on past observations and the MA part contributes the influence by past errors. In short note, for the optimal ARIMA-model, the equation \((1-\sum _{i=1}^p \varphi _i L^{i})(1-L)^d x_t=(1+\sum _{i=1}^q \theta _i L^{i}) \varepsilon _t\) is solved, where x denotes the data values, \(\varepsilon\) the errors (white noise), L the time-shift operator needed for a d-fold differencing towards stationary signals, p the lag-level considered and q the order of the moving-average model. \(\varphi _i\) are the linear autoregressive coefficients and \(\theta _i\) the moving average coefficients that apply to past errors. From this, the GARCH part then evaluates the variance of the process conditioned on the observed past data values and variances. The ARIMA(p,d,q)-model providing the optimal description of the data can then be used for forecasting. For example, given an ARIMA(2,1,0)-model, we obtain \(X_t=\mu +(\varphi _1 (x_{t-1} - \mu )) + (\varphi _2 (x_{t-2} -\mu )) + \varepsilon _t\) for the stationary time series, where \(\mu\) is the mean of time series \(x_t\) (expected to be small), and \(\varepsilon _t\) is white noise term with zero mean and constant variance. The GARCH part of the approach estimates the volatility based on a convex linear combination of its past values, its long-average persistence term, and the exponentially weighted moving average on past data; GARCH(a,b) models take care of contributions that lag further back in time a with respect to the generating time series, b with respect to the past variances. The determination of the \(\{p,d,q\}\) and \(\{a,b\}\) polynomials is made by using the Bayesian information (BIC) and the Akaike information (AIC) criteria37. The parameter-optimized ARIMA-GARCH model will be denoted in the following by ARIMA-GARCH\(^*\). When in the following we mention the calibration of an algorithm, we mean that its parameters are optimized over the entire or a part of the data.

Forecasting

A forecasting process consists of multiple estimations over a rolling window, i.e, in distinction to the parameter calibration over the full data set, we calibrate the ARIMA-GARCH and the Rulkov map on a window, or partition, of the data set. Given our weekly data, we take the full time series and consider only the first 52 data points (i.e., a trading year). We extract over that partition the best ARIMA-GARCH or Rulkov model, and we forecast the next value (i.e., the 53rd data point). In the next step, we consider the interval starting with the second data point to the 53rd data point, and redo all the calibration; with the calibrated model we obtain a new forecast for the 54th data point, etc. For the ARIMA-GARCH process, this involves two calibrations (for the x and the y separately). For the Rulkov map, this must be only done once.

A standard cost-effective zero-level benchmark against which more sophisticated model predictions can be compared is the naïve model36. This model forecasts a value equal to last period’s observed values, resulting in a in-sample one step ahead algorithm. A time series \({\hat{x}}_{N+t}=x_N (t \ge 1)\)

\(X = \lbrace x_h: h \in [1\,\,N] \rbrace\), is forecasted time t ahead as \({\hat{x}}_{N+t}=x_N \,\, (t \ge 1)\), which produces an error of size \(\dfrac{1}{N-1} \sum _{h=2}^{N} \left| x_h - x_{h-1} \right|\).

Elements of model results comparison

To compare modeling approaches, we use a set of metrics:

Normalized square root mean error

The root mean squared error (RMSE) defined as

\(RMSE=\sqrt{\frac{1}{N}\sum _{h=1}^N e^2_h}\), measures the accuracy of a model in terms of goodness of fit, where \(e_h\) denote the residuals between the observations and the simulations over N time steps. Hence, the values below unity indicate a good fit. RMSE, however, depends on the scale of the observed data and is, by this virtue, outlier-sensitive. For this reason, we will apply the normalized root mean squared error (NRMSE):

$$\begin{aligned} NRMSE= \dfrac{RMSE}{x_{\max } -x_{\min }}, \end{aligned}$$
(4)

where \(x_{\max }\) denotes the maximum value and \(x_{\min }\) the minimum value of the observed sample data.

Relative mean absolute error

The relative mean absolute error is

$$\begin{aligned} RMAE = \dfrac{MAE^{(1)}}{MAE^{(2)}}, \end{aligned}$$
(5)

where \(MAE^{(1)}\) is the mean absolute error of the model of interest and \(MAE^{(2)}\) is the error of some benchmark model38. A value greater than unity indicates that the chosen model performs worse than the benchmark.

Mean absolute percentage error

To measure the prediction quality of a model, often the mean absolute percentage error (MAPE)

$$\begin{aligned} MAPE=\frac{1}{N}\sum _{h=1}^N \biggl | \frac{e_h}{x_h}\biggr | \end{aligned}$$
(6)

is used, where \(e_h\) denotes the residuals between the observations \(x_h\) and their simulations over N time steps39. Table 1 exhibits the generally accepted MAPE accuracy levels40.

Table 1 MAPE interpretation guidelines.

Dynamic time warping distance

To scrutinize how close models are, dynamic time warping distance (DTWD) can be used. This technique finds the optimal alignment between two time series by stretching or shrinking on time series along the time axis, and then evaluates the Euclidean distance between the two41,42. This commonly used technique in finance and econometrics32,43,44 copes with slightly temporally varying responses in one of the time series45. This feature has made DTWD the predominant tool in speech recognition (for determining whether two waveforms represent the same spoken phrase) and one of the standard methods used for classification problems in data mining46,47. In our application, DTWD will include information on how much the predicted time series is shifted against the data modeled. In the comparison between different modelings, such effects may play a role, but are expected to be close to the ordinary Euclidean distance between time series. DTWD-results obtained for the artificial system displayed in Fig. 2 will later be used to assess how strong distinct modeling approaches of financial market data differ.

Figure 2
figure 2

Calibration of DTWD effects: (a) Perturbed sawtooth wave (blue), the same upon the addition of white Gaussian noise (black, \(\sigma =1/2\)) and, finally, of white Gaussian noise of doubled strength (red, \(\sigma =1\)). (b) and (c) Signals after DTWD application (same coloring).

Data investigated

In Table 2 we provide an overview of the empirical financial time series used for our numerical evaluations. Whereas common standard dataset abbreviations will be used otherwise, the abbreviation BAMLEM will refer to the Emerging Markets Corporate Plus Index Total Return Index Value.

Table 2 Datasets used: (i) The Financial Stress Index STLFSI248 measures the degree of financial stress in the markets.

Rulkov-map modeling of the financial stress index

In finance, the fast x and the slow y components correspond to the actual data and their memory, respectively, where the latter is approximated by the exponentially weighted moving average (EWMA). Since the two components follow well-separated time scales, this suggests to characterize the instability of each component individually by means of a one-dimensional Lyapunov exponent, although the underlying system is in fact two-dimensional.

The Financial Stress Index STLFSI248, see Fig. 3, will be our primary focus in financial data modeling. It is obtained from seven interest rates, six yield spreads, and five other indicators by means of principal component analysis (PCA), as each of these variables captures aspects of financial stress. Accordingly, whenever the level of financial stress in the economy changes, the data series are expected to change in a coherent manner. Positive stress values indicate higher-than-average stress in financial markets, while negative values indicate lower-than-average levels of stress. Our financial stress data covers the 2008–2009 financial crisis, as well as the ongoing turmoil in financial markets stemming from the fear and uncertainty associated with the COVID-19 pandemic. As the best ARIMA-GARCH models in terms of Bayesian information criterion (BIC) and Akaike information criterion (AIC) were those integrated of order one36, we calibrated all models investigated (Rulkov map, Naive model and the optimized ARIMA-GARCH model) over the first differences of the STLFSI2.

Figure 3
figure 3

(Source: FRED48).

Financial stress index STLFSI2. USA recessions correspond to shaded areas.

Below we report the accuracy of data modeling achieved by our Rulkov map approach, then the latter’s potential for making forecasts, and we finally present evidence of chaotic behavior in the data. For the Rulkov map, an \(\{x,y\}\)-pair provided the input, from which a corresponding output \(\{{\hat{x}}, {\hat{y}}\}\) was generated, where for with every x-value, an exponential weighted STLFSI2 moving average yielded the corresponding y-value. In the stochastic approach, after the two optimized ARIMA-GARCH\(^*\) models were determined (one for the x and one for the y-coordinate), similarly each \(\{x,y\}\)-pair was fed into the model, generating a corresponding output \(\{{\hat{x}}, {\hat{y}}\}\) (for model details, cf. “ARIMA-GARCH”). In Fig. 4 top row, we show how the Rulkov map given the properly calibrated parameters of Table 3, replicates the bursting dynamics of STLFSI2. In the second row, we display the measured error and, in the third row, we compare this to the one obtained from a calibrated ARIMA-GARCH\(^*\) model (where the properly calibrated parameters are given in Table 4). The greater simplicity of the model with substantial difference in the number of involved parameter, leads to a speed-up of the Rulkov calibration process by a factor of 20, compared to the ARIMA-GARCH calibration process.

Figure 4
figure 4

(a) Rulkov variable \({\hat{x}}\) (solid blue line) and STLFSI2’s immediate value x, dashed red line); (b) Rulkov variable \({\hat{y}}\) (solid blue line) and STLFSI2’s long term value y (approximated the moving average over \(m=52\) data points of x, dashed red line). (c) Difference between Rulkov \({\hat{x}}\)-value and STLFSI2’s value x; (d) Difference between Rulkov’s \({\hat{y}}\) and STLFSI2’s long term value y. (e) Difference between \({\hat{x}}\)-variable of the calibrated ARIMA-GARCH\(^*\) and the immediate value of STLFSI2 x; (f) Difference between the \({\hat{y}}\)-variable of the calibrated ARIMA-GARCH\(^*\) and STLFSI2’s long term value y.

Table 3 Parameters of the Rulkov map used for the STLFSI2 data set.
Table 4 ARIMA-GARCH* (2,1,2)-(1,1) coefficients for variable x and ARIMA-GARCH* (2,1,1)-(2,1) coefficients for variable y, listed according to the AR-I-MA-G-ARCH parts of the algorithm (where two of the constants appearing in the ARIMA as well as in the GARCH parts have been added, leading to 16 (instead of 20) parameters to be listed).

Among the remaining five financial market time series, there are twenty possible couplings. For space reasons, we will only display the most representative ones; the omitted ones follow their behavior closely. In their treatment, we occasionally observed correlations on the residuals, which led us into further investigations on integration, mean reversion, and ARCH effects, suggesting that some financial time series might not properly satisfy the finiteness condition imposed on the first four distribution moments49.

Chaotic or random data?

To check whether expected chaotic aspects are indeed in the data, and if so, how well they are reflected by the model, we followed the standard approach to evaluate for the data and the obtained simulations the maximal Lyapunov exponents (MLE)2,3,50,51,52,53,54,55. The MLE calculated on the first differences of STLFSI2 yields a value of 0.22, to be compared to 0.26 obtained for the Rulkov map. These values are close to those of the logistic map at nonlinearity parameter around 3.6 (chaotic phase), and point at the presence of low-dimensional chaos in the data. The approximate entropy (AE) is used to quantify the amount of regularity and the unpredictability of fluctuations over time-series56; applications to financial time series are wide-spread57,58,59. If the value is very small, this implies that the sequence is regular. As an example, for the binary signals S1 = [0 1 0 1 0 1 0 1 0 1 0 1 0 1 0 1] (as a regular alternation of 1 and 0’s) and S2 = [1 1 0 1 1 1 1 0 1 0 1 0 0 0 0 1] (as a part of a random sequence of 0 and 1’s) AE yields 0.0022 and 0.6283, respectively. Applied to the first data differences, the result obtained for STLFSI2 was 0.5045, versus 0.6368 of the Rulkov map modeling.

The correlation fractal dimension D is a standard measure of the dimensionality of the space occupied by a set of random points2,3,60,61. For the first differences of STLFSI2 and of the Rulkov map, the evaluated correlation dimensions are 4.37 and 4.65. These values are in line with what was reported earlier in Orlando62,63, see also59,64. Generally, the interpretation of such estimates, must, however, be taken with care, as differencing can be seen as filtering, which will generally affect the results22,65. Moreover, a difficulty of the method is to arrive at the required saturation of the curves for increasing embedding dimensions in the presence of noise. Therefore, the exhibited results are more of a corroborating nature. The Hurst exponent (H) of single time series is directly related to D by the relation \(D = 2 - H\)66,67,68. For the first difference of STLFSI2 to the corresponding Rulkov map, we found for the x-dynamics \(H=0.40\) and \(H=0.69\) for the y-dynamics, respectively. For various time-series, Table 5 lists the MLE, AE, D and H, together with the values obtained from the corresponding Rulkov map approach, evidencing a good agreement between the model and the empirical time-series data properties. For all these computations, the investigated time series were of a sufficient length.

Table 5 Chaotic descriptors Maximal Lyapunov Exponent MLE, Correlation Dimension D, Hurst exponent H of financial time series TS (all figures refer to differenced/detrendized data).

Details of model comparison

The autocorrelation function (ACF) and the partial autocorrelation function (PACF) on the residuals for state variables x and y of the Rulkov map and for the ARIMA-GARCH\(^*\) model (Fig. 5) indicate that the first order differencing is more effective in removing autocorrelation on the residuals for the Rulkov than for the ARIMA-GARCH\(^*\) approach. This could be seen as a sign of the greater explanatory power of the Rulkov map: Although sometimes correlated residuals are features of the research design or intrinsic to the studied phenomenon, “..the inclusion of correlated residuals in latent-variable models is often regarded as a statistical sleight of hand, if not an outright form of cheating”69. As STLFSI2 is a combination of other indices and, according to  Lo et al.70,71, there is significant evidence of positive serial correlation for weekly and monthly holding-period index returns. Generally, weekly and monthly stock returns are weakly negatively correlated, whilst daily weekly and monthly index returns are positively correlated72. This effect has been explained by positive cross-autocorrelations across individual securities across time, where it may happen that this cannot be removed completely72. The absence of a unit root, however, has been confirmed using the KPSS test73.

Figure 5
figure 5

Autocorrelations ACF and partial autocorrelations PACF for state variables x (first two columns) and y (last two columns), comparing the Rulkov map to ARIMA-GARCH\(^*\) (ARIMA(2,1,2)-GARCH(1,1)). While autocorrelation is almost absent in the Rulkov map, the ARIMA-GARCH\(^*\) displays significant autocorrelations. This could be taken as an indication of the greater explanatory power of the deterministic approach.

To determine how close the Rulkov and the ARIMA-GARCH\(^*\) approaches would be, we applied the commonly used the DTWD measure. We first calibrated DTWD at the simple example of a perturbed sawtooth wave that was replicated upon adding Gaussian white noise of variable strength (cf. Fig. 2). Figure 6 shows that the warped residuals of the Rulkov and the ARIMA-GARCH\(^*\) x-component are almost identical and closely follow the original timeline (in the sense that the warping did not essentially compress or stretch the dependent time series). For the y component, we obtained corroborating results (Table 6, corresponding plots are withheld for space reasons). With the only exception of the DTWD for \(x^j\) of the IBOV, our distances in Table 6 are below the DTWD generated by Gaussian white noise. From this, we conclude that the distinction between the ARIMA-GARCH\(^*\) and the Rulkov map approaches should be of, or even below, the order of the influence of noise.

Figure 6
figure 6

Rulkov map and ARIMA-GARCH\(^*\) residuals almost coincide.

In Table 6 we report the RMAE accuracy of the Rulkov map, of the Naive, and of the ARIMA-GARCH\(^*\) models, the NRMSE of the Rulkov map, of the ARIMA-GARCH\(^*\) model, and their DTWD. As may be expected, in all instances the Rulkov map performs substantially better than the Naive model, whereas a comparison between the Rulkov and the ARIMA-GARCH\(^*\) yields mixed results: Whereas the Rulkov map’s NMRSE is always lower, its MAE is occasionally higher. The differences are small in all cases, which is corroborated by the residuals’ DTWD.

Table 6 Error and DTWD measures of Rulkov map vs. Naive and vs. ARIMA-GARCH* models.

Forecasting power

Of particular interest is that the Rulkov map may easily be used for forecasting purposes. Indeed, once the values \(\{x_s,\;y_s\}\) at time s are known, the previsions at next time \(t>s\) are computed deterministically through Eq. (3). In our set-up the parameters are calibrated over the previous window \([s-m+1,s]\) of size m, opportunely chosen. Because we deal with a weekly frequency dataset, we set \(m=52\). In particular, starting from m, we predict the next values through a rolling window of fixed size m. The parameters are calibrated by a least squares regression. We run a robust estimation with the iteratively re-weighted least squares algorithm34 which, at each iteration, recomputes the weights based on the residual from the previous iteration. This process progressively continues until the weights converge. Using MAPE, the generated forecasts are compared with those of the ARIMA-GARCH\(^*\) model, see Table 7 and Fig. 7. The comparison provides evidence of fits of similar quality, both for the process and its trend, confirming that a deterministic approach performs as well as a stochastic approach.

Figure 7
figure 7

Absolute percentage error between the forecasts produced by the Rulkov map and the ARIMA-GARCH\(^*\) model, separately for \(x_t\) (plot a) and \(y_t\) (plot b).

Table 7 MAPE of the Rulkov and the ARIMA-GARCH* model.

Finally, we include a summary of results that we have obtained from the application of two more recent approaches to our data. The first is the G2++ model, a popular two-factor Gaussian model where the state process is given by the sum of two correlated Gaussian factors plus a properly chosen to fit real data deterministic function. The model is analytically tractable and its (conditional) distribution, as well as its moments, is given in closed formula. For the aforementioned reasons, Gaussian models like this G2++ model are useful in practice (see74, Chapter IV]). The second, the JOU-model, is a (mean-reverting) Ornstein-Uhlenbeck process with Poisson jumps75,76. The results in terms of error for the calibrated models with our optimized parameters evidences that in terms of RMAE and DTWD, and MAPE for forecast, the Rulkov map performs better than the two alternatives, although the JOU approach’s results were quite close to the Rulkov and ARIMA-GARCH* results (details of the comparison available upon request from the authors). The calibration time of the JOU approach was roughly thrice that of the Rulkov approach; G2++ required only slightly more time than Rulkov.

Conclusions

The low DTWD and the almost coinciding residuals of the Rulkov map and ARIMA-GARCH* exhibit a similar quality of the two approaches, indicating that the deterministic chaotic model performs as well as an ultimately refined stochastic one. Furthermore, the low-dimensional deterministic Rulkov map approach suffers less of residuals’ autocorrelation, hosts an improved NRMSE and a better DTW property. A main shortcoming of the ARIMA-GARCH is that generally it does not offer a nice handle towards the understanding of its building and the meaning of the resulting coefficients. The explicitness and simplicity of the Rulkov map opens, in contrast, the door for a deeper understanding of the drivers of the market dynamics. For a full understanding the consequences that a variation of the involved parameters will entrain in the context of the real-world financial markets, a more explicit mapping of the model parameters to financial market parameters will, however, be needed. This would require the building of a kind of conceptual real-world model, helped by the Rulkov equation, but replacing it. Once the essential market parameters would been identified in this way, the degree of precision of such a model can be assessed by using a recent high-level model testing77.

Regarding the long-standing debate whether we should consider financial times series as stochastic or rather chaotic78, our direct modeling approach yields evidence for a stronger chaotic data component than what one could probably expect. While both elements seem to be present, the chaotic component offers several options for understanding and, finally, monitoring financial processes, if choosing the appropriate time-scale. In particular, our approach may be expected to host an improved explanatory power to events of interest, such as shocks that emerge as endogenous, instead of the exogenous nature supposed by stochastic models24. Given this situation, we are happy to show that a substantial contribution of determinism is found in the data4,79. For forecasting, the latter is sufficient; the question whether the whole of the data should be termed chaotic or stochastic is, seen in this perspective, of a more ’academic’ nature1,63, where the contradicting arguments available may indicate a lack of appropriate technical or theoretical descriptors to deal with such type of data.

More generally, what we deal with here is similar to the one encountered in neuroscience, where initial ideas of simple chaotic or of power-law behavior characterizing the data, had to be replaced by more complicated and, in particular, scale-dependent characterizations80. This shifts the question towards determining exactly at what time-scales the two processes of low-dimensional chaos and of stochastic, respectively, are dominant, and to identify the horizon over which a deterministic prediction will yield optimal forecasting results32. In this way, our work may open a general avenue towards a better understanding and monitoring of the essential drivers in the data, not only in financial market dynamics, but also in similar processes beyond.