Abstract
Humans and animals can effortlessly coordinate their movements with external stimuli. This capacity indicates that sensory inputs can rapidly and flexibly reconfigure the ongoing dynamics in the neural circuits that control movements. Here, we develop a circuit-level model that coordinates movement times with expected and unexpected temporal events. The model consists of two interacting modules, a motor planning module that controls movement times and a sensory anticipation module that anticipates external events. Both modules harbor a reservoir of latent dynamics, and their interaction forms a control system whose output is adjusted adaptively to minimize timing errors. We show that the model’s output matches human behavior in a range of tasks including time interval production, periodic production, synchronization/continuation, and Bayesian time interval reproduction. These results demonstrate how recurrent interactions in a simple and modular neural circuit could create the dynamics needed to control timing behavior.
Similar content being viewed by others
Introduction
Sensorimotor coordination in humans is remarkably flexible. We can anticipate events based on few observations and use that information to adjust our movements. For example, musicians can use a metronome to adjust the tempo of their movements, and children can rapidly coordinate their movements during a clapping game. However, we still lack an understanding of how networks of neurons generate such coordinated movements.
Recent studies have proposed that the neural basis of sensorimotor coordination may be understood using the language of dynamical systems1,2,3. The key intuition is that recurrent neural networks in the motor cortex form dynamical systems4,5 whose output can be controlled by sensory inputs2,6,7,8,9. This idea has been explored in large-scale distributed recurrent neural network models10. However, the complexity of these network models often makes it difficult to understand their behavior from an algorithmic perspective.
Timing provides a prime example of sensorimotor coordination and is crucial in behaviors demanding the generation of delayed motor responses, generation of rhythmic movements with a desired tempo, and synchronization of movements to anticipated events. An early model proposed that the brain controls action timing by adjusting the speed of an internal clock whose ticks are integrated toward a fixed threshold11. Consistent with this proposal, experiments in animal models have found that neural activity in anticipation of a delayed response reaches a fixed threshold12,13,14 at a rate that is inversely proportional to the delay period7,15,16. These results suggest that the brain supports flexible timing by controlling the speed at which neural activity approaches a movement initiation threshold.
Recently, it was shown that flexible control of speed can be achieved through nonlinear interactions within a simple model consisting of a pair of units with reciprocal inhibitory connections7. In this model, the speed at which the output evolves toward a movement initiation threshold can be adjusted flexibly via a shared input (Fig. 1a). From a dynamical systems perspective, this model can be viewed as an open-loop controller that converts an instruction (i.e., shared input) to the desired dynamics (i.e., speed).
However, the larger utility of this model depends on whether it can be extended to accommodate additional constraints associated with temporal control of movements. In particular, temporal control of movements must accommodate noise in the nervous system17, prior expectations18, and sensorimotor delays imposed by internal19 and external20 temporal contingencies. These factors limit the capacity of open-loop systems to achieve robust temporal control. Efficient control systems often rely on sensory feedback to combat noise, and when sensory feedback is delayed, they rely on a mechanism to predict future sensory and motor states19,21,22. Here, we augment the original open-loop system with a sensory anticipation module (SAM) and a feedback mechanism, and show that the resulting neural circuit model can accommodates internal noise, prior expectations, and sensorimotor delays. Finally, we demonstrate that the model can capture key features of human behavior in a number of classic timing tasks (Fig. 1b–e).
Results
We will describe the full model in four steps. We start by introducing a basic circuit module (BCM) that acts as a flexible open-loop controller for producing desired time intervals. We extend the BCM to a motor planning module (MPM) capable of producing isochronous rhythms. We then introduce a SAM that provides the means for anticipating and predicting upcoming temporal events. Finally, we introduce the full model that combines the MPM and SAM to create a system that can dynamically coordinate motor plans and actions with anticipated and unanticipated stimuli in a range of behavioral tasks.
BCM for interval production
The BCM has been described in detail previously7. Briefly, the BCM includes three units, u, v, and y, each representing the average activity of a population of neurons (Fig. 1a, top). u and v inhibit one another and receive input, I, that is tonic, or constant over time. y receives excitatory input from u and inhibitory input from v, and leverages the nonlinear dynamics of mutual inhibition between u and v to generate ramp-like activity. Finally, the model initiates a “movement” when y crosses a fixed threshold, y0. The rate dynamics of u, v, and y are as follows:
WuI and WvI denote the strength with which I drives u and v, respectively. Wuv and Wvu denote the strength of inhibitory coupling from v to u, and from u to v, respectively. τ is the time constant of each unit. θ(x) is a sigmoidal function that maps the input to an output between 0 and 1 (see “Methods”). Finally, ηu, ηv, and ηy are stochastic synaptic inputs to each unit and are modeled as independent white noise with standard deviation σn.
We assume WuI = WvI = 6 (identical shared excitatory input), Wuv = Wvu = 6 (symmetric mutual inhibition), Wyu = Wyv = 1, and τ = 100 ms for all units, a value consistent with previous models. With these parameters, the model functions in a dynamical regime with three fixed points: an unstable point at u = v, and two stable fixed points with either u dominating or v dominating (Fig. 1a, middle). In this regime, u and v evolve, or change over time, toward one of the stable fixed points depending on the initial conditions. These dynamics lead to a ramp-like activity in y whose rate is inversely related to I (Fig. 1a, bottom). In other words, the BCM acts as an open-loop controller that converts an instruction conveyed by I to a dynamic output, y, that evolves toward the threshold, y0, at an appropriate speed (Fig. 1a). Therefore, by adjusting I, the BCM can flexibly adjust the movement initiation time.
MPM for periodic production
We next considered the production of multiple movements with a fixed inter-production-interval (IPI). This is challenging with the BCM which can only produce a single action. One solution is to use multiple concatenated BCMs, but that seems unrealistic as the number of modules would grow with the number of movements. We tackled this problem using a modification of the BCM, which we call the MPM. The MPM is identical to BCM except for an additional reset mechanism after each movement that allows the model to generate an arbitrary number of timed outputs (Fig. 2a). From a neurobiological perspective, the reset signal can be generated by the corollary discharge associated with the movement command23,24, or by sensory feedback triggered by the movement. We implemented the reset as a transient 10 ms pulse, Ip, activated right after the threshold crossing (Fig. 2a, dashed line). Mathematically, this can be formalized as follows:
In the MPM, up, vp, and yp evolve identically to the BCM between consecutive threshold crossings (Fig. 2b). In addition to producing an output at each threshold crossing, however, the model activates Ip, which resets up and vp, and causes a rapid drop in yp. This allows the circuit to restart the dynamics and generate another output. As expected, IPIs increase monotonically with I (Fig. 2c, d; r2 = 0.84; F(1, 160) = 828.6; p ≪ 0.01) within a suitable range of inputs (Supplementary Fig. 1).
In humans, IPIs are variable with a standard deviation that increases linearly with the mean25. To evaluate IPI variability in the model, we performed simulations in the presence of Gaussian noise (see “Methods”). When noise levels were not too high (Supplementary Fig. 2), the model exhibited a qualitatively similar behavior to that of humans (Fig. 2e): IPI variability increased monotonically with IPI (one-tailed F test; I = 0.75 to I = 0.76: p = 0.014, F(80, 70) = 0.60; I = 0.76 to I = 0.77: p < 0.01, F(57, 70) = 2.95; I = 0.77 to I = 0.78: p < 0.01, F(40, 57) = 10.93). However, this relationship was nonlinear in the model, which is unsurprising given the simplicity of our assumptions regarding the nature of noise in the brain and the underlying dynamics.
SAM for predicting future events
In many circumstances, the IPI is not known in advance and has to be adjusted based on external timing cues. Here, we considered a case in which the IPI has to be adjusted by the interval between external temporal events. In general, measuring time between events can be achieved in two ways. One way is to implement a circuit whose output is a ramp with a fixed slope. In such a circuit, the output level provides a moment-by-moment estimate of elapsed time (Supplementary Fig. 3a). Alternatively, the circuit may function predictively, and adjust the slope of the ramp such that the output reaches a certain expected level at the anticipated time of the next event (Supplementary Fig. 3b).
A recent study in monkeys found evidence in support of the predictive mechanism20: neural dynamics in the frontal cortex were adjusted so that responses reached an expected state at the expected time of the stimulus. Based on this finding, we developed a SAM that measures time predictively (Fig. 3a). The SAM behaves identically to the MPM except that it does not generate an action when its output, ys, reaches y0. Instead, the SAM adjusts the input so that ys would reach y0 exactly at the expected time of the sensory feedback. When I is too high, ys would increase at a slower pace than needed, and its value at the time of the sensory feedback would be below y0. Conversely, when I is too low, ys would go beyond y0 at the time of the feedback. Accordingly, the system must decrease I when ys < y0, and increase I when ys > y0. The SAM implements these adjustments by updating the value of I dynamically at a rate proportional to the error signal, ys−y0:
In this update rule, K scales the rate of change of I with the error signal, and s is a gating parameter that is set to 1 when the feedback is on and 0 otherwise. This formulation allows the model to update I only when the feedback is on (s = 1). Fig. 3b illustrates the response of the SAM to three equidistant sensory inputs with a short (400 ms) or long (1000 ms) inter-stimulus-interval (ISI). The first stimulus (S1) triggers a transient input, Is. This input does not alter I since the system has not yet measured the ISI, but resets us and vs (Fig. 3a; see “Methods”). Following S1, the output, ys, increases over time. For a short ISI, ys at the time of the second stimulus (S2) is lower than y0 causing I to decrease (Fig. 3b, left). In contrast, for a long ISI, ys is greater than y0 causing I to increase (Fig. 3b, right). The transient input, Is, triggered by S2 also resets us and vs, allowing the dynamics to evolve with the updated speed (Fig. 3b). By iterating this process after each sensory input, the SAM dynamically adjusts its output such that ys will eventually match y0 at the precise time of each stimulus input. These results are robust with respect to when the first stimulus is presented (750 ms in Fig. 3b and subsequent simulations; see Supplementary Fig. 4e for other delays).
Using simulations of the SAM, we analyzed the effect of K and I0 on the root-mean-squared-error (RMSE) between the predicted time (the time at which ys crosses y0) and the actual time of the stimuli (Fig. 3c; see “Methods” for simulation details). When I0 is large and K is small, the SAM generates large errors. Intermediate values of I0 and K lead to smaller errors, and a specific combination of (\({K}^{* },{I}_{0}^{* }\)) minimizes the RMSE (see Fig. 3c). Exploration of the parameter space revealed three key results. First, \({I}_{0}^{* }\) is largely independent of noise level and the number of stimuli (Fig. 3d, left). This is consistent with I0 serving as an initial estimate before any feedback is integrated. Second, K* decreases with increasing σn (Fig. 3d, right). This is expected because, when internal noise increases, the error signals generated by the SAM are less reliable and should be appropriately discounted. Third, as N increases, K* decreases, which reduces the weight given to each error and allows the model to integrate across inputs (Fig. 3d, right). Finally, we verified that the SAM updating process is robust with respect to the units’ initial conditions (u0 and v0) (Supplementary Fig. 4a–d), and can be generalized to different dynamical regimes (Supplementary Fig. 5). Therefore, the SAM provides a plausible mechanism for measuring time intervals predictively.
Sensorimotor updating by combining the SAM and MPM
So far, we found that the MPM can produce different IPIs depending on the level of input, and the SAM can adjust the input level based on the anticipated time of sensory events. Accordingly, we reasoned that the SAM and MPM might together be able to generate timed outputs that are in register with incoming sensory events: the SAM would adjust the input based on sensory events and the MPM would use that input to adjust IPI. We therefore connected the SAM to the MPM by having them share the input, I (Fig. 4a), and measured IPI as we changed the ISI randomly in a blocked fashion (Fig. 4b). The circuit was able to successfully track ISI throughout the run (Fig. 4c; see “Methods”). After each block transition, the SAM detects the error between IPI and ISI and adjusts I so that the MPM can gradually bring IPI closer to the new ISI (Fig. 4c). This mechanism allows IPIs to increase lawfully with the ISI (r2 = 0.53; p ≪ 0.01; Fig. 4d).
Full circuit model for synchronization
Coupling between the MPM and SAM allowed the circuit to adaptively match its output frequency (IPI) to the input frequency (ISI). However, it failed to match the input and output in terms of phase, as evidenced by the uniform distribution of phase differences between inputs and outputs (Fig. 4e; Rayleigh test of uniformity, p = 0.10). The source of this problem is that the circuit has no mechanism to determine whether the MPM output leads or lags the stimulus. Previous models have proposed highly nonlinear systems for phase adjustment26,27,28,29. We found that our model provides a simple solution to this problem. Since the output of the SAM (ys) and MPM (yp) are synchronized to sensory and motor events, respectively, the difference between them provides a graded signal reflecting their relative phase. Specifically, when ys > yp, the MPM is lagging and should be sped up, and when ys < yp, the MPM is leading and should be slowed down. To implement this solution, we augment the input I with a signal ΔI of the form
where α controls the learning rate. This adjustment can be realized by a unit which receives excitatory and inhibitory input form yp and ys, respectively (Fig. 5a, cyan). Fig. 5b demonstrates this correction scheme. Initially, yp slightly lags behind ys. This asynchrony generates a biphasic ΔI whose value is transiently positive (between the stimulus onset and motor output) and then negative until the next stimulus onset. The addition of this error signal reduces the total tonic drive to the MPM and increases the output speed relative to the SAM. After this adjustment is applied over several ISIs/IPIs, the MPM becomes increasingly synchronized with the stimulus. This strategy allows the circuit to gradually reduce asynchrony and narrow the distribution of phase errors (Rayleigh test of uniformity, p ≪ 0.01; Fig. 5c).
We examined the behavior of the model for different values of α and K. Larger values of α make the model more sensitive to phase errors and allow the model to cancel asynchronies more effectively (Fig. 5d, left panel). However, larger values of α cause rapid beat-by-beat adjustments in IPI, which can compromise rhythmicity (Fig. 5d, middle panel). As such, the optimal value of α depends on the relative cost of minimizing phase asynchrony and IPI error. In our simulations, values between 0.1 and 0.15 allow the model to minimize the sum squared error of phase and IPI (Fig. 5d, right panel). The parameter K, on the other hand, has little effect on how α influences phase asynchrony and IPI error (Fig. 5d, different colors).
This mechanism allowed the model to capture two counterintuitive observations in human behavior. First, as shown in Fig. 5c, the MPM had a persistent phase lead relative to the stimulus (−27.14° ± 71.45°) in a manner similar to humans performing analogous tasks30. Second, the interaction of α with the level of noise caused the model to occasionally skip a beat (Supplementary Fig. 6), which occurs occasionally in humans performing similar tasks31.
Model responses to perturbations are similar to humans
We compared the behavior of the model to that of humans in synchronization tasks during which the rhythmic input was perturbed in one of three different ways: a step change in ISI (Fig. 6a, top), a phase shift (Fig. 6b, top), and a jitter in the timing of a single event (Fig. 6c, top).
When facing a sudden change in ISI (Fig. 6a, top), response times have to be adjusted in both phase and frequency. For example, after an uncued increase in ISI, the first motor response would lead the stimulus. To regain synchrony, one has to delay the response to cancel the phase lead and additionally increase the IPI to match the new ISI. Human subjects concurrently reduce asynchrony and adjust IPI such that after a few samples, actions are in sync with sensory inputs30,32,33,34. A hallmark of this error-correcting strategy is that IPIs exhibit a transient overshoot relative to the new ISI (Fig. 6a, middle)35,36,37.
To test the behavior of the model in response to a change in ISI, we used the following protocol: we allowed the model to reach steady-state for an ISI of 800 ms, stepped the ISI to 1000 ms, and measured subsequent IPIs produced by the model. We repeated this procedure 1000 times and measured how the average IPI changed over time after the step change. Qualitatively, the model exhibits the transient IPI overshoot relative to ISI that is observed in humans, and gradually adjusts the IPI to match the new ISI (Fig. 6a, bottom). Quantitatively, the degree of overshoot depends on the model’s two learning rates, K for frequency (Fig. 6a, bottom), and α for phase (Supplementary Fig. 7).
Next, we analyzed the behavior of the model in response to a phase shift, in which a single ISI is increased or decreased, causing all subsequent stimuli to occur at a different phase without any change to the subsequent ISIs (Fig. 6b, top). In this case, humans adjust their response times so that the initial mismatch is gradually reduced (Fig. 6b, middle)38. This model was able to capture this response pattern, and the speed of recovery could be adjusted by α (Fig. 6b, bottom).
Finally, we considered a perturbation in which a single stimulus is jittered temporally. This perturbation alters two consecutive ISIs in equal and opposite directions without changing either the phase or the ISI of the subsequent stimuli (Fig. 6c, top). In response to the perturbation, human subjects exhibit a characteristic change in IPI relative to ISI. For example, when a single stimulus is delayed, subjects detect the error and delay their next response accordingly. However, since the perturbation is transient, subjects have to then undo their corrective response, which is done gradually over the course of the subsequent stimuli39,40 (Fig. 6c, middle). Again, the model was able to capture this response pattern, and the dynamics of the error correction was moderately dependent on α (Fig. 6c, bottom, circles).
These results demonstrate that the circuit model can capture key qualitative aspects of human behavior in response to various perturbations of stimulus timing. Behavioral studies have found individual differences in how subjects respond to these perturbations35,36,37,38,39,40. Accordingly, we verified that our model can capture this behavioral diversity through adjustments of model parameters (K and α) with an accuracy that is comparable to previously proposed algorithmic models30,34,36 (Supplementary Fig. 8).
Circuit model implements Bayesian interval reproduction
In the absence of any prior knowledge about the ISI of the first few beats, there is no way for the circuit (or a human) to choose an informed initial speed, and only sensory feedback can guide motor timing. However, when an observer has some prior expectation about the possible values of ISIs, the optimal strategy, as prescribed by Bayesian integration, is to integrate sensory inputs with that prior knowledge. This integration enables the system to reduce variability due to internal noise25,41 and to estimate the ISI with greater precision.
This Bayesian integration strategy is a hallmark of human behavior in time interval reproduction experiments18,42. In these experiments, subjects are typically provided with a sample interval, ts, drawn from a fixed prior distribution, and are asked to produce a matching interval, tp. The behavior of an optimal Bayesian observer performing this task exhibits two characteristic features. First, tp values are biased toward the mean of the ts distribution. Second, the magnitude of biases becomes smaller when measurements of ts are more reliable.
Recently, we tested humans in a time interval reproduction experiment in which ts was sampled from a fixed discrete uniform distribution ranging between 600 and 1000 ms31. The experiment involved two trial types. In one type (Fig. 7a, left), which we refer to as “1-2-Go,” subjects were presented with the first two beats of an isochronous rhythm (1-2) and were asked to produce the third omitted beat (Go). In the second type (Fig. 7a, right), referred to as “1-2-3-Go,” subjects were presented with the first three beats (1-2-3) and had to synchronize their response with the fourth omitted beat. Results provided clear evidence that subjects relied on their prior knowledge, since tp values were biased toward the mean of the ts distribution (Fig. 7b, purple). Moreover, the presentation of two ISIs in the 1-2-3-Go compared to one ISI in the 1-2-Go condition allowed subjects to measure ts more accurately and reduce the bias (Supplementary Fig. 9).
We simulated the model to test if it can emulate these characteristics. The SAM received input pulses representing the beats of an isochronous rhythm (2 pulses for the 1-2-Go task and 3 for the 1-2-3-Go task) with the ISI sampled from the same distribution that was used in the human experiment. We defined the production interval tp of the model as the interval between the final input pulse and the next time the SAM’s output exceeded y0.
The behavior of the model in this task depends on three key parameters: the initial input, I0, the update constant, K, and the level of noise in the model, σn. I0 determines the speed of dynamics before the first measurement of ts. This reflects the model’s initial guess about ts based on prior expectations. K determines the strength with which the SAM updates the input based on the difference between anticipated and observed ts. Finally, σn determines the variability of the model’s behavior due to internal noise. Here, we assume that the statistics of the environment are stationary (i.e., the prior distribution is fixed), an assumption consistent with the experimental design31. Under these conditions, a single set of parameters is required to optimize the performance of the model. Note that we do not explicitly model the learning process that optimizes model parameters for a specific prior distribution.
Together, the parameters I0, K, and σn allow the model to capture a range of behaviors observed in human subjects (see Supplementary Table 1). Results in Fig. 7b show the behavior of the model that was fit to the data from two human subjects in the 1-2-Go and 1-2-3-Go tasks (for parameter values, see Supplementary Table 1). Evidently, the model captured several key features of Bayesian integration present in the human behavior (Fig. 7b, magenta): (1) Average tp increased monotonically with ts (Fig. 7b, black circles), (2) noise in the model caused tp to vary on a trial-by-trial basis for the same ts (Fig. 7b, black error bars), (3) average tp was biased toward the mean of the prior distribution (Fig. 7b, deviation from the dashed unity line), and (4) the magnitude of bias was smaller in the 1-2-3-Go compared to the 1-2-Go trials (z = 2.7, p < 0.01, one-sided Wilcoxon signed rank test; Supplementary Fig. 9). An alternative implementation of this task using the full circuit architecture yields similar production times (Supplementary Fig. 10).
To further compare the model’s behavior to that of the human subjects, we partitioned the root-mean-square error (RMSE) between tp and ts to two terms, a BIAS term that measures the overall bias, and a VAR term that measures average variability across all values of ts (see “Methods”). The model was able to capture the observed BIAS and VAR across subjects in both 1-2-Go and 1-2-3-Go (Fig. 7c), indicating that it can correctly implement the bias-variance trade-off exhibited by humans during interval reproduction. In this model, the update constant, K, plays a central role in determining the bias-variance trade-off (compare Fig. 7b, top and bottom). For a Bayesian observer, the magnitude of bias is determined by noise in the measurements of ts and by the imposed cost function. Accordingly, the value of K in the model was adjusted to achieve a level of bias-variance trade-off that is inversely related with the inherent noise in the model (σn) and the operative cost function (r = − 0.94, p = 0.0002; see Supplementary Table 1). These results provide a potential mechanistic account of linear updating algorithms proposed previously to describe Bayesian timing behaviors43 (Supplementary Fig. 11).
Bayesian synchronization/continuation
As our final test, we examined the model in a Bayesian synchronization/continuation task that demands sensory anticipation, motor timing, and Bayesian integration. In this task, subjects are asked to tap synchronously with a metronome (synchronization phase) and continue to tap with the same tempo without the metronome until instructed to stop (continuation phase). In the classic version of this task, subjects have no uncertainty about the ISI of the metronome as it is kept constant across trials. Recently, humans were tested on a variant of this task in which ISI for each trial was sampled randomly from a prior distribution44. While performing this task, humans exhibit distinct patterns of biases in their IPIs. During synchronization, responses were biased toward the mean of the ISI distribution and the magnitude of the bias decreased with the number of synchronization stimuli (Fig. 8a, purple lines). These observations are consistent with the Bayesian integration scheme that we discussed previously in the context of a time interval reproduction task (Fig. 7). During the continuation phase, the bias persisted and its magnitude increased substantially (Fig. 8a, red lines) for all subjects (Fig. 8b). This increase in bias was due to a drop in the sensitivity of IPI to ISI (shallower slope relating IPI to ISI) as well as an overall bias toward shorter IPIs (Fig. 8c, left).
We fitted the model parameters (I0, K, and α) to individual subject’s behavior in this task (see “Methods”; Supplementary Table 2). Model fits generated IPIs that were biased towards the mean ISI. Consistent with subjects’ behavior, the magnitude of the bias decreased during synchronization and increased during continuation (Fig. 8a, black line). Moreover, like humans, the increase in bias during continuation was due to a combination of decreased sensitivity to the ISI (decrease in slope of IPI-ISI relationship, z = 2.1, p = 0.02, one-sided Wilcoxon signed rank test) and an overall shift towards shorter IPIs (decrease in IPI of the middle interval, z = 2.1, p = 0.02, one-sided Wilcoxon signed rank test; Fig. 8c, right; Supplementary Fig. 12).
In the model, this pattern of biases is explained by the augmented input, ΔI, to the MPM, which is necessary for the full model to synchronize its outputs with sensory inputs. Following the final stimulus input in the synchronization phase, the SAM no longer resets and instead proceeds toward the terminal fixed point. As a result, its output, ys, becomes fixed at a value greater than y0. However, yp continues to oscillate between values smaller than y0 because it is reset by the motor output. This mismatch makes ΔI negative and decreases the overall input to the MPM. The decreased input, in turn, speeds up the dynamics of the MPM and shortens IPIs during the continuation phase across all ISIs. Further, the impact of this negative ΔI is integrated over time, leading to an asymmetric impact on the speed of MPM dynamics associated with a short ISI compared to a long ISI. Specifically, longer ISIs lead to a greater increase in speed than shorter ISIs, accounting for the decreased sensitivity to the ISI during continuation. It is important to note that these results require that the SAM does not reset. Therefore, an alternate circuit configuration in which the output of the MPM resets the SAM will not be able to capture the increase in bias during continuation. This finding validates our assumption that resetting the SAM relies on sensory inputs and not motor outputs. Further, this feature of the circuit model allowed it to more accurately capture the observed pattern of bias than simple linear updating algorithms that have previously been used to model human synchronization behavior (Supplementary Fig. 13).
Discussion
To coordinate movements with external stimuli, the dynamics of neural activity must be adjusted based on sensory inputs7,15,45. However, precise control of dynamics is challenging because neural signals are subject to internal noise, and timing cues are often discrete and delayed21. Here, we proposed a simple neural circuit model that can address these challenges in a wide range of timing tasks.
The model is comprised of two modules, an MPM that generates timed actions, and a SAM that predicts the time of upcoming external events. Both modules were engineered to emulate a recently discovered speed control mechanism in the frontal cortex of monkeys during a flexible motor timing task7. The MPM uses this speed control mechanism in conjunction with a reset to produce rhythmic outputs. The SAM uses the same speed control mechanism in conjunction with an error-correction scheme that enables it to predict the timing of external events.
Coupling the MPM and SAM creates a closed-loop control system with versatile timing capacities in the presence of noise and delayed sensory feedback. For example, the model was able to capture human behavior in classic beat synchronization tasks30,32,33,34,37. In this case, the function of the SAM was to detect and correct discrepancies between the model’s output and external beats so that the MPM could adjust the outputs accordingly. The model was also able to emulate the biases associated with Bayesian time interval reproduction in the presence of noisy measurements when sample intervals were drawn from a fixed prior distribution18,31. To do so, the initial input to the model had to be adjusted based on the prior distribution, and the model’s updating parameters had to take the magnitude of the noise into account. Finally, a phase-correction mechanism between the SAM and MPM enabled the model to capture several non-trivial features of human behavior during a Bayesian synchronization/continuation task44 that previous algorithmic models could not.
Previous studies have proposed algorithmic models for how humans coordinate their motor outputs with sensory inputs. The basic algorithm is to compare the time of each motor output to the time of the corresponding sensory input, and use the error between the two to adjust the timing of the next movement31,32,33,34,43. This algorithm is conceptually similar to how our model functions, and its error-correcting behavior in beat synchronization tasks is comparable to our model30,43 (Supplementary Fig. 8). However, by virtue of implementing this algorithm using a neural circuit, our model’s predictions can be readily compared to neural data. Moreover, our model provides a mechanistic understanding of hypothesized predictive processes that algorithmic models have attributed to human behavior during perceptual timing tasks43,46,47.
Previous circuit models of timing largely fall into three classes, each with their own strengths and weaknesses. The first class is based on the accumulation of ticks of a central clock11,48,49,50. Similar to our circuit model, these clock-accumulator models rely on ramping activity that is observed in individual neurons12,13. However, these models do not explain how the recurrent circuit interactions lead to such ramping activity. The second class uses large recurrent neural circuits capable of producing rich dynamics2,7,51,52,53. These models can produce activity patterns that are strikingly similar to those observed in local populations of neurons, but it is unclear how they can flexibly integrate sensory and motor feedback. The third class uses a system of coupled oscillatory units26,27,28,54,55. These models can produce sophisticated timing behaviors and can integrate sensory information across a range of time scales26. However, the activity profile of neurons in the brain regions causally involved in timing is typically not oscillatory7. Our model provides an understanding of the link between these model classes. First, it explains the ramping activity in terms of recurrent dynamics due to interactions between neurons. Second, the model’s dynamics can be flexibly adjusted based on incoming sensory input and motor feedback. Third, the model generates oscillations at its output. However, unlike abstract models comprised of inherently oscillatory units, our model generates oscillations through circuit-level interactions that are consistent with the speed control mechanism observed in single neurons in the frontal cortex7.
Finally, the wiring of our circuit model might provide insight into the individual functional contributions of the cortical and subcortical systems important to action and perceptual timing56. One node, the basal ganglia, has been heavily implicated in action timing through lesion studies57 and physiological evidence7,45,58. An important principle of the basal ganglia function is the inhibition of downstream neural activity59. Given their proposed function as a competitive selection mechanism60, these inhibitory pathways may be the substrate for implementing the mutual inhibitory interactions needed for the temporal control of movements6. Another key node, the cerebellum, has also been linked to timing through lesion studies61, physiology62,63, and modeling44. A hallmark of cerebellar function is the detection and correction of sensory errors during sensorimotor control64,65, which is the key function of the SAM in our model. Finally, the output of both the basal ganglia and cerebellum are sent trans-thalamically to regions of the frontal cortex involved in sensory and motor timing66. Accordingly, we speculate that the integration of the sensory anticipation and motor production modules in our model may rely on interaction between the basal ganglia, cerebellum, and frontal cortical areas66.
Methods
Modeling and analyses were performed in Python 3.5.4 and Matlab R2017a.
BCM for interval production
The fundamental circuit architecture consists of three rate units, u, v, and y, which are governed by the following set of equations:
where τ, the time constant of each unit, was set to 100 ms and θ(x), the activation function of each unit, was specified by \(\theta (x)=1/[1+\exp (-x)]\), and I is a tonic input. WuI, Wuv, Wvu, and WvI specify the weighting of the interactions between units and each was set to 6 for all simulations. Similarly, we set Wyu = Wyv = 1. Noise in the system was modeled by 0 mean Gaussian white noise inputs, represented by the variables ηu, ηv, and ηy for unit u, v, and y, respectively. These variables were independently sampled at every time point from a Gaussian distribution with mean set to 0 and standard deviation specified by σn. The units were initialized at u = 0.7, v = 0.2, y = 0.5, and I = I0, where I0 is a free parameter. All simulations were carried out using Euler’s method with a step size of 10 ms.
Motor planning module
The MPM consists of four rate units I, up, vp, and yp, which are simulated based on the following equations:
Ip specifies a transient input to up and vp that serves to reset the system when yp > y0. We set Ip to 50 and y0 to 0.7 for all simulations. The activity level of I controls the speed of the dynamics of the MPM. The units were initialized at up = 0.7, vp = 0.2, yp = 0.5. Noise, represented by the variables \({\eta }_{{u}_{{\rm{p}}}}\), \({\eta }_{{v}_{{\rm{p}}}}\), and \({\eta }_{{y}_{{\rm{p}}}}\), was injected into each unit as in the BCM. All other parameters are the same as the BCM.
Sensory anticipation module
The SAM consists of four rate units I, us, vs, and ys. The system evolves according to the equations
where K is a free parameter and η indicates noise input to each unit as in the BCM. The module receives a binary input sIs, where s represents the visual stimulus (s = 1 when the stimulus is on and s = 0 when the stimulus is off) and Is is set to 50. This input serves to reset the values of us and vs at the time of each stimulus. All other parameters are the same as the BCM.
Sequences of stimuli were implemented by setting s = 1 periodically. Each stimulus presentation lasts 10 ms. The units us, vs, and ys were initialized as in the MPM and I was initialized at I0, a free parameter. The dynamics were allowed to run forward for 750 ms before the first stimulus was presented. This initialization period serves to put the circuit closer to steady-state so that the effect of reset signals is consistent across different ISIs. I does not update during the first presentation of the stimulus (by setting K = 0 during the first stimulus presentation).
Optimization of SAM parameters was achieved by simulating the module with 100 different pairs of I0 and K for each σn. K was uniformly sampled from values between 1 and 8.0, and I0 was sampled from values between 0.77 and 0.79. N stimuli were presented to the model, where N = 2, 3,…,10. ISIs were sampled from a discrete uniform distribution with five values between 600 and 1000 ms.
For each set (σn, I0, K, N), the model was run as described above and the RMSE was calculated according to
where Ns = 500, the total number of iterations across all ISIs. The pair (I0, K) that results in the smallest RMSE was recorded. The procedure was repeated 10 times for each σn to obtain a distribution of the optimal (I0, K).
In Fig. 3c, I0 and K were picked from a 30 × 30 grid spanning the the same range as the optimization procedure, and Ns = 5000.
Behavioral tasks
We used a suite of behavioral tasks to test circuit model timing performance. In all tasks, we performed numerical simulations of the dynamical equations expressed above using Euler’s method and a time step of 10 ms. All behavioral experiments were performed with the approval of the Committee on the Use of Humans as Experimental Subjects at MIT after receiving informed consent.
Periodic production task
To achieve periodic production, we simulated the MPM in isolation at four different levels of input, I, uniformly spaced from 0.75 to 0.78, for 40 s at each level. To measure the performance of the model, we calculated the IPI as
where tn was the time of the nth action produced by the model. For all levels of input, we set σn = 0.01. To compare the timing behavior of the model across levels of input, we calculated the mean and standard deviation of the first 40 IPIs generated by the circuit.
ISI tracking task
Circuit model synchronization was tested in a randomized ISI tracking task. The ISI was defined as
where mn was the time of the nth stimulus. On each trial of the task, we initialized the ISI at 800 ms and simulated a sensory input with 20 consecutive 800 ms intervals. After the first block of 20 ISIs, the ISI for the next 20 intervals was selected at random from a discrete uniform distribution between 600 and 900 ms with four possible values. This process was repeated four times, to generate five blocks of ISIs, corresponding to a sequence of 100 total intervals.
Circuit parameters were selected as follows: I0 was set to 0.771; K was selected to be 2; α was chosen to be 0 for simulations with no augmented input (e.g. Fig. 4) and 0.1 for simulations with augmented input (e.g. Fig. 5). I0 was set such that the MPM generated an 800 ms IPI on average, matching the initial ISI of every trial. This ensured that differences in circuit output resulted from updates to the value of I, rather than the value of I0 selected.
The nth IPI was quantified as in Eq. (19) and compared to the nth ISI. The asynchrony between an action and the stimulus presented at time mn was defined as
where tn was the time of the action which was closest in time to mn. The relative phase of the nth action, ϕn, was calculated as
where the phase is in units of degrees. For all simulations, we set σn = 0.01.
ISI perturbation task
The basis of the mechanisms behind human synchronization behavior are generally probed using perturbations of the ISI after an experimental subject has reached steady-state synchronization performance to a given ISI (see Repp30 for a review). To compare the circuit model synchronization performance to that of humans, we explored the behavior of the model in response to three common ISI perturbations: (1) a step change in the ISI, where the circuit synchronized to a stimulus with an ISI of 800 ms before the ISI was stepped to 1000 ms for all subsequent ISI. (2) A phase shift in the stimulus timing, where the circuit synchronized to a stimulus with a 500 ms ISI before the timing of stimuli was shifted by increasing a single ISI to 600 ms. All subsequent ISIs were 500 ms. (3) Stimulus jitter, where the circuit synchronized to an ISI of 500 ms before the timing of a single stimulus is perturbed, while subsequent stimuli remained in phase with the stimulus before the perturbation. To accomplish this, we perturbed two successive ISIs; the first was increased to 600 ms and second was decreased to 400 ms.
We simulated 1000 trials of each perturbation type and calculated the asynchrony associated with each action as in Eq. (21) and the mean IPI between actions as in Eq. (19). To ensure the circuit model was fully synchronized before a perturbation, we simulated 30 ISIs of the same duration before applying the perturbation. The circuit model was simulated with I0 = 0.771 and σn = 0.005. The values of K and α were varied to explore the behavior of the model with different sensitivities to errors in simulation and synchronization.
1-2-Go and 1-2-3-Go tasks
We compared the circuit model interval reproduction behavior to that of humans performing a timing task we refer to as 1-2-Go and 1-2-3-Go. The methods used for testing human interval reproduction are summarized briefly here. Please see our behavioral paper for a full description of the task and methods31.
The behavior of nine human subjects was analyzed. Subjects sat in a dark, quiet room at a distance of approximately 50 cm from a display monitor. The display monitor had a refresh rate of 60 Hz, a resolution of 1920 by 1200, and was controlled by a custom software (MWorks; http://mworks-project.org/) on an Apple Macintosh platform.
The interval reproduction task consisted of two randomly interleaved trial types “1-2-Go” and “1-2-3-Go”. On 1-2-Go trials, two flashes demarcated a sample interval (ts). On 1-2-3-Go trials, three flashes demarcated ts twice. For both trial types, subjects had to reproduce ts immediately after the last flash by pressing a button on a standard Apple keyboard. On all trials, subjects had to initiate their response proactively and without any additional cue. Subjects received graded feedback on their accuracy.
Each trial began with the presentation of a 0.5° circular fixation point at the center of a monitor display. The color of the fixation cued the trial type. After a 2 second delay, a warning stimulus and a trial cue were presented. The warning stimulus was a white circle that subtended 1.5° and was presented 10° to the left of the fixation point. The trial cue consisted of 2 or 3 small rectangles 0.6° above the fixation point (subtending 0.2° × 0.4°, 0.5° apart) for the 1-2-Go and 1-2-3-Go trials, respectively. After a random delay with a uniform hazard (250 ms minimum plus an interval drawn from an exponential distribution with mean of 500 ms), flashes demarcating ts were presented. Each flash lasted 6 frames (~100 ms) and was presented as an annulus around the fixation point with an inside and outside diameter of 2.5° and 3°, respectively. ts was sampled from a discrete uniform distribution ranging between 600 and 1000 ms with a 5 samples. To help subjects track the progression of events throughout the trial, after each flash, one rectangle from the trial cue would disappear (starting from the rightmost). The produced interval (tp) was measured as the interval between the time of the last flash and the time when the subject pressed the response key. We discarded any trial when the subject responded before the second (for 1-2-Go) or third (for 1-2-3-Go) flash and any trial where the response was 1000 ms after the veridical ts. We further used a model based approach to identify “lapse trials,” or trials in which tp was not related to ts31. We then calculated the mean tp conditioned on ts and trial type.
The SAM was run for 100 simulated trials for each ts (600 ms, 700 ms, 800 ms 900 ms, 1000 ms). To simulate the 1-2-Go task, the SAM was presented with 2 stimuli that were separated by the selected ts. The 1-2-3-Go task was simulated similarly but with 3 stimuli. The production interval tp of the model was defined as the interval between the final input pulse and the next time ys exceeded y0.
For each subject, we calculated the mean and standard deviation of tp for each ts in both the 1-2-Go and 1-2-3-Go tasks (10 task conditions in total). We call these values μ1,subject, STD1,subject, . . . , μ10,subject, STD10,subject (one pair of values for each condition).
For each set of parameters (σn, I0, K), the model was run as described above. We calculated the mean and standard deviation of the model’s tp for each ts, and for both the 1-2-Go and 1-2-3-Go tasks. We call these values μ1,model, STD1,model, . . . , μ10,model, STD10,model.
The fitting was done by alternating between fitting σn and jointly fitting (I0, K). The parameters were uniformly and independently sampled from the following intervals: [1.0, 8.0] for K, [0.77, 0.79] for I0 and [0.005, 0.4] for σn. For each step, 100 sets of parameters were sampled and the optimal set was used to update the parameters.
To evaluate the anticipated ts associated with I0 and σn, we initialized the model with I0, presented a single stimulus to the model 750 ms after initialization, and determined the interval between this stimulus and the next time ys exceeded y0.
We defined BIAS and VAR for each subject and model simulation as
where N = 5 is the number of target intervals, \({\overline{t}}_{{{\rm{p}}}_{i}}\) and \({\sigma }_{i}^{2}\) are the mean and variance of production intervals that correspond to the target interval \({t}_{{{\rm{s}}}_{i}}\).
Synchronization/continuation task
We compared the circuit model to the behavior of humans performing a synchronization/continuation task in which the inter-stimulus-interval (ISI) was selected at random each trial from a set distribution. The methods used for testing human experiments are summarized briefly here. Please see Narain et al. 44 for a full description of the task and methods44.
We analyzed the behavior of six human participants. Stimuli were viewed from a distance of approximately 67 cm in a dark, quiet room. The display monitor had a refresh rate of 60 Hz, a resolution of 1920 by 1200, and was controlled by a custom software (MWorks; http://mworks-project.org/) on an Apple Macintosh platform.
Each trial began with the presentation of a red circular fixation stimulus (diameter = 0.75° visual angle) in the center of the screen. After a variable delay (200 ms plus an additional amount of time which was exponentially distributed with mean = 300 ms and a maximum value of 2300 ms), a synchronization stimulus was flashed four times with an ISI chosen from a discrete uniform distribution (five intervals, minimum = 550 ms, maximum = 817 ms). The flashing stimulus was a gray annulus centered around fixation with an inner diameter of 1° and outer diameter of 1.25°. Participants were instructed to tap a response key synchronously with the third and fourth synchronization flashes and continue tapping to reflect the ISI until instructed to stop. The number of continuation taps required was three plus an exponentially distributed number with a mean of nine and a maximum of 22. The first IPI was defined as the interval between the middle of the second flash and the first key press. Subsequent IPIs were defined as the interval between successive key presses.
Biases in IPIs were calculated according to the following procedure. For each interval, ISIi, we found the mean IPI for the kth interval in the sequence \(({\overline{{\rm{IPI}}}}_{i,k})\). The mean squared bias for each k was defined as
where N = 5 is the number of distinct ISI values.
We simulated 21 trials of the full synchronization model for each ISI. In each trial, the stimuli presented to the model were a series of three flashes separated by the ISI of that trial. After the three flashes, the stimulus was set to s = 0 and the model was run and allowed to produce 17 more productions. The IPI of the circuit was defined as the time difference between successive productions.
We fixed σn at 0.01 and varied three free parameters K, I0, and α. For each subject, we simulated 100 different parameter combinations with K randomly selected from the interval [0.01, 5], I0 from the interval [0.76, 0.78] and α from the interval [0.01, 0.1]. For each combination of parameters, the model was run as described above and bias was calculated as in Eq. (27). We then found the combination of parameters that minimized the mean squared errors between the BIASk observed in the subject and the BIASk of the circuit, across k = 1, …, 20.
Synchronization bias was quantified as the bias for k = 3, and continuation bias was the bias for k = 7.
Linear behavioral algorithm for sensorimotor timing
For comparison purposes, we also developed a linear updating model for synchronization based on previous work30. In the absence of any error, the model computes the (n + 1)th production time, tn+1, by adding the expected ISI, Tn, to the nth production time, tn. When there is a discrepancy, the updating rule incorporates two additional error terms. The first error term measures the asynchrony, which is the difference between tn and the time of the nth metronomic input, mn. The second error term measures the difference between Tn and the observed ISI, ISIn. In the model, these two error terms are weighted by βAsynch and βISI, respectively.
We simulated this timing model for different values of βAsynch and βISI to compare this standard algorithm to the behavior of the circuit model.
In simulations of the 1-2-Go and 1-2-3-Go tasks, we added a noise term, η, to tn+1. η was sampled independently at every time point from a Gaussian distribution with mean 0 and standard deviation σn. For fitting, βAsynch was set to 0 and parameters were uniformly and independently sampled from the following intervals: [0.0, 1.0] for βISI, [700, 900] for T0, and [30, 120] for σn. The fitting procedure is the same as described for the circuit model.
In simulations of the synchronization/continuation task, the production times tn during the synchronization phase were simulated like the 1-2-3-Go task. For the continuation phase, tn+1 = tn + TN + η, where η was sampled independently at every time point from a Gaussian distribution with mean 0 and standard deviation σn, and TN is the value of T at the final stimulus presentation. For fitting, σn was set to 50, and parameters were uniformly and independently sampled from the following intervals: [600, 1200] for T0, [0, 2] for βISI, [0, 1] for βAsynch. The fitting procedure is the same as described for the circuit model.
Reporting summary
Further information on research design is available in the Nature Research Reporting Summary linked to this article.
Data availability
No new data were collected. Data for previously published work is available at https://jazlab.org/resources/.
Code availability
Code used in this study is publicly available at https://jazlab.org/resources/.
References
Remington, E. D., Egger, S. W., Narain, D., Wang, J. & Jazayeri, M. A dynamical systems perspective on flexible motor timing. Trends Cogn. Sci.22, 938–952 (2018).
Remington, E. D., Narain, D., Hosseini, E. A. & Jazayeri, M. Flexible sensorimotor computations through rapid reconfiguration of cortical dynamics. Neuron98, 1005–1019.e5 (2018).
Churchland, M. M. et al. Neural population dynamics during reaching. Nature487, 51–56 (2012).
Fetz, E. E. Are movement parameters recognizably coded in the activity of single neurons? Behav. Brain Sci.15, 679–690 (1992).
Shenoy, K. V., Sahani, M. & Churchland, M. M. Cortical control of arm movements: a dynamical systems perspective. Annu. Rev. Neurosci.36, 337–359 (2013).
Murray, J. M. & Escola, G. S. Learning multiple variable-speed sequences in striatum via cortical tutoring. Elife6, e26084 (2017).
Wang, J., Narain, D., Hosseini, E. A. & Jazayeri, M. Flexible timing by temporal scaling of cortical responses. Nat. Neurosci.21, 102–110 (2018).
Sohn, H., Narain, D., Meirhaeghe, N. & Jazayeri, M. Bayesian computation through cortical latent dynamics. Neuron103, 934–947.e5 (2019).
Sauerbrei, B. A. et al. Cortical pattern generation during dexterous movement is input-driven. Nature577, 386–391 (2020).
Hardy, N. F., Goudar, V., Romero-Sosa, J. L. & Buonomano, D. V. A model of temporal scaling correctly predicts that motor timing improves with speed. Nat. Commun.9, 4732 (2018).
Killeen, P. R. & Fetterman, J. G. A behavioral theory of timing. Psychol. Rev.95, 274–295 (1988).
Tanji, J. & Evarts, E. V. Anticipatory activity of motor cortex neurons in relation to direction of an intended movement. J. Neurophysiol.39, 1062–1068 (1976).
Schultz, W. & Romo, R. Role of primate basal ganglia and frontal cortex in the internal generation of movements i. preparatory activity in the anterior striatum. Exp. Brain Res.91, 363–384 (1992).
Riehle, A. & Requin, J. The predictive value for performance speed of preparatory changes in neuronal activity of the monkey motor and premotor cortex. Behav. Brain Res.53, 35–49 (1993).
Mita, A., Mushiake, H., Shima, K., Matsuzaka, Y. & Tanji, J. Interval time coding by neurons in the presupplementary and supplementary motor areas. Nat. Neurosci.12, 502–507 (2009).
Murakami, M., Vicente, M. I., Costa, G. M. & Mainen, Z. F. Neural antecedents of self-initiated actions in secondary motor cortex. Nat. Neurosci.17, 1574–1582 (2014).
Gallistel, C. R. & Gibbon, J. Time, rate, and conditioning. Psychol. Rev.107, 289–344 (2000).
Jazayeri, M. & Shadlen, M. N. Temporal context calibrates interval timing. Nat. Neurosci.13, 1020–1026 (2010).
Miall, R. C., Weir, D. J., Wolpert, D. M. & Stein, J. F. Is the cerebellum a smith predictor? J. Mot. Behav.25, 203–216 (1993).
Egger, S. W., Remington, E. D., Chang, C.-J. & Jazayeri, M. Internal models of sensorimotor integration regulate cortical dynamics. Nat. Neurosci.22, 1871–1882 (2019).
Todorov, E. Optimality principles in sensorimotor control. Nat. Neurosci.7, 907–915 (2004).
Scott, S. H. Optimal feedback control and the neural basis of volitional motor control. Nat. Rev. Neurosci.5, 532–546 (2004).
Sommer, M. A. & Wurtz, R. H. What the brain stem tells the frontal cortex: I. oculomotor signals sent from superior colliculus to frontal eye field via mediodorsal thalamus. J. Neurophysiol.91, 1381–1402 (2004).
Sommer, M. A. & Wurtz, R. H. What the brain stem tells the frontal cortex. II. role of the SC-MD-FEF pathway in corollary discharge. J. Neurophysiol.91, 1403–1423 (2004).
Ivry, R. B. & Hazeltine, R. E. Perception and production of temporal intervals across a range of durations: evidence for a common timing mechanism. J. Exp. Psychol. Hum. Percept. Perform.21, 3–18 (1995).
Large, E. W. & Kolen, J. F. Resonance and the perception of musical meter. Conn. Sci.6, 177–208 (1994).
Large, E. W. & Jones, M. R. The dynamics of attending: How people track time-varying events. Psychol. Rev.106, 119–159 (1999).
Todd, N. P. M., Lee, C. S. & O’Boyle, D. J. A sensorimotor theory of temporal tracking and beat induction. Psychol. Res.66, 26–39 (2002).
Bose, A., Byrne, Á. & Rinzel, J. A neuromechanistic model for rhythmic beat generation. PLoS Comput. Biol.15, e1006450 (2019).
Repp, B. H. Sensorimotor synchronization: a review of the tapping literature. Psychon. Bull. Rev.12, 969–992 (2005).
Egger, S. W. & Jazayeri, M. A nonlinear updating algorithm captures suboptimal inference in the presence of signal-dependent noise. Sci. Rep.8, 12597 (2018).
Mates, J. A model of synchronization of motor acts to a stimulus sequence I. Timing and error corrections. Biol. Cybern.70, 463–473 (1994).
Semjen, A., Vorberg, D. & Schulze, H.-H. Getting synchronized with the metronome: comparisons between phase and period correction. Psychol. Res.61, 44–55 (1998).
Pressing, J. Error correction processes in temporal pattern production. J. Math. Psychol.42, 63–101 (1998).
Michon, J. A. Timing in Temporal Tracking (Institute for Perception RVO-TNO Soesterberg, The Netherlands, 1967).
Thaut, M. H., Miller, R. A. & Schauer, L. M. Multiple synchronization strategies in rhythmic sensorimotor tasks: phase vs period correction. Biol. Cybern.79, 241–250 (1998).
Bavassi, M. L., Tagliazucchi, E. & Laje, R. Small perturbations in a finger-tapping task reveal inherent nonlinearities of the underlying error correction mechanism. Hum. Mov. Sci.32, 21–47 (2013).
Repp, B. H. Compensation for subliminal timing perturbations in perceptual-motor synchronization. Psychol. Res.63, 106–128 (2000).
Repp, B. H. Automaticity and voluntary control of phase correction following event onset shifts in sensorimotor synchronization. J. Exp. Psychol. Hum. Percept. Perform.28, 410–430 (2002).
Repp, B. H. Phase correction in sensorimotor synchronization: nonlinearities in voluntary and involuntary responses to perturbations. Hum. Mov. Sci.21, 1–37 (2002).
Getty, D. J. Discrimination of short temporal intervals: a comparison of two models. Percept. Psychophys.18, 1–8 (1975).
Miyazaki, M., Nozaki, D. & Nakajima, Y. Testing bayesian models of human coincidence timing. J. Neurophysiol.94, 395–399 (2005).
Shi, Z. & Burr, D. Predictive coding of multisensory timing. Curr. Opin. Behav. Sci.8, 200–206 (2016).
Narain, D., Remington, E. D., De Zeeuw, C. I. & Jazayeri, M. A cerebellar mechanism for learning prior distributions of time intervals. Nat. Commun.9, 469 (2018).
Mello, G. B. M., Soares, S. & Paton, J. J. A scalable population code for time in the striatum. Curr. Biol.25, 1113–1122 (2015).
McAuley, J. D. & Jones, M. R. Modeling effects of rhythmic context on perceived duration: a comparison of interval and entrainment approaches to short-interval timing. J. Exp. Psychol. Hum. Percept. Perform.29, 1102–1125 (2003).
Di Luca, M. & Rhodes, D. Optimal perceived timing: integrating sensory information with dynamically updated expectations. Sci. Rep.6, 28563 (2016).
Treisman, M. Temporal discrimination and the indifference interval: Implications for a model of the "internal clock”. Psychological Monogr.: Gen. Appl.77, 1–31 (1963).
Meck, W. H. Neuropharmacology of timing and time perception. Brain Res. Cogn. Brain Res.3, 227–242 (1996).
Simen, P., Balci, F., de Souza, L., Cohen, J. D. & Holmes, P. A model of interval timing by neural integration. J. Neurosci.31, 9238–9253 (2011).
Buonomano, D. V. Timing of neural responses in cortical organotypic slices. Proc. Natl Acad. Sci. U. S. A.100, 4897–4902 (2003).
Karmarkar, U. R. & Buonomano, D. V. Timing in the absence of clocks: encoding time in neural network states. Neuron53, 427–438 (2007).
DePasquale, B., Cueva, C. J., Rajan, K., Escola, G. S. & Abbott, L. F. full-FORCE: a target-based method for training recurrent networks. PLoS ONE13, e0191527 (2018).
Miall, C. The storage of time intervals using oscillating neurons. Neural Comput.1, 359–371 (1989).
Large, E. W., Herrera, J. A. & Velasco, M. J. Neural networks for beat perception in musical rhythm. Front. Syst. Neurosci.9, 159 (2015).
Merchant, H., Harrington, D. L. & Meck, W. H. Neural basis of the perception and estimation of time. Annu. Rev. Neurosci.36, 313–336 (2013).
Malapani, C. et al. Coupled temporal memories in parkinson’s disease: a dopamine-related dysfunction. J. Cogn. Neurosci.10, 316–331 (1998).
Rao, S. M., Mayer, A. R. & Harrington, D. L. The evolution of brain activation during temporal processing. Nat. Neurosci.4, 317–323 (2001).
Hikosaka, O. & Wurtz, R. H. Visual and oculomotor functions of monkey substantia nigra pars reticulata. IV. relation of substantia nigra to superior colliculus. J. Neurophysiol.49, 1285–1301 (1983).
Frank, M. J. Computational models of motivated action selection in corticostriatal circuits. Curr. Opin. Neurobiol.21, 381–386 (2011).
Ivry, R. B. & Keele, S. W. Timing functions of the cerebellum. J. Cogn. Neurosci.1, 136–152 (1989).
Halverson, H. E., Khilkevich, A. & Mauk, M. D. Cerebellar processing common to delay and trace eyelid conditioning. J. Neurosci.38, 7221–7236 (2018).
Ohmae, S., Uematsu, A. & Tanaka, M. Temporally specific sensory signals for the detection of stimulus omission in the primate deep cerebellar nuclei. J. Neurosci.33, 15432–15441 (2013).
Herzfeld, D. J., Kojima, Y., Soetedjo, R. & Shadmehr, R. Encoding of action by the purkinje cells of the cerebellum. Nature526, 439–442 (2015).
Herzfeld, D. J., Kojima, Y., Soetedjo, R. & Shadmehr, R. Encoding of error and learning to correct that error by the purkinje cells of the cerebellum. Nat. Neurosci.21, 736–743 (2018).
Strick, P. L., Dum, R. P. & Fiez, J. A. Cerebellum and nonmotor function. Annu. Rev. Neurosci.32, 413–434 (2009).
Repp, B. H. Phase correction following a perturbation in sensorimotor synchronization depends on sensory information. J. Mot. Behav.34, 291–298 (2002).
Acknowledgements
The authors thank Jing Wang, Morteza Saraf, Nicolas Meirhaeghe, and Alexandra Ferguson for their insightful comments on this manuscript and Michal De-Medonsa for her editorial assistance. M.J. is supported by NIH (NINDS-NS078127), the Klingenstein Foundation, the Simons Foundation, the McKnight Foundation, and the McGovern Institute.
Author information
Authors and Affiliations
Contributions
S.W.E. conceived of the extension of the basic circuit module. S.W.E. developed, implemented and characterized the motor planning module. N.M.L. developed, implemented and characterized the sensory anticipation module. S.W.E. developed the circuit for the synchronization task and characterized model performance. N.M.L. compared model performance to humans in the 1-2-Go, 1-2-3-Go, and synchronization/continuation tasks. M.J. supervised the project. All authors contributed to the interpretation of results and writing of the manuscript.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Peer review informationNature Communications thanks Sean Escola and the other, anonymous, reviewer(s) for their contribution to the peer review of this work. Peer reviewer reports are available.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Egger, S.W., Le, N.M. & Jazayeri, M. A neural circuit model for human sensorimotor timing. Nat Commun 11, 3933 (2020). https://doi.org/10.1038/s41467-020-16999-8
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41467-020-16999-8
This article is cited by
-
Functional architecture of executive control and associated event-related potentials in macaques
Nature Communications (2022)
-
A biophysical counting mechanism for keeping time
Biological Cybernetics (2022)
-
Two views on the cognitive brain
Nature Reviews Neuroscience (2021)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.