Abstract
Multiple learning systems allow individuals to flexibly respond to opportunities and challenges present in the environment. An evolutionarily conserved “Pavlovian” learning mechanism couples valence and action, promoting a tendency to approach cues associated with reward and to inhibit action in the face of anticipated punishment. Although this default response system may be adaptive, these hard-wired reactions can hinder the ability to learn flexible “instrumental” actions in pursuit of a goal. Such constraints on behavioral flexibility have been studied extensively in adults. However, the extent to which these valence-specific response tendencies bias instrumental learning across development remains poorly characterized. Here, we show that while Pavlovian response biases constrain flexible action learning in children and adults, these biases are attenuated in adolescents. This adolescent-specific reduction in Pavlovian bias may promote unbiased exploration of approach and avoidance responses, facilitating the discovery of rewarding behavior in the many novel contexts that adolescents encounter.
Similar content being viewed by others
Introduction
From an early age, individuals can rely on distinct forms of learning to maximize rewards and avoid punishments in their environments. Through Pavlovian learning, a neutral cue that predicts reward or punishment acquires the positive or negative value associated with its outcome1. This learned association between a cue and its predicted outcome can then drive reflexive behavioral responses that are elicited in a valence-dependent manner. Expectations of reward typically lead to approach behaviors2,3, whereas expectations of punishment tend to foster the inhibition of action4. For instance, by learning that a sweet smell often precedes enjoying a scrumptious dessert, you might be drawn toward the fragrant scent in anticipation of an afternoon treat. Or by learning that the sound of a horn often is heard before a car accident, you might freeze upon hearing the noise in anticipation of a negative event. This evolutionarily conserved learning mechanism allows approach or action inhibition to be readily deployed as “default” reactions to anticipated rewards or threats, respectively, without any need to evaluate the efficacy of these responses through experience5. Whereas Pavlovian learning couples expectations of reward or punishment with reflexive reactions that have no causal influence on the outcomes that actually occur, instrumental learning enables the discovery and deployment of actions that can directly influence the likelihood of obtaining reward or avoiding punishment. By learning which actions bring about beneficial outcomes, an individual can flexibly engage active or inactive responses based on their causal efficacy, rather than emitting reactive responses that are controlled by cues in their environment.
Pavlovian and instrumental learning systems do not operate in isolation6. When active approach responses are required to secure reward, or one must inhibit action to avoid punishment, the Pavlovian reactions aligned with these instrumental contingencies can facilitate learning. However, when Pavlovian tendencies conflict with instrumental contingencies, these default reactions can impede the learning of adaptive instrumental actions. For example, freezing at the sound of a car horn might be beneficial if you are about to step into oncoming traffic, but might be harmful if it prevents you from running out of the way if you are already in the road. Studies across species highlight the difficulty of learning to perform instrumental actions that conflict with Pavlovian expectancies, as the evolutionarily conserved tendencies to approach cues predictive of reward or to inhibit action to cues predictive of punishment can be too robust to overcome7,8,9,10,11,12. In active avoidance paradigms, rodents have difficulty learning to cross to the opposite side of a conditioning chamber in order to avoid being shocked, due to their tendency to freeze following a threat-predictive cue13,14,15. Similarly, in the appetitive domain, rats have difficulty learning to earn a food reward by not approaching a cue that predicts the reward16. These studies demonstrate that Pavlovian reactions to anticipated reward or punishment can pose constraints on instrumental action, disrupting flexible goal-directed learning.
To date, the biasing effects of Pavlovian learning on behavior have been studied primarily in adult rodents and humans10,17,18,19,20,21,22,23. While both Pavlovian and instrumental learning are evident from early childhood24,25,26,27, relatively few studies have investigated the interaction between Pavlovian and instrumental learning across development in humans28. Thus, it is unclear when Pavlovian constraints on action learning emerge and how they change from childhood to adulthood.
In the present study, we tested the extent to which Pavlovian reactions differentially biased instrumental learning across development. We hypothesized that these Pavlovian learning constraints would decrease from childhood to adulthood, as the cognitive capabilities and neural circuits that support goal-directed instrumental learning are refined25,29,30,31,32. To test this hypothesis, we had 61 participants, 8–25 years of age (20 children aged 8–12, 20 adolescents aged 13–17, and 21 adults aged 18–25; see methods for detailed demographics of participants), complete a probabilistic Go/No-Go task in which valence and action were orthogonalized. We adapted a well-validated paradigm20,21,33 for use in a developmental cohort by using a child-friendly narrative to frame the task. The goal of the task was to earn as many “tickets” as possible by interacting with four different colored robots. Valence (i.e., the potential to either win or lose a ticket) and action (i.e., the need to either press or not press a button) were orthogonalized across the four robots. Two of the robots were associated with the potential to win a ticket (“ticket givers”), and two robots were associated with the potential loss of a ticket (“ticket takers;” Fig. 1). For both “ticket givers” and “ticket takers,” the correct response for one of the robots was to press a button, whereas the correct response for the other robot was to withhold a button press. Participants were instructed that robots could be either “ticket givers” or “ticket takers” and that the correct action for each robot could be learned through feedback. For the “ticket givers,” a correct response resulted in winning a ticket 80% of the time but no ticket (the “null” outcome) 20% of the time; whereas for the “ticket takers,” a correct response avoided the loss of a ticket (the “null” outcome) 80% of the time but resulted in the loss of a ticket 20% of the time. Incorrect responses yielded feedback with reversed outcome probabilities.
The Pavlovian default responses to approach expected reward or to inhibit action in the face of potential punishment were aligned with the correct instrumental response on “Go to Win” and “No-Go to Avoid Losing” trials, whereas the default responses were in conflict with the optimal instrumental actions on “Go to Avoid Losing” and “No-Go to Win” trials. Critically, a Pavlovian bias was evident if performance was better for the robots for which Pavlovian tendencies and correct instrumental responses were congruent than those for which they were incongruent. If Pavlovian learning did not bias instrumental action learning, then performance would be comparable across all four trial types.
Results
To characterize patterns of age-related change in task performance, we assessed through model comparison whether each measure of choice behavior was best captured by a statistical model that included age alone (i.e., a linear model) or a model that included an additional nonlinear age-squared term, as in previous studies34,35,36. The best-fitting model thus indicates whether the developmental trajectory of choice behavior exhibits a continuous linear progression from childhood to adulthood, or whether behavior shows either adolescent-specific effects (i.e., children’s and adults’ behavior are more similar to each other than to adolescents’ behavior) or adolescent-emergent effects (i.e., behavior of adolescents is more similar to adults’ behavior than to children’s). Age and age-squared were included as continuous variables in all analyses. However, age is represented categorically (i.e., grouping children, adolescents, and adults) in some figures for the purpose of depicting results, and we use categorical age terminology in our interpretation and discussion of the findings.
Behavioral analyses
First, we tested whether there was a relationship between the number of tickets won and participant age. A linear model that included both age and age-squared as predictors of number of tickets won significantly improved the fit compared to a model that included age alone (F(1,58) = 15.959, p = 0.0002). We found significant effects of age (β value = 7.523, s.e. = 1.804, t(58) = 4.170, p < 0.001, Cohen’s ƒ2 = 0.300) and age-squared (β value = −7.334, s.e. = 1.836, t(58) = -3.995, p < 0.001, Cohen’s ƒ2 = 0.275) on the number of tickets won, indicating a peak in overall task performance during late adolescence (Fig. 2a).
We next examined how participants’ choice behavior gave rise to this nonlinear age-related change in task performance. The qualitative pattern of performance on each trial type is depicted for children, adolescents, and adults in Fig. 2b (and Supplementary Fig. S1). To test how performance for each trial type differs as a function of age, we performed four separate linear regressions in which age and age-squared predict accuracy. The model that included age alone provided the best fit for accuracy on “Go to Win” (GW) trials, although age was not a significant predictor (F(1,58) = 0.557, p = 0.458; age: β value = 0.013, s.e. = 0.020, t(58) = 0.647, p = 0.52, Cohen’s ƒ2 = 0.007. For all other trial types (i.e., Go to Avoid Losing (GAL), No-Go to Win (NGW), No-Go to Avoid Losing (NGAL)), the model that included an age-squared term provided a better fit than the model with age alone (GAL: F(1,58) = 9.937, p = 0.003; NGW: F(1,58) = 8.828, p = 0.004; NGAL: F(1,58) = 10.004, p = 0.002). For these three trial types, age and age-squared were significant predictors of accuracy (GAL age: β value = 0.089, s.e. = 0.020, t(58) = 4.447, p < 0.001, Cohen’s ƒ2 = 0.341; GAL age-squared: β value = − 0.064, s.e. = 0.020, t(58) = − 3.152, p = 0.003, Cohen’s ƒ2 = 0.171; NGW age: β value = 0.084, s.e. = 0.038, t(58) = 2.204, p = 0.032, Cohen’s ƒ2 = 0.084; NGW age-squared: β value = − 0.116, s.e. = 0.039, t(58) = − 2.971, p = 0.004, Cohen’s ƒ2 = 0.152; NGAL age: β value = 0.080, s.e. = 0.018, t(58) = 4.402, p < 0.001, Cohen’s ƒ2 = 0.334; NGAL age-squared: β value = − 0.058, s.e. = 0.018, t(58) = − 3.163, p = 0.002, Cohen’s ƒ2 = 0.172). Thus, apart from GW for which accuracy was the highest and was comparable across age, performance on the other three trial types exhibited significant nonlinear improvements from childhood into adulthood.
To quantify the Pavlovian influence on instrumental learning for each individual, we first calculated a Pavlovian performance bias score by averaging how often reward-related cues invigorated action (number of Go responses to Win cues/total number of Go responses) and how often punishment-related cues suppressed action (number of No-Go responses to Avoid Losing cues/total number of No-Go responses). A bias score of 0.5 indicates the absence of a Pavlovian bias; whereas higher scores reflect a greater Pavlovian bias on action, with 1 being the maximum bias score. The best-fitting linear model predicting age-related changes in this bias score included both age and age-squared as regressors, of which age-squared was a significant predictor of Pavlovian performance bias (F(1,58) = 4.136, p = 0.047; age: β value = −0.018, s.e. = 0.014, t(58) = − 1.271, p = 0.209, Cohen’s ƒ2 = 0.028; age-squared: β value = 0.03, s.e. = 0.015, t(58) = 2.034, p = 0.047, Cohen’s ƒ2 = 0.071). This analysis revealed that children and adults exhibited a greater Pavlovian bias than adolescents (Fig. 3). This bias appears to be driven comparably by both a reward-driven invigoration of action and a punishment-driven suppression of action (Supplementary Figure S2). A mixed-effects logistic regression corroborated this adolescent-specific attenuation of Pavlovian bias on learning (Supplementary Table S1).
Computational modeling
By formalizing the value computations involved in learning from valenced outcomes, computational models can disentangle the contribution of a Pavlovian bias on choice behavior from differences in value updating, choice stochasticity, a value-independent bias toward action, or sensitivity to reward. We used reinforcement-learning models to dissociate the component processes of learning in the task and test whether the age-related changes observed in choice behavior could be attributed specifically to Pavlovian biases on action-value computation over the course of the learning task. We compared a set of nested reinforcement-learning models that were fit to participants’ data to determine which model best explained choice behavior across all participants. The models were chosen based on prior work in adults using a variant of this task (see Supplementary Table S2 for full list of models tested)21,33. Median Akaike Information Criterion (AIC) values, for which lower values reflect a better fit, were used to determine which model provided the best explanation of behavior37.
The model that best fit choice behavior included a learning rate, lapse rate, Go bias, Pavlovian bias, and a single reinforcement sensitivity term (Supplementary Table S2; see Methods for details on the other models tested). This model estimated an action-value (Q value or Q(a,s)), for each potential action (\(a\); i.e., Go or No-Go) for each of the four robot stimuli (s). These Q value estimates were updated on every trial (t) using an error-driven learning function (Eq. 1). Outcomes (either 1, 0, -1) were included in the model through the reward term (r) and multiplied by a reinforcement sensitivity parameter (ρ) that scaled the effective size of rewards and punishments, with higher values magnifying differences between Q values. The difference between the effective reward on that trial (ρrt) and the previous Q value estimate (Qt-1(\(a\)t,st)) indexes the reward prediction error, indicating whether the effective outcome was better or worse than expected. The Q value estimate was incrementally updated following each outcome by adding the reward prediction error scaled by a learning rate (α).
Two weights were introduced that altered this action value estimate (weighted action estimate (Wt(\(a\)t,st))). The first was a Go bias (b), which captured an increased tendency to press the button (Eq. 2).
The second was a Pavlovian bias term (π) that was multiplied by the stimulus value estimate (Vt(s)) for each robot (Eq. 3). The Pavlovian bias parameter indexed the degree to which action was facilitated for cues associated with reward and inhibited for those associated with punishment. The stimulus value estimate was updated on each trial through reinforcement, in a similar manner to the Q value estimate (Eq. 4).
Action values from the model were transformed into choice probabilities using a squashed softmax choice function38 that included a lapse rate (ξ), which captures the effects of inattention (Eq. 5).
We then investigated patterns of age-related change for each free parameter in the model by performing linear regressions including age alone or both age and age-squared as predictors. Lapse rate, learning rate, and reinforcement sensitivity were best fit by linear regression models that included age alone (Table 1; Supplementary Fig. S3; lapse rate: F = 3.192, p = 0.079; learning rate: F = 0.121, p = 0.729; reinforcement sensitivity: F = 1.057, p = 0.308). However, age was not a significant predictor of lapse rate or learning rate, indicating no evidence of age-related changes in inattention or updating the value of an action (lapse rate: β value = − 0.027, s.e. = 0.026, t(58) = − 1.02, p = 0.312, Cohen’s ƒ2 = 0.018; learning rate: β value = 0.056, s.e. = 0.033, t(58) = 1.672, p = 0.1, Cohen’s ƒ2 = 0.047). There was a significant effect of age on reinforcement sensitivity, revealing greater effective reinforcement with age (β value = 0.935, s.e. = 0.396, t(58) = 2.362, p = 0.022, Cohen’s ƒ2 = 0.095). The inclusion of age-squared in the model significantly improved the fit for the Go and Pavlovian bias parameters (Go bias: F = 4.447, p = 0.039; Pavlovian bias: F = 11.916, p = 0.001). Age-squared, but not linear age, was a significant predictor of the Go bias (age: β value = − 0.039, s.e. = 0.112, t = − 0.350, p = 0.728, Cohen’s ƒ2 = 0.002; age-squared: β value = 0.24, s.e. = 0.114, t = 2.109, p = 0.039, Cohen’s ƒ2 = 0.077). The Pavlovian bias significantly decreased linearly with age (β value = − 0.903, s.e. = 0.223, t = − 4.044, p = 0.001, Cohen’s ƒ2 = 0.282) and also exhibited a significant effect of age-squared (β value = 0.784, s.e. = 0.227, t = 3.452, p = 0.001, Cohen’s ƒ2 = 0.205). These nonlinear effects revealed that the Go bias and the Pavlovian bias parameter estimates were both attenuated in adolescents, relative to younger and older individuals. The significant linear relationship between Pavlovian bias and age suggests that, in addition to an adolescent-specific effect, this bias is highest in the youngest individuals and decreases with age. By using the parameter estimates from all compared models to simulate choice behavior for each participant, the best-fitting model qualitatively reproduced the pattern of behavior evident in Fig. 2b (see Supplementary Fig. S4 for predictions from all models). Although parameters from the best-fitting model exhibited some degree of correlation (Supplementary Table S3), parameter recovery revealed that all parameters were highly recoverable (see Parameter Recovery section of the supplement).
Finally, analysis of response times revealed that older participants made faster responses, and responses to reward-associated stimuli were faster when the correct response was “Go” (Supplementary Table S4).
Discussion
In this study, participants were tasked with learning the optimal response to cues that were associated with either positive or negative outcomes. We investigated how hard-wired tendencies to approach cues associated with reward and to withhold action to cues associated with punishment might differentially constrain participants’ ability to learn optimal actions over the course of development. Unexpectedly, we found that adolescents’ learning was less biased, relative to both younger and older individuals, in two distinct ways. Whereas children and, to a lesser extent, adults exhibited a robust Pavlovian bias on their action value learning, this interference was reduced in adolescents. Adolescents also displayed a diminished bias toward action over inaction (the “Go” bias).
Adolescents in our study showed better performance in the “No-Go to Win” condition, relative to both adults and children. Consistent with our findings, past work in adults using variants of this task has observed the greatest evidence of Pavlovian interference with instrumental learning in this condition, with performance typically approaching chance in this condition18,21,39. Better performance on “Go to Win” than “No-Go to Win” trials can arise not only through reward invigoration of action, but also from an enhanced ability to “assign credit” for rewarding outcomes to past actions, rather than to inaction. Previous studies using task designs capable of dissociating these features of learning find that adults exhibit both a Pavlovian bias on action learning, as well as difficulty linking reward with inaction39,40. Thus, it is possible that in addition to a reduced Pavlovian bias, adolescents might also exhibit a heightened ability to learn associations between rewards and past inaction. Future studies should clarify whether developmental differences in the ability to assign credit to passive responses might contribute to adolescents’ superior performance in this task.
Children and adults in our sample, but not adolescents, exhibited a robust Pavlovian bias that was, in part, driven by a suppression of action to punishment-associated cues. In environments in which outcomes are stochastic, actions that are beneficial, on average, may yield occasional negative outcomes. Strong Pavlovian biases may prevent the continued engagement with previously punished contexts that is necessary to discover the opportunities for reward that they present41,42,43. Such tendencies to reflexively withdraw from or avoid situations associated with negative outcomes represent a type of “learning trap” that hinders an individual from sampling sufficiently to discern the true statistics of an environment, allowing potential rewards to go undiscovered44,45. Given the many novel contexts that adolescents encounter during their transition toward parental independence46,47, a reduced Pavlovian bias might facilitate exploration and unbiased evaluation of action values, allowing adolescents to approach uncertain or ambiguous situations to discover their true value48. However, such willingness to continue sampling in stochastic environments that may yield negative outcomes may contribute to a seemingly heightened willingness to take risks during this developmental stage49,50.
Pavlovian responses serve critical survival functions51,52. However, their expression can also constrain the flexibility of learning. Theoretical accounts propose that the degree of control afforded in learning environments may be used to optimally calibrate the expression of Pavlovian versus instrumental responses53,54,55,56. Instrumental learning is only adaptive in controllable environments, in which actions can leverage the true contingent causal structure of the environment to bring about beneficial outcomes. In contexts where outcomes are not contingent, the latent assumptions of causality inherent in instrumental learning are overly complex and inaccurate57, and reliance on Pavlovian “default” response tendencies may instead be optimal. In active avoidance paradigms, which afford the opportunity for instrumental control of outcomes, adolescent rodents learn to proactively prevent punishment (e.g., by shuttling) better than younger and older animals, whose performance is impaired to a greater degree by the competing Pavlovian tendency to freeze in the face of threat15,58. However, in uncontrollable environments, in which there is no effective instrumental response to an anticipated threat, adolescents readily learn and express Pavlovian responses that are particularly resistant to subsequent extinction59,60. A parsimonious account of these seemingly inconsistent findings may be that adolescents are particularly effective at detecting the controllability of a given situation and calibrating their reliance on Pavlovian versus instrumental action accordingly. Our finding of an adolescent-specific reduction in Pavlovian bias accords with this explanation, as adolescents might infer the highly controllable nature of our task and adaptively deploy instrumental action. However, further studies are needed to clarify whether the degree of environmental controllability modulates the expression of motivated behaviors differentially across development.
Studies examining the neural substrates of Pavlovian bias in adults suggest how age-related changes in the brain might give rise to the developmental pattern of behavior observed here. Consistent with computational models proposing that the value of both stimuli and actions are evaluated during learning61,62,63, the ventral and dorsal striatum, respectively, encode signals consistent with these computations64, and facilitate Pavlovian and instrumental behaviors65,66. The prefrontal cortex, through its projections to and from the striatum, is proposed to exhibit sensitivity to contexts in which stimulus and action values conflict and to enable attenuation of Pavlovian reactive biases in such contexts18,21,39. Corticostriatal circuitry undergoes marked structural and functional age-related changes from childhood into adulthood25,32,67,68,69,70,71,72,73,74. Reduced corticostriatal connectivity in children might constrain their ability to modulate Pavlovian biases on action learning, leading to the heightened expression of valence-action coupling in our youngest participants. Developmental increases in the integration between the prefrontal cortex and striatum may contribute to the linear reductions in these reactive responses with age. Dopamine, a neurotransmitter that innervates both the prefrontal cortex and the striatum, appears to modulate the expression of Pavlovian learning biases33,40. The dopaminergic system exhibits substantial reorganization across adolescence75,76, including nonlinear changes in dopamine receptor and transporter density77,78,79,80,81. Moreover, dopaminergic projections from the striatum to the prefrontal cortex continue to develop in adolescence82,83. Such nonlinear changes in the dopaminergic system and corticostriatal connectivity might contribute to the adolescent-specific attenuation of the Pavlovian bias. While the combined influence of these linear and nonlinear changes in corticostriatal circuits and dopaminergic signaling likely contributes to the developmental trajectory of Pavlovian bias expression, future studies are needed to understand the neural mechanisms underlying the age-related pattern observed here.
Although our study focused on interactions between Pavlovian and instrumental learning, an extensive literature further distinguishes two forms of instrumental action that are proposed to reflect distinct underlying computations84,85. “Goal-directed” actions are proposed to arise from a “model-based” learning process that computes the value of an action by prospectively searching a mental model of task contingencies and outcomes. In contrast, “habitual” action learning, forgoes the use of a model, instead allowing rewarding outcomes to directly reinforce associations between stimuli and responses. Habitual behavior is well approximated by a “model-free” reinforcement-learning algorithm that incrementally updates a stored estimate of the value of an action. Importantly, these two forms of instrumental action differ in their sensitivity to Pavlovian influence. Pavlovian reactions interfere with habitual behavior to a greater extent than goal-directed action22,86. Across tasks that assess these distinct forms of learning, adults who rely more on model-based strategies also exhibit reduced Pavlovian bias87. Moreover, pharmacological manipulation of dopamine alters both Pavlovian bias33,40 and the use of model-based strategies88, suggesting that a common neural mechanism contributes to the expression these distinct forms of value-based learning. While our task cannot directly differentiate goal-directed and habitual forms of instrumental learning, both goal-directed behavior and the model-based evaluations proposed to support it have been found to increase with age across development25,89,90,91,92. Thus, children’s greater reliance on model-free instrumental learning may confer heightened vulnerability to Pavlovian interference, with a shift toward model-based computation conferring greater resistance to Pavlovian bias with age.
The present study examined the extent to which Pavlovian responses constrain flexible action learning in healthy development. However, strong Pavlovian influences on instrumental action are characteristic of many forms of psychopathology that typically emerge during adolescence93,94, including substance abuse and anxiety disorders95,96,97,98,99. Adolescence is a period of heightened plasticity in the neural circuits that govern motivated behavior24,100,101,102,103, which may render adolescents’ expression of Pavlovian action biases particularly sensitive to experiential variation, such as stress exposure103,104. Consistent with this notion, past studies in rodents have observed that adolescents’ tendency to imbue cues that predict reward or punishment with inherent motivational value (i.e., “sign-tracking”), is reduced relative to adult animals105,106, but that this age-related pattern is reversed when adolescent animals are exposed to adverse rearing environments107,108. This experiential sensitivity aligns with evidence in humans that Pavlovian biases on action learning are exacerbated following adolescent exposure to trauma109. Further studies examining how life experience interacts with neural plasticity during adolescence to shape learning may shed light on mechanisms that promote resilience or susceptibility to psychopathology.
In this study, we characterized the developmental trajectory of reflexive Pavlovian constraints on flexible action learning. We found that the extent to which these valence-specific response tendencies interfered with instrumental action changed nonlinearly with age, showing a selective attenuation during adolescence. This decreased expression of hard-wired behavioral responses enables greater behavioral flexibility as adolescents learn to respond adaptively to opportunities and challenges in their environment. The influence of Pavlovian learning biases on fundamental behaviors that change over development such as exploration and planning, as well as their mechanistic role in multiple forms of psychopathology that typically emerge prior to adulthood, underscores the importance of a deeper understanding of how interactions between Pavlovian and instrumental learning systems contribute to adaptive and maladaptive behavior over the course of development.
Methods
Participants
Previous studies in adults using variants of this task found robust valence-action coupling in sample sizes of 20 participants total20,21 or group differences with 20–30 participants per group110,111. Thus, we targeted a sample size of 20 participants in each age bin, for a total of 60 participants. This is in line with sample sizes for previous studies that have examined developmental changes in evaluative processes36,89,92. Participants were recruited from the New York City metropolitan area through flyers and outreach events. Participants were screened to exclude for a diagnosis of mood or anxiety disorders, a learning disability, current use of beta-blockers or psychoactive medications, or colorblindness. Sixty-two individuals participated in the study. Data from one child were excluded due to a technical error. Our final sample size was 61 participants, consisting of 20 children (8–12 years old, n = 10 female, mean age: 10.55, SD: 1.34), 20 adolescents (13–17 years old, n = 11 female, mean age: 15.38, SD: 1.45), and 21 adults (18–25 years old, n = 10 female, mean age: 21.28, SD: 2.19). The sample included 31 Caucasians (50.82%), 14 individuals of mixed race (22.95%), 12 Asians (19.67%), 3 African Americans (4.92%), and 1 Pacific Islander (1.64%). Eleven participants identified as Hispanic (18.03%). The study protocol was approved by New York University’s Institutional Review Board, the University Committee on Activities Involving Human Subjects. All research was performed in accordance with the relevant guidelines and regulations. All adult participants and parents of minors provided written informed consent and minors provided assent prior to the study.
Task details
On each trial, participants saw the cue, which was one of the four robots (1,000 ms), followed by a fixation cross (250–3,500 ms), and then the target (described to participants as the robot’s “button”). When participants saw the robot’s “button”, they could decide whether to press the button via a keyboard press (“Go” response) or not press the button (“No-Go” response). Participants had 1,500 ms to respond. If they made a “Go” response, the border of the target would enlarge for the remainder of the 1,500 ms. Following the target, another fixation cross appeared on the screen (1,000 ms) prior to receiving probabilistic feedback (2000 ms). During the inter-trial interval, a fixation cross was presented on the screen (750–1,500 ms).
There were 45 trials for each of the four trial types, resulting in 180 trials total. The colors of the robots were randomized across participants. Stimuli were presented in pseudo-random order ensuring that each robot appeared fifteen times in each of three blocks. After every block of 60 trials, participants were given a break. Prior to the task, participants practiced making “Go” responses by pressing the button and withholding their button press to make “No-Go” responses. Participants also practiced pressing a button and not pressing a button for each type of robot, experiencing that a given robot would give (or take) a ticket for one action and not give (or take) a ticket for the other action. The probabilistic nature of the reinforcements was learned through instruction and experience during the actual game, although no information about the reinforcement probabilities was provided. To encourage learning, participants were instructed that the greater number of tickets won would result in more bonus money at the end of the study, though they were not informed of the specific relationship between tickets and money. In reality, all participants received a $5 bonus regardless of their performance on the task. This experiment was programmed using Cogent 2000, a MATLAB toolbox.
Statistical analysis
Age and age-squared were z-scored prior to inclusion in any linear regression. To calculate age-squared, age was first z-scored and then squared. All p-values reported within the manuscript reflect a two-tailed significance test unless otherwise indicated. Cohen’s ƒ2 is a measure of effect size for regression analyses and can be used to calculate the local effect size of predictors in multiple regression. Small, medium, and large effect sizes are represented by ƒ2 ≥ 0.02, ƒ2 ≥ 0.15, and ƒ2 ≥ 0.35, respectively112.
Computational models
Model fitting procedures and parameter estimation were performed using maximum a posteriori method using the fmincon function in Matlab 9.1.0. In addition to the best-fitting model, we also compared models that divided the reinforcement sensitivity term into separate reward sensitivity (ρrew) and punishment sensitivity (ρpun) parameters, capturing asymmetry in the effective size of positive and negative reinforcement. Whereas the best-fitting model included weights on the action estimate, in the simplest models the weighted action estimate (Wt(at,st)) was equivalent to the Q value estimate.
The parameters were constrained as follows: the lapse rate and learning rate between 0 to 1; the reinforcement sensitivity parameters (ρ, ρrew, ρpun) and the Pavlovian bias between 0 and ∞; the Go bias parameter was not constrained (-∞ to ∞). Priors were chosen to be minimally informative and were based on previous reinforcement learning studies113. A prior of Beta(1.1, 1.1) was employed for parameters constrained between 0 and 1. A prior of Normal(0, 1) was employed for parameters constrained between negative infinity and infinity. A prior of Gamma(2, 3) was employed for parameters constrained between 0 and infinity.
Data availability
Data are available on Open Science Framework: https://osf.io/4h6ne/.
Code availability
Code to reproduce all analyses in the manuscript can be found on Open Science Framework: https://osf.io/4h6ne/.
References
Pavlov, I. P. Conditional reflexes: an investigation of the physiological activity of the cerebral cortex (Oxford University Press, Oxford, 1927).
Hershberger, W. A. An approach through the looking-glass. Anim. Learn. Behav. 14, 443–451 (1986).
Williams, D. R. & Williams, H. Auto-maintenance in the pigeon: sustained pecking despite contingent non-reinforcement. J. Exp. Anal. Behav. 12, 511–520 (1969).
Gray, J. A. & McNaughton, N. The neuropsychology of anxiety: an enquiry into the function of the septo-hippocampal system (Oxford University Press, Oxford, 2000).
Bolles, R. C. Species-specific defense reactions and avoidance learning. Psychol. Rev. 77, 32–48 (1970).
O’Doherty, J. P. Multiple systems for the motivational control of behavior and associated neural substrates in humans. In: Behavioral neuroscience of motivation (eds. Simpson, E. H. & Balsam, P. D.) 291–312 (Springer International Publishing, Berlin, 2016). https://doi.org/10.1007/7854_2015_386.
Boureau, Y.-L. & Dayan, P. Opponency revisited: competition and cooperation between dopamine and serotonin. Neuropsychopharmacology 36, 74–97 (2011).
Breland, K. & Breland, M. The misbehavior of organisms. Am. Psychol. 16, 681–684 (1961).
Dickinson, A. & Balleine, B. Motivational control of goal-directed action. Anim. Learn. Behav. 22, 1–18 (1994).
Estes, W. K. Discriminative conditioning. I. A discriminative property of conditioned anticipation. J. Exp. Psychol. 32, 150–155 (1943).
Lovibond, P. F. Facilitation of instrumental behavior by a Pavlovian appetitive conditioned stimulus. J. Exp. Psychol. Anim. Behav. Process. 9, 225 (1983).
Rescorla, R. & Solomon, R. Two-process learning theory: relationships between Pavlovian conditioning and instrumental learning. Psychol. Rev. 74, 151–182 (1967).
Choi, J.-S., Cain, C. K. & LeDoux, J. E. The role of amygdala nuclei in the expression of auditory signaled two-way active avoidance in rats. Learn. Mem. 17, 139–147 (2010).
Galatzer-Levy, I. R. et al. Heterogeneity in signaled active avoidance learning: substantive and methodological relevance of diversity in instrumental defensive responses to threat cues. Front. Syst. Neurosci. 8, 179 (2014).
Stavnes, K. & Sprott, R. L. Effects of age and genotype on acquisition of an active avoidance response in mice. Dev. Psychobiol. 8, 437–445 (1975).
Holland, P. C. Differential effects of omission contingencies on various components of Pavlovian appetitive conditioned responding in rats. J. Exp. Psychol. Anim. Behav. Process. 5, 178–193 (1979).
Bray, S., Rangel, A., Shimojo, S., Balleine, B. & O’Doherty, J. P. The Neural mechanisms underlying the influence of pavlovian cues on human decision making. J. Neurosci. 28, 5861–5866 (2008).
Cavanagh, J. F., Eisenberg, I., Guitart-Masip, M., Huys, Q. & Frank, M. J. Frontal theta overrides pavlovian learning biases. J. Neurosci. 33, 8541–8548 (2013).
Crockett, M. J., Clark, L. & Robbins, T. W. Reconciling the role of serotonin in behavioral inhibition and aversion: acute tryptophan depletion abolishes punishment-induced inhibition in humans. J. Neurosci. 29, 11993–11999 (2009).
Guitart-Masip, M. et al. Action dominates valence in anticipatory representations in the human striatum and dopaminergic midbrain. J. Neurosci. 31, 7867–7875 (2011).
Guitart-Masip, M. et al. Go and no-go learning in reward and punishment: interactions between affect and effect. Neuroimage 62–334, 154–166 (2012).
Holland, P. C. Relations between Pavlovian-instrumental transfer and reinforcer devaluation. J. Exp. Psychol. Anim. Behav. Process. 30, 104–117 (2004).
Talmi, D., Seymour, B., Dayan, P. & Dolan, R. J. Human Pavlovian-instrumental transfer. J. Neurosci. 28, 360–368 (2008).
Hartley, C. A. & Lee, F. S. Sensitive periods in affective development: nonlinear maturation of fear learning. Neuropsychopharmacology 40, 50–60 (2015).
Raab, H. A. & Hartley, C. A. The development of goal-directed decision-making. in Goal-Directed Decision Making (eds. Morris, R., Bornstein, A. & Shenhav, A.) 279–308 (Academic Press, Cambridge, 2018). https://doi.org/10.1016/B978-0-12-812098-9.00013-9.
Rovee-Collier, C. K. & Gekoski, M. J. The economics of infancy: a review of conjugate reinforcement. Adv. Child Dev. Behav. 13, 195–255 (1979).
Shechner, T., Hong, M., Britton, J. C., Pine, D. S. & Fox, N. A. Fear conditioning and extinction across development: evidence from human studies and animal models. Biol. Psychol. 100, 1–12 (2014).
Moutoussis, M. et al. Change, stability, and instability in the Pavlovian guidance of behaviour from adolescence to young adulthood. PLOS Comput. Biol. 14, e1006679 (2018).
Bunge, S. A. & Wright, S. B. Neurodevelopmental changes in working memory and cognitive control. Curr. Opin. Neurobiol. 17, 243–250 (2007).
Diamond, A. The early development of executive functions. in Lifespan Cognition: Mechanisms of Change. Bialystok E, Craik F, editors. 70–95 (Oxford University Press, Oxford 2006).
Luna, B. Developmental changes in cognitive control through adolescence. in Advances in Child Development and Behavior (ed. Bauer, P.) vol. 37 233–278 (JAI, Amsterdam, 2009).
Somerville, L. H. & Casey, B. Developmental neurobiology of cognitive control and motivational systems. Curr. Opin. Neurobiol. 20, 236–241 (2010).
Guitart-Masip, M. et al. Differential, but not opponent, effects of L -DOPA and citalopram on action learning with reward and punishment. Psychopharmacology 231, 955–966 (2014).
Nook, E. C., Sasse, S. F., Lambert, H. K., McLaughlin, K. A. & Somerville, L. H. The nonlinear development of emotion differentiation: granular emotional experience is low in adolescence. Psychol. Sci. 29, 1346–1357 (2018).
Rodman, A. M., Powers, K. E. & Somerville, L. H. Development of self-protective biases in response to social evaluative feedback. Proc. Natl. Acad. Sci. 114, 13158–13163 (2017).
Somerville, L. H. et al. The medial prefrontal cortex and the emergence of self-conscious emotion in adolescence. Psychol. Sci. 24, 1554–1562 (2013).
Akaike, H. A new look at the statistical model identification. IEEE Trans. Autom. Control 19, 716–723 (1974).
Sutton, R. S. & Barto, A. G. Reinforcement Learning. (MIT Press, 1998).
Swart, J. C. et al. Frontal network dynamics reflect neurocomputational mechanisms for reducing maladaptive biases in motivated action. PLOS Biol. 16, e2005979 (2018).
Swart, J. C. et al. Catecholaminergic challenge uncovers distinct Pavlovian and instrumental mechanisms of motivated (in) action. eLife 6, e22169 (2017).
Huys, Q. J. M. et al. Bonsai trees in your head: how the pavlovian system sculpts goal-directed choices by pruning decision trees. PLOS Comput. Biol. 8, e1002410 (2012).
Huys, Q. J. M. et al. Interplay of approximate planning strategies. Proc. Natl. Acad. Sci. 112, 3098–3103 (2015).
Lally, N. et al. The neural basis of aversive Pavlovian guidance during planning. J. Neurosci. 37, 10215–10229 (2017).
Denrell, J. & March, J. G. Adaptation as information restriction: the hot stove effect. Organ. Sci. 12, 523–538 (2001).
Rich, A. S. & Gureckis, T. M. The limits of learning: exploration, generalization, and the development of learning traps. J. Exp. Psychol. Gen. 147, 1553 (2018).
Casey, B., Duhoux, S. & Cohen, M. M. Adolescence: what do transmission, transition, and translation have to do with it?. Neuron 67, 749–760 (2010).
Spear, L. P. The adolescent brain and age-related behavioral manifestations. Neurosci. Biobehav. Rev. 24, 417–463 (2000).
Tymula, A. et al. Adolescents’ risk-taking behavior is driven by tolerance to ambiguity. Proc. Natl. Acad. Sci. 109, 17135–17140 (2012).
Rosenbaum, G. M. & Hartley, C. A. Developmental perspectives on risky and impulsive choice. Philos. Trans. R. Soc. B Biol. Sci. 374, 20180133 (2019).
Steinberg, L. A social neuroscience perspective on adolescent risk-taking. Dev. Rev. DR 28, 78–106 (2008).
Bach, D. R. & Dayan, P. Algorithms for survival: a comparative perspective on emotions. Nat. Rev. Neurosci. 18, 311–319 (2017).
LeDoux, J. & Daw, N. D. Surviving threats: neural circuit and computational implications of a new taxonomy of defensive behaviour. Nat. Rev. Neurosci. 19, 269–282 (2018).
Huys, Q. J. M. & Dayan, P. A Bayesian formulation of behavioral control. Cognition 113, 314–328 (2009).
Lloyd, K. & Dayan, P. Safety out of control: dopamine and defence. Behav. Brain Funct. 12, (2016).
Moscarello, J. M. & Hartley, C. A. Agency and the calibration of motivated behavior. Trends Cogn. Sci. 21, 725–735 (2017).
Rigoli, F., Pezzulo, G. & Dolan, R. J. Prospective and Pavlovian mechanisms in aversive behaviour. Cognition 146, 415–425 (2016).
Dorfman, H. M. & Gershman, S. J. Controllability governs the balance between Pavlovian and instrumental action selection. Nat. Commun. 10, 1–8 (2019).
Bauer, R. H. Ontogeny of two-way avoidance in male and female rats. Dev. Psychobiol. 11, 103–116 (1978).
McCallum, J., Kim, J. H. & Richardson, R. Impaired extinction retention in adolescent rats: effects of D-Cycloserine. Neuropsychopharmacology 35, 2134–2142 (2010).
Pattwell, S. S. et al. Altered fear learning across development in both mouse and human. Proc. Natl. Acad. Sci. 109, 16318–16323 (2012).
Barto, A. G. Adaptive critics and the basal ganglia. in Models of information processing in the basal ganglia 215–232 (The MIT Press, Cambridge, 1995).
Barto, A. G., Sutton, R. S. & Anderson, C. W. Neuronlike adaptive elements that can solve difficult learning control problems. IEEE Trans. Syst. Man Cybern. 13, 834–846 (1983).
Maia, T. V. Two-factor theory, the actor-critic model, and conditioned avoidance. Learn. Behav. 38, 50–67 (2010).
O’Doherty, J. et al. Dissociable roles of ventral and dorsal striatum in instrumental conditioning. Science 304, 452–454 (2004).
Cardinal, R. N., Parkinson, J. A., Hall, J. & Everitt, B. J. Emotion and motivation: the role of the amygdala, ventral striatum, and prefrontal cortex. Neurosci. Biobehav. Rev. 26, 321–352 (2002).
Dayan, P. & Balleine, B. W. Reward, motivation, and reinforcement learning. Neuron 36, 285–298 (2002).
van den Bos, W., Cohen, M. X., Kahnt, T. & Crone, E. A. Striatum-medial prefrontal cortex connectivity predicts developmental changes in reinforcement learning. Cereb. Cortex 22, 1247–1255 (2012).
van Duijvenvoorde, A. C. K., Achterberg, M., Braams, B. R., Peters, S. & Crone, E. A. Testing a dual-systems model of adolescent brain development using resting-state connectivity analyses. NeuroImage 124, 409–420 (2016).
Gogtay, N. et al. Dynamic mapping of human cortical development during childhood through early adulthood. Proc. Natl. Acad. Sci. U. S. A. 101, 8174–8179 (2004).
Larsen, B., Verstynen, T. D., Yeh, F.-C. & Luna, B. Developmental changes in the integration of affective and cognitive corticostriatal pathways are associated with reward-driven behavior. Cereb. Cortex 28, 2834–2845 (2018).
Liston, C. et al. Frontostriatal microstructure modulates efficient recruitment of cognitive control. Cereb. Cortex N. Y. N 1991(16), 553–560 (2006).
Mills, K. L. et al. Structural brain development between childhood and adulthood: convergence across four longitudinal samples. NeuroImage 141, 273–281 (2016).
Raznahan, A. et al. Longitudinal four-dimensional mapping of subcortical anatomy in human development. Proc. Natl. Acad. Sci. 111, 1592–1597 (2014).
Sowell, E. R., Thompson, P. M., Holmes, C. J., Jernigan, T. L. & Toga, A. W. In vivo evidence for post-adolescent brain maturation in frontal and striatal regions. Nat. Neurosci. 2, 859–861 (1999).
Hoops, D. & Flores, C. Making dopamine connections in adolescence. Trends Neurosci. 40, 709–719 (2017).
Wahlstrom, D., White, T. & Luciana, M. Neurobehavioral evidence for changes in dopamine system activity during adolescence. Neurosci. Biobehav. Rev. 34, 631–648 (2010).
Andersen, S. L., Rutstein, M., Benzo, J. M., Hostetter, J. C. & Teicher, M. H. Sex differences in dopamine receptor overproduction and elimination. NeuroReport 8, 1495 (1997).
Andersen, S. L., Thompson, A. T., Rutstein, M., Hostetter, J. C. & Teicher, M. H. Dopamine receptor pruning in prefrontal cortex during the periadolescent period in rats. Synapse 37, 167–169 (2000).
Meng, S. Z., Ozawa, Y., Itoh, M. & Takashima, S. Developmental and age-related changes of dopamine transporter, and dopamine D1 and D2 receptors in human basal ganglia. Brain Res. 843, 136–144 (1999).
Montague, D. M., Lawler, C. P., Mailman, R. B. & Gilmore, J. H. Developmental regulation of the dopamine D 1 receptor in human caudate and putamen. Neuropsychopharmacology 21, 641–649 (1999).
Teicher, M. H., Andersen, S. L. & Hostetter, J. C. Evidence for dopamine receptor pruning between adolescence and adulthood in striatum but not nucleus accumbens. Dev. Brain Res. 89, 167–172 (1995).
Hoops, D., Reynolds, L. M., Restrepo-Lozano, J.-M. & Flores, C. Dopamine development in the mouse orbital prefrontal cortex is protracted and sensitive to amphetamine in adolescence. eNeuro 5, (2018).
Reynolds, L. M. et al. DCC receptors drive prefrontal cortex maturation by determining dopamine axon targeting in adolescence. Biol. Psychiatry 83, 181–192 (2018).
Balleine, B. W. & O’Doherty, J. P. Human and rodent homologies in action control: corticostriatal determinants of goal-directed and habitual action. Neuropsychopharmacology 35, 48–69 (2010).
Dickinson, A. Actions and habits: the development of behavioural autonomy. Phil. Trans. R. Soc. Lond. B 308, 67–78 (1985).
Yin, H. H. & Knowlton, B. J. The role of the basal ganglia in habit formation. Nat. Rev. Neurosci. 7, 464–476 (2006).
Sebold, M. et al. Don’t think, just feel the music: individuals with strong pavlovian-to-instrumental transfer effects rely less on model-based reinforcement learning. J. Cogn. Neurosci. 28, 985–995 (2016).
Wunderlich, K., Smittenaar, P. & Dolan, R. J. Dopamine enhances model-based over model-free choice behavior. Neuron 75, 418–424 (2012).
Decker, J. H., Otto, A. R., Daw, N. D. & Hartley, C. A. From creatures of habit to goal-directed learners tracking the developmental emergence of model-based reinforcement learning. Psychol. Sci. 27, 848–858 (2016).
Kenward, B., Folke, S., Holmberg, J., Johansson, A. & Gredebäck, G. Goal directedness and decision making in infants. Dev. Psychol. 45, 809–819 (2009).
Klossek, U. M. H., Russell, J. & Dickinson, A. The control of instrumental action following outcome devaluation in young children aged between 1 and 4 years. J. Exp. Psychol. Gen. 137, 39–51 (2008).
Potter, T. C. S., Bryce, N. V. & Hartley, C. A. Cognitive components underpinning the development of model-based learning. Dev. Cogn. Neurosci. 25, 272–280 (2017).
Kessler, R. C. et al. Lifetime prevalence and age-of-onset distributions of dsm-iv disorders in the national comorbidity survey replication. Arch. Gen. Psychiatry 62, 593–602 (2005).
Paus, T., Keshavan, M. & Giedd, J. N. Why do many psychiatric disorders emerge during adolescence?. Nat. Rev. Neurosci. 9, 947–957 (2008).
Carter, B. L. & Tiffany, S. T. Meta-analysis of cue-reactivity in addiction research. Addict. Abingdon Engl. 94, 327–340 (1999).
Everitt, B. J., Dickinson, A. & Robbins, T. W. The neuropsychological basis of addictive behaviour. Brain Res. Rev. 36, 129–138 (2001).
Garbusow, M. et al. Pavlovian-to-instrumental transfer effects in the nucleus accumbens relate to relapse in alcohol dependence. Addict. Biol. 21, 719–731 (2016).
Mineka, S. & Oehlberg, K. The relevance of recent developments in classical conditioning to understanding the etiology and maintenance of anxiety disorders. Acta Psychol. (Amst.) 127, 567–580 (2008).
Mkrtchian, A., Aylward, J., Dayan, P., Roiser, J. P. & Robinson, O. J. Modeling avoidance in mood and anxiety disorders using reinforcement learning. Biol. Psychiatry 82, 532–539 (2017).
Dahl, R. E. Adolescent brain development: a period of vulnerabilities and opportunities. Keynote Address. Ann. N. Y. Acad. Sci. 1021, 1–22 (2004).
Fuhrmann, D., Knoll, L. J. & Blakemore, S.-J. Adolescence as a sensitive period of brain development. Trends Cogn. Sci. 19, 558–566 (2015).
Meyer, H. C. & Lee, F. S. Translating developmental neuroscience to understand risk for psychiatric disorders. Am. J. Psychiatry 176, 179–185 (2019).
Romeo, R. D. & McEwen, B. S. Stress and the adolescent brain. Ann. N. Y. Acad. Sci. 1094, 202–214 (2006).
Lupien, S. J., McEwen, B. S., Gunnar, M. R. & Heim, C. Effects of stress throughout the lifespan on the brain, behaviour and cognition. Nat. Rev. Neurosci. 10, 434–445 (2009).
Anderson, R. I. & Spear, L. P. Autoshaping in adolescence enhances sign-tracking behavior in adulthood: Impact on ethanol consumption. Pharmacol. Biochem. Behav. 98, 250–260 (2011).
Doremus-Fitzwater, T. L. & Spear, L. P. Amphetamine-induced incentive sensitization of sign-tracking behavior in adolescent and adult female rats. Behav. Neurosci. 125, 661–667 (2011).
Anderson, R. I., Bush, P. C. & Spear, L. P. Environmental manipulations alter age differences in attribution of incentive salience to reward-paired cues. Behav. Brain Res. 257, 83–89 (2013).
DeAngeli, N. E., Miller, S. B., Meyer, H. C. & Bucci, D. J. Increased sign-tracking behavior in adolescent rats. Dev. Psychobiol. 59, 840–847 (2017).
Ousdal, O. T. et al. The impact of traumatic stress on Pavlovian biases. Psychol. Med. 48, 327–336 (2018).
de Boer, L. et al. Dorsal striatal dopamine D1 receptor availability predicts an instrumental bias in action learning. Proc. Natl. Acad. Sci. 116, 261–270 (2019).
Guitart-Masip, M. et al. Action controls dopaminergic enhancement of reward representations. Proc. Natl. Acad. Sci. 109, 7511–7516 (2012).
Cohen, J. A power primer. Psychol. Bull. 112, 155–159 (1992).
Daw, N. D., Gershman, S. J., Seymour, B., Dayan, P. & Dolan, R. J. Model-based influences on humans’ choices and striatal prediction errors. Neuron 69, 1204–1215 (2011).
Acknowledgements
This work was supported by a Jacobs Foundation Early Career Fellowship, a NARSAD Young Investigator Award, a Klingenstein-Simons Fellowship in Neuroscience, the National Science Foundation (CAREER Grant 1654393 to C.A.H. and Graduate Research Fellowship to H.A.R.), the NYU Vulnerable Brain Project, and the NYU High Performance Computing resources, service, and staff expertise. We are grateful to Shivani Hiralall for her assistance with participant recruitment and testing, Marc Guitart-Masip for sharing the original version of the task, and the families who participated in this study.
Author information
Authors and Affiliations
Contributions
H.A.R. and C.A.H. designed the study. H.A.R. implemented the task, collected the data, and performed analyses under the supervision of C.A.H. H.A.R. and C.A.H. wrote the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Raab, H.A., Hartley, C.A. Adolescents exhibit reduced Pavlovian biases on instrumental learning. Sci Rep 10, 15770 (2020). https://doi.org/10.1038/s41598-020-72628-w
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-020-72628-w
This article is cited by
-
Compulsive avoidance in youths and adults with OCD: an aversive pavlovian-to-instrumental transfer study
Translational Psychiatry (2024)
-
Emotional cues reduce Pavlovian interference in feedback-based go and nogo learning
Psychological Research (2024)
-
Examining the Association between Punishment and Reward Sensitivity and Response Inhibition to Previously-Incentivized Cues across Development
Journal of Youth and Adolescence (2024)