Introduction

Social anxiety disorder (SAD) and major depressive disorder (MDD) are prevalent, impairing internalizing disorders [1]. Effective treatment can reduce this burden and cognitive behavioral therapy (CBT) is considered the ā€˜gold standardā€™ psychotherapy for internalizing disorders [2]. Yet, CBT outcome varies and many remain clinically symptomatic following treatment [3,4,5], thereby highlighting the need to delineate markers that prospectively identify who is most likely to benefit from CBT.

An important construct in understanding heterogeneity in CBT outcomes is attentional control. Attentional control facilitates goal-directed behavior by inhibiting task-irrelevant information that competes with goals and is considered a form of implicit emotion regulation [6]. Patients with SAD and MDD demonstrate deficiencies in attentional control over negative information, resulting in attentional bias for negative stimuli [7, 8]. CBT utilizes an amalgam of techniques that draw on executive functions including attentional control [9] and is hypothesized to strengthen top-down regulatory processes [10]. Thus, transdiagnostic baseline individual differences in attentional control mechanisms in the context of task-irrelevant negative stimuli (i.e., emotional interference) may interact with CBT-specific techniques to predict symptom improvement.

The anterior cingulate cortex (ACC) plays a key role in attentional control (i.e., resolving emotional interference) [11,12,13] and subregion specialization constitutes the dorsal ACC (dACC), which is involved in conflict monitoring and the deployment of cognitive control, and rostral ACC (rACC), which is involved in emotional conflict resolution and the assessment of motivationally relevant information [11,12,13]. Separate lines of research indicate ACC is a promising biomarker of CBT outcomes. In SAD, more baseline dACC activity in response to task-irrelevant (>task-relevant) emotional faces [14] and task-irrelevant threat (>neutral) faces [15] has been shown to correspond with greater symptom improvement following CBT. Also, more baseline dACC activity in response to task-irrelevant numbers during a cognitive interference paradigm predicts greater reduction in social anxiety severity following CBT with or without pharmacotherapy [16].

We are not aware of emotional interference studies that haveĀ evaluated ACC as a predictor of CBT outcome in MDD. However, meta-analytic studies suggest pre-treatment rACC engagement is a general marker of treatment response in MDD [17, 18]. For example, more rACC activity at rest foretells greater symptom improvement following pharmacological [19], sleep deprivation [20], and neuromodulation [21] interventions. However, findings regarding ACC as a predictor of CBT response in MDD are mixed. One study found less baseline dACC response to sad faces [22] corresponded with more symptom improvement following CBT. Other studies report similar relations, though in subgenual ACC (sgACC), an rACC subregion [23]. Specifically, less sgACC during rest [24, 25] or in response to self-referential stimuli [26, 27] predicts more symptom improvement. The ACC is involved in various functions beyond attentional control (e.g., self-referential processing [28], emotion perception [29]), which likely contribute to mixed results along with other methodological differences between studies. Even so, findings suggest ACC serves as a treatment outcome marker in MDD. An important next step is determining whether ACC as it relates to attentional control is a transdiagnostic marker of CBT outcome.

When evaluating attentional control, it is important to consider task difficulty as attention is a limited capacity system [30, 31]. According to load theory, when perceptual load is low (e.g., task is easy to perform), emotional interference is high as residual attentional resources are ā€˜left overā€™ to process task-irrelevant information. Conversely, when perceptual load is high (e.g., task is difficult to perform), emotional interference is low as resources are restricted [32]. Therefore, when evaluating attentional control in the context of negative ā€˜distractorsā€™, individual differences are likely to be observed when perceptual load is low. In support, associations have been observed between rACC or dACC activity to threat (>neutral) face distractors, and self-reported trait anxiety, behavioral inhibition, and subjective attentional control in healthy participants or patients with internalizing disorders under low, but not high, perceptual load [33,34,35]. This said, when evaluating CBT outcome in SAD, more baseline dACC activity to threat (>neutral) face distractors under high, but not low, load predicted greater pre-to-post CBT decrease in social anxiety severity [15]. Findings suggest variance in ACC to threat distractors under low load is sensitive to individual differences in constructs that factor into psychopathology, though variance in the more cognitively effortful high load condition may also interact with CBT.

The current study sought to extend previous findings by examining whether ACC as a predictor of symptom improvement was transdiagnostic and unique to CBT versus psychotherapy in general. Specifically, CBT involves a range of strategies to reduce maladaptive cognitions and behaviors in addition to common factors shared across psychotherapies such as rapport and treatment adherence [36]. Therefore, supportive therapy (ST) was included as a ā€˜common factorsā€™ comparator treatment arm to examine whether ACC interacted with CBT-specific techniques or common factors. Also, we examined whether baseline dACC or rACC activity to threat (>neutral) face distractors predicted trajectory of symptom change over the course of treatment as opposed to one endpoint (i.e., post-treatment symptom severity), which does not allow for the characterization of patterns of change over time (e.g., [37, 38]). This may be particularly important when comparing two psychotherapies, as the shape of symptom trajectories may differ even if post-treatment symptom severity is the same.

Based on prior CBT-related findings in SAD, we hypothesized that greater baseline dACC activity in response to threat distractors under high perceptual load would correspond with steeper trajectories of symptom improvement. In addition, we hypothesized that the relation between baseline ACC and symptom trajectories would not differ between patients with SAD and MDD. Finally, as CBT explicitly targets regulatory processes [9], we expected ACC would predict symptom improvement for patients randomized to CBT, but not ST.

Materials and methods

Participants

Seventy-two treatment-seeking patients with a principal diagnosis of SAD or MDD were recruited as part of an ongoing parallel-group randomized control trial (1:1 schedule) examining neural predictors and mechanisms of CBT and ST treatment response (ClinicalTrials.gov Identifier: NCT03175068). All participants were required to be between 18- and 65- years old and have a diagnosis of SAD (nā€‰=ā€‰39) or MDD (nā€‰=ā€‰33) based on DSM-5 diagnostic criteria. Patients were also required to either surpass a symptom severity threshold (i.e., a Liebowitz Social Anxiety Scale (LSAS; [39]) score ā‰„60 for SAD patients; a Hamilton Depression Rating Scale (HAMD; [40]) score ā‰„17 or a Beck Depression Inventory-II [41] score ā‰„16 for MDD patients) or report clinically significant levels of impairment (based on diagnostic interview). Regarding demographic characteristics, the average age for SAD patients was 25.51 (SDā€‰=ā€‰7.50), and 61.5% identified as women. For MDD patients, the average age was 29.82 (SDā€‰=ā€‰10.33) and 69.7% identified as women.

Although comorbidity was allowed, patients with MDD were not permitted to have comorbid SAD and vice-versa. SeeĀ Supplementary Materials for exclusion criteria. All participants tested negative for substance use and pregnancy before the MRI scan.

Participants were recruited between September 2017 and December 2020 from a local outpatient clinic and community. After obtaining informed consent, a psychiatric interview was conducted consisting of the Structured Clinical Interview for DSM-5 [42], LSAS, and HAMD. A Best-Estimate/Consensus Panel of at least three study staff members determined study eligibility and participants were informed of assignment to CBT or ST by a non-treating clinician. All study and treatment procedures were conducted at the University of Illinois at Chicago and were approved by the university Institutional Review Board and complied with the Helsinki Declaration. All participants were compensated for their time.

Treatment procedures

Patients were randomized to receive either 12 weeks of CBT or 12 weeks of ST using a covariate adaptive randomization (i.e., minimization) approach. See Fig.Ā 1 for information regarding study flow and patient retention. As seen, 37 CBT patients and 35 ST patients completed treatment and all baseline and follow-up measures for analyses.

Fig. 1: CONSORT flow diagram illustrating patient recruitment, treatment allocation, follow-up, and analysis.
figure 1

CBT cognitive behavioral therapy, ST supportive therapy.

At the time of interim analysis, the majority of the proposed final patient sample (nā€‰=ā€‰140) had completed treatment before COVID-19, thereby reducing possible confounds. Also, the sample sizes for each treatment arm exceeds the sample sizes of previous studies that either observed an effect of ACC activation during attentional control on CBT treatment outcome in patients with SAD (ns: 21, 32; [14, 15]) or observed treatment differences (i.e., CBT with versus without SSRIs) in the relation between ACC activity and SAD symptom improvement (ns: per arm: 23, 24; [16]).

Patients received 12 weekly 60-min psychotherapy sessions from trained PhD-level licensed clinical psychologists. Participants randomized to CBT received manualized treatment for SAD or MDD (e.g. [43,44,45]); the core content of sessions involved psychoeducation, cognitive restructuring, behavioral activation (for MDD), in vivo exposures (for SAD), and relapse prevention. ST resembled client-centered therapy (e.g. [46, 47]) and emphasized reflective listening and elicitation of effect as appropriate. SeeĀ Supplementary Materials for information regarding treatment fidelity and competency.

Clinical symptoms

The LSAS and HAMD served as primary outcome measures and were administered to patients by a trained clinician blinded to the treatment arm before treatment, as well as at Sessions 2, 4, 6, 8, 10, and 12 (i.e., 7 time points) to assess social anxiety and depression severity, respectively. As expected, patients with SAD exhibited higher social anxiety symptoms, t(70)ā€‰=ā€‰10.92, pā€‰<ā€‰0.001, and patients with MDD exhibited higher depression symptoms, t(70)ā€‰=ā€‰3.45, pā€‰=ā€‰0.001, at baseline. Therefore, to examine clinical change across treatment, we utilized a composite score of patientsā€™ symptoms by summing the LSAS and HAMD total scores to create an outcome measure that was sensitive to clinical change for both diagnostic groups. Standardized z scores are not recommended in longitudinal studies, as standardization at each time point prevents examination of mean-level changes [48]. Therefore, consistent with other treatment-outcome studies combining different symptom measures [49, 50], we utilized the proportion of maximum scaling method [51] to transform LSAS and HAMD before creating composite scores for each assessment.

Emotional faces interference task

The emotional faces interference task (EFIT) has been used in prior studies [15, 33, 35]. In brief, participants were presented with a string of six letters superimposed on a task-irrelevant face distractor and instructed to identify target letters (N or X) via button press as quickly and accurately as possible. Task-irrelevant faces consisted of fearful, angry, andĀ neutral expressions from eight different individuals from a standardized stimulus set [52]. In low perceptual load trials, the string was comprised entirely of target letters whereas under high perceptual load, the string included a single target letter and five non-target consonants (e.g., HKMWZ) in randomized order. There were two image acquisition runs, each comprising 12 blocks of 5 trials. A mixed block/event-related design was employed whereby perceptual load (low vs. high) varied across blocks and the facial expressions varied within blocks on a trial-by-trial basis. Images were presented for 200ā€‰ms followed by a fixation cross presented for 1800 ms. Within blocks, the intertrial interval varied randomly between 2 and 6ā€‰s; trials between blocks were separated by 4ā€“8ā€‰s.

fMRI data acquisition, preprocessing, and analysis

Scanning during the EFIT task was conducted on a 3.0 Tesla MR 750 scanner (General Electric Healthcare; Waukesha, WI) using a standard radiofrequency coil. Blood oxygen-level dependent (BOLD) functional images were acquired using a gradient-echo echo-planar imaging sequence with the following parameters: repetition time (TR)ā€‰=ā€‰2ā€‰s, echo time (TE)ā€‰=ā€‰25ā€‰ms, flip angleā€‰=ā€‰900, field of viewā€‰=ā€‰22ā€‰Ć—ā€‰22ā€‰cm, acquisition matrix 64ā€‰Ć—ā€‰64; 44 axial, 3-mm-thick slices with no gap. The first 4 volumes from each run were discarded to allow for T1 equilibration effects. For anatomical localization, a high-resolution, T1-weighted volumetric anatomical scan was acquired.

Conventional preprocessing steps were used in Statistical Parametric Mapping (SPM12) software package (Wellcome Trust Centre for Neuroimaging, London www.fil.ion.ucl.ac.uk/spm). Images were temporally corrected to account for differences in slice time collection, spatially realigned to the first image of the first run, coregistered to the anatomical, normalized to a Montreal Neurological Institute (MNI) space using warping based on the anatomical, resampled to 2ā€‰Ć—ā€‰2ā€‰Ć—ā€‰2ā€‰mm voxels, and smoothed with an 8ā€‰mm isotropic Gaussian kernel.

A general linear model identifying the low and high perceptual load blocks for angry, fearful, and neutral faces was applied to the time series, convolved with the canonical hemodynamic response function and with a 128ā€‰s high-pass filter. Nuisance regressors comprising six motion parameters were included to correct for motion artifacts. To maximize the signal across threat content, angry and fearful faces were collapsed. Contrasts of interest were task-irrelevant threat (i.e., fearful/angry) faces>neutral faces under low (i.e., Threat Low>Neutral Low) and high (i.e., Threat High>Neutral High) perceptual load conditions. A priori regions of interest (ROI) were used to test hypotheses. The Automatic Anatomical Labeling (AAL) atlas 3 [53] was used to generate masks (Supplementary Fig.Ā 1) such that rACC comprised sgACC and pregenual ACC (pgACC) (search volumeā€‰=ā€‰21,704ā€‰mm3). Supracallosal ACC and AAL 3 median cingulate anterior to yā€‰=ā€‰0 delineated dACC (search volumeā€‰=ā€‰19256ā€‰mm3). MarsBaR [54] was used to extract activation (Ī² weights, arbitrary units [a.u.]) from all ROIs which were submitted to tests in Statistical Package for the Social Sciences (Chicago, IL version 26) and Amos (version 27).

Analytic plan

Analyses regarding task performance (i.e., EFIT accuracy and reaction time) are presented inĀ Supplementary Materials.

To evaluate whether contrasts of interest probed ACC in general, whole-brain analyses were conducted. SeeĀ Supplementary Materials for the analytic plan.

Hypotheses were tested using a series of latent growth curve models in treatment completers. First, in our base model we examined if internalizing composite symptom scores changed over the course of treatment and whether the treatment arm (CBT vs. ST) predicted symptom trajectory. To determine whether symptom improvement followed linear or non-linear trajectories of improvement, linear, quadratic, and cubic symptom trajectories were examined. Log-linear trajectories were also examined to account for trajectories indicating sudden gains. Second, we examined whether pre-treatment rACC and/or dACC activity to threat (>neutral) face distractors under low and high perceptual load predicted the intercept (i.e., baseline symptom severity) and slope (i.e., trajectories of symptom change) over the course of treatment (see Supplementary Fig.Ā 2). Third, we examined whether the treatment arm moderated the impact of baseline rACC and/or dACC response on symptom intercept and slope. Finally, to examine ACC as a transdiagnostic predictor, we examined whether diagnosis (SAD vs. MDD) moderated the impact of baseline ACC response on symptom intercept and slope (see Supplementary Fig.Ā 3 for model). A Bonferroni correction (adjusted pā€‰=ā€‰0.0125) was implemented for all primary analyses to control for the number of examined ROIs (i.e., rACC and dACC under both high and low perceptual load).

Given evidence baseline sgACC activity predicts CBT outcome in MDD [24,25,26,27], post hoc analysis was performed for significant rACC results to evaluate whether a specific subregion explained findings. Specifically, both sgACC and pgACC were simultaneously examined as predictors of the intercept and slope of symptom change within the same model to examine the unique influences of rACC subdivisions. Finally, as a test of robustness, post hoc analysis was conducted for significant findings to see if relations between ACC response and symptom trajectories were maintained when statistically controlling for baseline symptom composite scores.

Results

Task performance

Accuracy was greater and reaction time faster across participants when detecting task-relevant target letters under low relative to high perceptual load. Results are inĀ Supplementary Materials.

Intent to treat analysis

Analyses revealed racial identity and baseline symptom differences between treatment completers and non-completers, such that non-completers were more like to be Asian and to have lower depression symptoms at baseline. No other demographic or clinical differences were observed. SeeĀ Supplementary Materials and Supplementary TableĀ 1 for details.

Participant characteristics

Patient demographic and clinical characteristics are presented in TableĀ 1. As seen, patients randomly assigned to CBT or ST did not differ in any demographic characteristics, principal diagnosis, or symptom severity across any of the assessments. Regarding diagnostic differences, SAD patients were younger than MDD patients, t(70)ā€‰=ā€‰āˆ’2.04, pā€‰=ā€‰0.045, but did not differ from MDD patients on any other demographic characteristics (psā€‰ā‰„ā€‰0.13). Therefore, patient age was entered as a covariate for all analyses. Compared to MDD patients, SAD patients demonstrated higher symptom severity (i.e., composite score) across all assessments, tsā€‰=ā€‰3.08ā€“6.56, all pā€™sā€‰ā‰¤ā€‰0.003.

Table 1 Descriptive statistics and clinical and demographic characteristics.

Whole-brain analyses

One-sample t-tests for Threat Low>Neutral Low and Threat High>Neutral High were performed to determine whether contrasts of interest probed ACC. See full results inĀ Supplementary Materials, Supplementary TableĀ 2, and Supplementary Fig.Ā 4.

Preliminary analyses

Linear, quadratic, cubic, and log-linear growth models of symptom change were fit to the data. The linear growth model was determined to be the best fit and was, therefore, used in all subsequent analyses (seeĀ Supplementary Materials, Supplementary TableĀ 3, and Supplementary Fig.Ā 5 for model comparison). Results indicated that means for both symptom intercept, Mā€‰=ā€‰0.77, SEā€‰=ā€‰0.11, pā€‰<ā€‰0.001, and slope, Mā€‰=ā€‰āˆ’0.05, SEā€‰=ā€‰0.017, pā€‰=ā€‰0.003, were significant, suggesting symptoms linearly decreased throughout treatment. Treatment arm was not associated with symptom intercept, Ī²ā€‰=ā€‰āˆ’0.02, pā€‰=ā€‰0.87, or slope, Ī²ā€‰=ā€‰0.19, pā€‰=ā€‰0.16, suggesting patients in both arms demonstrated similar levels of baseline symptoms and similar symptom trajectories (see Fig.Ā 2). Patient age was not associated with symptom intercept, Ī²ā€‰=ā€‰āˆ’0.19, pā€‰=ā€‰0.13, or slope, Ī²ā€‰=ā€‰āˆ’0.04, pā€‰=ā€‰0.79, suggesting there was not an effect of age.

Fig. 2: Figure illustrating change in symptom severity across treatment for patients assigned to CBT or ST.
figure 2

CBT cognitive behavioral therapy, ST supportive therapy, Composite Symptom Score Hamilton Depression Rating Scale and Liebowitz Social Anxiety Scale Composite Score.

Baseline neural activity and symptom change

Next, a series of analyses examining baseline ACC as a predictor of symptom change was conducted. See TableĀ 2 for the result summary. Results indicated that greater baseline rACC response under low load was negatively associated with symptom slope, such that greater rACC activation predicted a steeper trajectory of symptom decrease, Ī²ā€‰=ā€‰āˆ’0.39, pā€‰=ā€‰0.002. Post-hoc analysis indicated that this main effect of rACC response on symptom slope was maintained when statistically controlling for baseline symptoms, Ī²ā€‰=ā€‰āˆ’0.31, pā€‰=ā€‰0.027. No other main effects of neural activation on symptom slope (all pā€™sā€‰ā‰„ā€‰0.02) survived Bonferroni correction.

Table 2 Impact of neural activation on symptom trajectories throughout treatment.

Post hoc analyses were conducted to examine whether rACC subdivisions (i.e., subgenual, pregenual) under low load were unique predictors of symptom slope. Results indicated greater sgACC activation, Ī²ā€‰=ā€‰āˆ’0.39, pā€‰=ā€‰0.009, but not pgACC activation, Ī²ā€‰=ā€‰āˆ’0.15, pā€‰=ā€‰0.32, was uniquely associated with steeper slope of symptom change. Post-hoc findings showed the sgACC effect was maintained when controlling for baseline symptom severity, Ī²ā€‰=ā€‰āˆ’0.38, pā€‰=ā€‰0.02 (see Fig.Ā 3).

Fig. 3: Figure illustrating subgenual anterior cingulate cortex (sgACC)Ā activationĀ as a predictor of symptom trajectories.
figure 3

A Anatomic mask for sgACC used as region of interest. B Figure depicting projected trajectories of symptom change (across Sessions 2ā€“12) as a function of patient baseline sgACC activation to threat (>neutral) emotional face distractors under low perceptual load while statistically controlling for covariates (i.e., treatment arm, patient age, pregenual anterior cingulate cortex activation, and baseline symptom severity). Simple slopes are depicted for low (āˆ’1 SD), average, and high (+1 SD) levels of sgACC activation.

Moderation by treatment arm

Primary analyses were re-conducted with the interaction between neural activation and arm included to test whether the treatment arm moderated the impact of baseline ACC on symptom improvement. None of the ROIā€‰Ć—ā€‰arm interactions significantly predicted symptom slope (all pā€™sā€‰ā‰„ā€‰0.28).

Moderation by diagnosis

Finally, primary analyses were re-conducted with patient diagnosis and the ROIā€‰Ć—ā€‰diagnosis interaction included to examine whether the relation between baseline ACC and symptom improvement was disorder specific. None of the ROIā€‰Ć—ā€‰ diagnosis interactions predicted symptom slope (all pā€™sā€‰ā‰„ā€‰0.31).

Discussion

The primary aim of the current study was to examine whether more baseline ACC activity during attentional control predicted symptom improvement over the course of CBT or ST for patients with SAD or MDD. Behavioral findings showed the manipulation of task difficulty was effective as accuracy was higher and reaction times faster for low perceptual load (i.e., high emotional interference) relative to high perceptual load (i.e., low emotional interference) conditions. In addition, results revealed greater baseline ACC activation to threat (>neutral) face distractors under low, but not high, the perceptual load was associated with a steeper trajectory of symptom improvement (i.e., rate of change) regardless of principal diagnosis or treatment modality. Therefore, findings partially support hypotheses.

Our hypothesis that more baseline ACC activity to threat distractors would predict greater symptom improvement over the course of psychotherapy was supported. Findings are consistent with previous SAD studies [14,15,16] and evidence that ACC as a predictor of psychotherapy outcome extends to MDD highlights the possibility that ACC during attentional control may be a transdiagnostic biomarker of treatment response. Findings build upon an increasing body of literature that focuses on neural correlates of treatment response across disorders, rather than within categorically defined disorders [55, 56]. It will be important for future studies to determine if our preliminary results replicate in more diagnostically heterogenous samples.

While the direction of activity was expected, the specific ACC sub-region and its relation to perceptual load was not supported, as greater rate of symptom improvement was predicted by more rACC, but not dACC, activity to threat distractors under low, but not high, perceptual load. This may in part be due to methodological differences, as other prior studies examined dACC response in the context of task-irrelevant numbers [16] or used a whole-brain regression approach to identify regions during emotional interference that foretold pre-to-post symptom improvement [14, 15]. In the current study, we focused on a priori ROIs, examined trajectories of symptom change, included ST for comparison, and included patients with MDD. That said, current findings are not entirely unexpected, as links between rACC response to threat distractors and participant characteristics (e.g., trait anxiety, behavioral inhibition, self-reported attentional control) have been observed under low, but not high, perceptual load [33,34,35], suggesting the high emotional interference condition may be sensitive to individual differences. Furthermore, the specificity of findings to rACC, rather than dACC, under low load is not surprising, given rACCā€™s role in resolving emotional conflict [11,12,13]. Larger, well-powered studies are needed to replicate and confirm the reliability of findings.

Post-hoc analysis suggests the observed rACC effect was driven by sgACC. The sgACC is strongly connected with regions important for emotion processing (e.g., amygdala) and is known to engage during emotional conflict [11, 12, 57]. Subgenual ACC engagement may reflect top-down inhibition of amygdala response to task-irrelevant emotional stimuli or bottom-up amygdala habituation to task-irrelevant stimuli leading to sgACC activation (i.e., ā€˜disinhibitedā€™ sgACC response; [57]). Although further study is needed to clarify the underpinnings of sgACC as a predictor of symptom change, evidence of sgACC as a biomarker is consistent with studies showing sgACC activation during rest [24, 25] or self-referential emotion processing [26, 27] corresponds with CBT response in depressed patients. Methodological differences across studies do not permit comparison with the current study and likely contribute to differences in the directionality of sgACC findings. Even so, our results indicate sgACC activity during emotional interference may be a transdiagnostic marker of treatment outcome.

We expected ACC activity would be predict CBT, not ST, outcome. However, there was no effect of treatment arm on symptom trajectories and no interaction with baseline ACC activity. Current findings bring up questions regarding mechanisms of change for CBT. CBT-specific strategies aim to improve patientsā€™ emotion regulation strategies [9] and CBT is hypothesized to exert its effect via strengthening top-down regulation [10]. However, meta-analyses show post-treatment symptom improvement in CBT is comparable to ST [58], though CBT exhibits superior long-term efficacy [59, 60], highlighting the possibility that CBT may contribute to therapeutic gains via both common factors shared across psychotherapies (e.g., therapeutic alliance, treatment adherence) and CBT-specific techniques (e.g., emotion regulation skills) [61]. Accordingly, our findings also suggest that CBT may, at least in part, contribute to symptom improvement via common factors as ACC appears to interact with common factors shared across psychotherapies to predict individual differences in treatment response. This said, the current study focused on ACC as a predictor of treatment change. Future research is needed to examine whether other baseline neural markers interact with CBT-specific strategies to predict symptom improvement.

The current study is not without limitations, which should be acknowledged. One limitation is the utilization of neutral faces as a ā€˜baselineā€™ comparison, given increasing evidence that emotionally ambiguous faces are not truly ā€˜neutralā€™, but elicit aberrant ACC activation in patients [62]. Also, the lack of a waitlist control group or pharmacotherapy arm precludes firm conclusions from being drawn regarding the extent to which the relation between pre-treatment ACC activation and symptom improvement can be specifically ascribed to psychotherapy response. Therefore, conclusions should remain tentative pending replication in studies utilizing such controls. Findings may not generalize to patients with other internalizing conditions. Patients with MDD were older than those with SAD; although age was covaried in analysis and there was no evidence of an age effect, we cannot rule out the potential that age may have impacted findings. Depression severity was mild overall; therefore, results may not generalize to a severely depressed population. Also, non-treatment completers had less baseline depression than completers suggesting low depression levels may have factored into attrition. Furthermore, task-irrelevant positive faces were not included, which may have reduced our ability to detect diagnostic-specific effects. In addition, the study focused on ACC; it is possible that baseline activity in other brain regions predict treatment outcome. Lastly, the lack of a healthy comparator group in the current study precludes conclusions from being drawn regarding whether baseline ACC activation connotes aberrant response.

Despite important limitations, preliminary findings indicate more rACC/sgACC engagement during threat interference under low, but not high, perceptual load predicts symptom improvement over the course of psychotherapy in SAD and MDD, pointing to a transdiagnostic neural marker of treatment outcome. Though current results suggest this pattern of activation may predict positive response to psychotherapy in general, future studies with larger sample sizes and control groups are needed to confirm the predictive utility of this biomarker as general- versus modality-specific marker of treatment response. Finally, the objective of the current study was to contribute to theory regarding transdiagnostic neural predictors of psychotherapy outcome in light of gaps in the literature and previous work that point to the relevance of ACC as a predictor of treatment outcome. However, it will be important that future studies take a data-driven approach (e.g., machine learning) in large samples to determine if ACC predicts treatment outcome at the single-subject level.