Hostname: page-component-848d4c4894-xfwgj Total loading time: 0 Render date: 2024-07-02T19:04:32.045Z Has data issue: false hasContentIssue false

A core component of psychological therapy causes adaptive changes in computational learning mechanisms

Published online by Cambridge University Press:  08 June 2023

Quentin Dercon*
Affiliation:
MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK UCL Institute of Mental Health, University College London, London, UK
Sara Z. Mehrhof
Affiliation:
MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
Timothy R. Sandhu
Affiliation:
MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK Department of Psychology, University of Cambridge, Cambridge, UK
Caitlin Hitchcock
Affiliation:
MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK Melbourne School of Psychological Sciences, University of Melbourne, Melbourne, Australia
Rebecca P. Lawson
Affiliation:
MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK Department of Psychology, University of Cambridge, Cambridge, UK
Diego A. Pizzagalli
Affiliation:
Department of Psychiatry, Harvard Medical School, Boston, MA, USA Center for Depression, Anxiety, and Stress Research, McLean Hospital, Belmont, MA, USA
Tim Dalgleish
Affiliation:
MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK Cambridgeshire and Peterborough NHS Foundation Trust, Cambridgeshire, UK
Camilla L. Nord
Affiliation:
MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
*
Corresponding author: Quentin Dercon; Email: [email protected]
Rights & Permissions [Opens in a new window]

Abstract

Background

Cognitive distancing is an emotion regulation strategy commonly used in psychological treatment of various mental health disorders, but its therapeutic mechanisms are unknown.

Methods

935 participants completed an online reinforcement learning task involving choices between pairs of symbols with differing reward contingencies. Half (49.1%) of the sample was randomised to a cognitive self-distancing intervention and were trained to regulate or ‘take a step back’ from their emotional response to feedback throughout. Established computational (Q-learning) models were then fit to individuals' choices to derive reinforcement learning parameters capturing clarity of choice values (inverse temperature) and their sensitivity to positive and negative feedback (learning rates).

Results

Cognitive distancing improved task performance, including when participants were later tested on novel combinations of symbols without feedback. Group differences in computational model-derived parameters revealed that cognitive distancing resulted in clearer representations of option values (estimated 0.17 higher inverse temperatures). Simultaneously, distancing caused increased sensitivity to negative feedback (estimated 19% higher loss learning rates). Exploratory analyses suggested this resulted from an evolving shift in strategy by distanced participants: initially, choices were more determined by expected value differences between symbols, but as the task progressed, they became more sensitive to negative feedback, with evidence for a difference strongest by the end of training.

Conclusions

Adaptive effects on the computations that underlie learning from reward and loss may explain the therapeutic benefits of cognitive distancing. Over time and with practice, cognitive distancing may improve symptoms of mental health disorders by promoting more effective engagement with negative information.

Type
Original Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
Copyright © The Author(s), 2023. Published by Cambridge University Press

Introduction

Emotion regulation difficulties occur across psychiatric disorders, improve following effective psychological treatment (Sloan et al., Reference Sloan, Hall, Moulding, Bryce, Mildred and Staiger2017), and may predict treatment response (Siegle, Carter, & Thase, Reference Siegle, Carter and Thase2006). Cognitive distancing is a core therapeutic strategy to facilitate emotion regulation, forming a central component of cognitive behavioural therapy (CBT) (Papa, Boland, & Sewell, Reference Papa, Boland, Sewell, O'Donohue and Fisher2012), mindfulness-based cognitive therapy (Hamidian, Omidi, Mousavinasab, & Naziri, Reference Hamidian, Omidi, Mousavinasab and Naziri2016), and dialectical behaviour therapy (Neacsiu, Eberle, Kramer, Wiesmann, & Linehan, Reference Neacsiu, Eberle, Kramer, Wiesmann and Linehan2014), among others. When cognitive distancing, patients are encouraged to view negative thoughts from afar, such as by ‘adopt[ing] the position of a neutral observer’ (Staudinger, Erk, Abler, & Walter, Reference Staudinger, Erk, Abler and Walter2009). With practice, distancing promotes disengagement from intense emotions in favour of a more experiential perspective, reducing distress (Mennin, Ellard, Fresco, & Gross, Reference Mennin, Ellard, Fresco and Gross2013) and depressed thoughts (Kross, Gard, Deldin, Clifton, & Ayduk, Reference Kross, Gard, Deldin, Clifton and Ayduk2012).

Despite major advances in understanding the computations underpinning pharmacological treatments in psychiatry, whether these mechanisms parallel those involved in psychological treatment is unknown. Reward learning is compromised in numerous psychiatric disorders (Halahakoon et al., Reference Halahakoon, Kieslich, O'Driscoll, Nair, Lewis and Roiser2020; Huys, Pizzagalli, Bogdan, & Dayan, Reference Huys, Pizzagalli, Bogdan and Dayan2013; Maia & Frank, Reference Maia and Frank2011), and is a target of pharmacological interventions for conditions including depression (Hales, Houghton, & Robinson, Reference Hales, Houghton and Robinson2017), schizophrenia (Insel et al., Reference Insel, Reinen, Weber, Wager, Jarskog, Shohamy and Smith2014), and bipolar disorder (Volman et al., Reference Volman, Pringle, Verhagen, Browning, Cowen and Harmer2020). Reward learning may also be a target of psychological interventions (Delgado, Gillis, & Phelps, Reference Delgado, Gillis and Phelps2008). CBT has been shown to directly affect reward learning (Brown et al., Reference Brown, Zhu, Solway, Wang, McCurry, King-Casas and Chiu2021), altering how patients assign value to actions and the extent to which these are updated by different environmental signals. These computations are formalised in a reinforcement learning framework as expected values (EVs) and prediction errors (PEs) respectively (Sutton & Barto, Reference Sutton and Barto1998).

Cognitive distancing has been specifically shown to affect midbrain reward representations through top-down prefrontal cortex input (Staudinger et al., Reference Staudinger, Erk, Abler and Walter2009; Staudinger, Erk, & Walter, Reference Staudinger, Erk and Walter2011). However, the computational underpinnings of these representations (Huys, Maia, & Frank, Reference Huys, Maia and Frank2016) are not known. For example, does cognitive distancing alter the extent to which PEs update the EV of future decisions (learning rate), or does distancing alter the extent to which these EVs drive decisions (inverse temperature)? Here, in a large online sample (n = 995) broadly representative of the UK population, we test whether and how cognitive distancing alters reinforcement learning, using a probabilistic selection task (PST) (Frank, Seeberger, & O'Reilly, Reference Frank, Seeberger and O'Reilly2004; Frank, Moustafa, Haughey, Curran, & Hutchison, Reference Frank, Moustafa, Haughey, Curran and Hutchison2007) combined with established computational models to decompose components of participant behaviour. Our study methods and analyses were preregistered (https://osf.io/fd4qu).

Methods and materials

Recruitment and study design

Participants were recruited on Prolific (Palan & Schitter, Reference Palan and Schitter2018) over three weeks in April–May 2021. Prolific pre-screeners were used to recruit batches of male and female UK nationals with and without a self-reported prior diagnosis of a psychiatric disorder across five age-groups (18–24, 25–34, 35–44, 45–54, and 55+), with target numbers for each of the twenty batches calculated based on UK population data (Office of National Statistics, 2016; Stansfeld et al., Reference Stansfeld, Clark, Bebbington, King, Jenkins, Hinchliffe, McManus, Bebbington, Jenkins and Brugha2016) (see online Supplementary Methods). The online study was written in JavaScript using the jsPsych library (de Leeuw, Reference de Leeuw2015), and consisted of a reinforcement learning task (Fig. 1a), followed by a working memory task (digit span; see online Supplementary Methods) and a psychiatric questionnaire battery. A total of 995 participants completed all study components. Participants were paid a fixed rate of £9 (approximately £6/h), plus a bonus contingent on task performance to encourage engagement (£1 if in the top 30% of points or £2 in the top 10%). The study was approved by the University of Cambridge Human Biology Research Ethics Committee (HBREC.2020.40) and jointly sponsored by the University of Cambridge and Cambridge University Hospitals NHS Foundation Trust (IRAS ID 289980). All participants provided written informed consent through an online form, in line with approved University of Cambridge Human Biology Research Ethics Committee procedures for online studies.

Figure 1. Reinforcement learning task design, transdiagnostic factor derivation, and comparison of factor scores to previous samples. a. The probabilistic selection task (PST) in the present study consisted of six blocks of sixty trials (training phase) where participants were instructed to choose between Hiragana characters presented as three pairs (AB, CD, EF; twenty of each per block), and given feedback (‘Correct!’ or ‘Incorrect.’). One character in each of the pairs was consistently more likely to be correct (reward probabilities of 0.8/0.2, 0.7/0.3, and 0.6/0.4 for A/B, C/D, and E/F respectively. 497 participants (49.9%) were randomised to a self-distancing intervention, and additionally received a prompt to ‘Distance yourself…’ with the fixation cross at the start of each training trial. The training phase was followed by a sixty-trial test phase without feedback where the twelve other possible character combinations were added (i.e. four of each of the fifteen pairs). All participants saw the same six characters, but the pairs themselves were randomised for each participant, and the order of the pairs was counterbalanced across trials. One of three affect questions was asked after each trial, with unlimited time to answer. b. Multi-target lasso regression with five-fold cross-validation was used to predict the three transdiagnostic dimension factor scores from different subsets of the 209 questions in the original dataset (Gillan et al., Reference Gillan, Kosinski, Whelan, Phelps and Daw2016, study 2). 78 questions were found to predict the three factor scores with high predictive accuracy (colours correspond to heatmap labels). c. Heatmap of five-fold cross-validated multi-target lasso regression coefficient weights for each of the included 78 questions across eight psychiatric questionnaires used to predict the three transdiagnostic symptom dimensions (i.e. weights for all other questions fixed at zero; see online Supplementary Methods for full details). d. Comparison between the factor score distributions for the n = 935 non-excluded participants in the present study (darker colours), and those previously obtained by Gillan et al. (Reference Gillan, Kosinski, Whelan, Phelps and Daw2016) in n = 1413 participants (lighter colours). Inset plots show the predictive validity of the subset of 78 questions in predicting these scores in the original dataset.

Exclusion criteria

Exclusion criteria included self-reported diagnosis of a neurological disorder (n = 18); English proficiency below B2 level (good command or working knowledge; n = 3); and a digit span of 0 (n = 5). Based on recent recommendations (Zorowitz, Solis, Niv, & Bennett, Reference Zorowitz, Solis, Niv and Bennett2021), two harder catch questions were included in the psychiatric questionnaire battery (e.g. ‘In the past week, I would (have) avoided… Eating mouldy food’; expected answer ‘Usually’), in addition to two standard catch questions (e.g. ‘Please answer ‘Strongly Agree’ to this question.’). Participants were excluded if they got one of the standard questions wrong (n = 16), or both harder questions wrong (n = 20).

Reinforcement learning task

In the present study, the PST (Frank et al., Reference Frank, Seeberger and O'Reilly2004, Reference Frank, Moustafa, Haughey, Curran and Hutchison2007) consisted of a training (six blocks of sixty trials) and test phase (sixty trials; Fig. 1a). Prior to starting the task, all participants viewed a one-minute instructional video, and had to correctly answer a multiple-choice question on these instructions. Participants randomised to the control arm (n = 498) then proceeded to the first training trial, while participants randomised to the intervention (n = 497) were also shown a second ninety-second video (see ‘Cognitive distancing manipulation' below) before advancing. On each trial, participants were presented with one of three pairs of Japanese Hiragana characters (there were exactly twenty of each pair per block) and instructed to choose the left or right character via key press. The symbols within each stimulus pair were associated with different chances of reward: symbol ‘A’ was correct 80% of the time, and ‘B’ 20% of the time, with respective probabilities of 0.7/0.3 and 0.6/0.4 for the ‘C’/‘D’ and ‘E’/‘F’ pairs. Participants were instructed to pick the symbol they felt was most likely correct. After receiving feedback (‘Correct!’ or ‘Incorrect.’, plus twenty-five points if the answer was correct), participants were asked to indicate their current feelings (0–100) with respect to one of three affect adjectives (happy/confident/engaged), with unlimited time to answer. The training phase was followed by a sixty-trial test phase without feedback, which included the twelve other possible character combinations. All participants saw the same six characters, but the pairs themselves were randomised for each participant, and the order of the pairs was counterbalanced across trials. Twelve participants (1.2%) reported prior familiarity with Hiragana characters.

Cognitive distancing manipulation

Half of the sample (n = 497) was randomly allocated to the cognitive distancing manipulation. After viewing the instructional video and correctly answering the multiple-choice question on these instructions as detailed above, participants randomised to the cognitive distancing intervention were also shown a second ninety-second video. In this video, the concept of cognitive distancing was introduced as ‘the ability to take mental ‘step back’ from your immediate reactions to events, and view these events from a broader, calmer, and less emotional perspective’, and participants were suggested to practice it by trying to ‘imagine yourself as an external observer, watching yourself perform the task from a distance.’ It was then explained that they should still try to win as many points as possible, but that whenever they had an emotional response to a trial, they should ‘try to distance yourself from your immediate reaction, by taking a step back from how you are feeling.’ Following these instructions (see online Supplementary Methods for full text), the task proceeded in the same manner to that taken by non-distanced participants, except that they received an additional reminder to ‘Distance yourself…’ with the fixation cross that preceded each training trial (Fig. 1a). In the test phase, this reminder was omitted, as there was no feedback.

Psychiatric questionnaires and transdiagnostic factor derivation

After completing the reinforcement learning task, all participants completed a psychiatric questionnaire battery. Using an approach recently termed computational factor modelling (Wise, Robinson, & Gillan, Reference Wise, Robinson and Gillan2022), we estimated scores for participants on each of three well-characterised transdiagnostic symptom dimensions – anxiety/depression, compulsive behaviour, and social withdrawal (Gillan, Kosinski, Whelan, Phelps, & Daw, Reference Gillan, Kosinski, Whelan, Phelps and Daw2016) – which could then be related to learning parameters. Though these scores were originally derived from 209 individual questionnaire items, we used a method described previously (Wise & Dolan, Reference Wise and Dolan2020) to identify a subset of items which could predict scores in the original dataset (Gillan et al., Reference Gillan, Kosinski, Whelan, Phelps and Daw2016) with high accuracy. Specifically, a multi-target lasso regression model was trained on the raw question ratings of Gillan et al.'s original dataset (study 2) with the three factor scores as the responses, with differing numbers of question coefficients fixed to zero. Five-fold cross-validation was then used to assess the predictive accuracy of these question subsets (Wise & Dolan, Reference Wise and Dolan2020). We found that 78 questions from eight different questionnaires (online Supplementary Table S1), represented a good compromise between number of questions and predictive accuracy (R 2 ≥ 0.9 for all three dimensions; Fig. 1b). Coefficient estimates from the multi-target lasso regression model fit to the original dataset (Fig. 1c) were then used to predict factor scores on each of the transdiagnostic symptom dimensions for all individuals in our sample, based on their answers to the included 78 questions (Fig. 1d).

Computational modelling

Q-learning models

Model-free reinforcement learning in the PST is commonly modelled using Q-learning models (Watkins & Dayan, Reference Watkins and Dayan1992) with single or dual learning rates (α) (Frank et al., Reference Frank, Moustafa, Haughey, Curran and Hutchison2007). The key distinction is that, in the dual learning rate model, EVs (termed Q-values) are assumed to update at different rates depending on whether the feedback is negative (i.e. reward r t = 0, so PE δ t < 0; termed α loss) or positive (i.e. reward r t = 1, so PE δ t ≥ 0; termed α reward). Both models assume that EVs assigned to the stimuli in each pair are updated in response to feedback; higher learning rate parameter values (α) indicate increased sensitivity of choices to recent feedback (single learning rate), or in dual learning rate models, differential sensitivity to positive or negative feedback (Frank et al., Reference Frank, Moustafa, Haughey, Curran and Hutchison2007) (reward or loss learning rate; α reward or α loss). These EVs are converted to probabilities of choosing one symbol over another via a softmax logistic function, with the extent to which differences in Q-values determined choices weighted by an inverse temperature parameter (β) (higher values suggest choices more determined by EV differences). In addition to modelling ‘training alone’, these models can also be extended to include test trials (‘training-plus-test’). In the absence of feedback, EVs are assumed to be fixed at the end of training, but in these models, the log probability density was additionally incremented on test phase choices. As such, parameter values from these models are interpreted as those which best reflect choices across training plus those made in the subsequent test phase. See online Supplementary Methods for further details.

Model fitting and outcome generalised linear models

Models were fit in a hierarchical Bayesian manner (Ahn, Krawitz, Kim, Busemeyer, & Brown, Reference Ahn, Krawitz, Kim, Busemeyer and Brown2011; Gelman et al., Reference Gelman, Carlin, Stern, Dunson, Vehtari and Rubin2013) via CmdStan (Stan Development Team, Reference Stan Development Team2021), separately for distanced and non-distanced participants (Valton, Wise, & Robinson, Reference Valton, Wise and Robinson2020). Stan model code was adapted from the hBayesDM (Ahn, Haines, & Zhang, Reference Ahn, Haines and Zhang2017) repository. Models were fit using Markov chain Monte Carlo (MCMC) with four parallel chains, and 4000 warm-up plus 20 000 sampling draws per chain. Numerical and visual MCMC diagnostics, including expected sample size (ESS) and split R-hat (Vehtari, Gelman, Simpson, Carpenter, & Burkner, Reference Vehtari, Gelman, Simpson, Carpenter and Burkner2021) were used to assess chain mixing and convergence. Individuals' parameters were omitted from subsequent analyses if they had either split R-hat ⩾ 1.1 or bulk ESS < 100 for any parameter (this applied to no more than two individuals per fit). Model posterior predictions were also compared to observed choices for all individuals, and the ability of all models to recover known parameter values from simulated data was also assessed (see online Supplementary Methods). Models were compared using two numerical metrics of out-of-sample predictive accuracy: expected log posterior density (ELPD; higher is better), and the leave-one-out information criterion (LOOIC; lower is better) (Vehtari, Gelman, & Gabry, Reference Vehtari, Gelman and Gabry2016).

To quantify the evidence for associations between model parameters and cognitive distancing or transdiagnostic factor scores, Bayesian GLMs were fit via CmdStan (Stan Development Team, Reference Stan Development Team2021), using Stan models and priors from the rstanarm R package (Goodrich, Gabry, Ali, & Brilleman, Reference Goodrich, Gabry, Ali and Brilleman2020). Due to positively skewed distributions (Fig. 2b, c), the association between learning rate(s) and cognitive distancing was assessed using gamma family GLMs with log link functions, while the association between the inverse temperatures and outcomes was assessed using standard linear regression. All models were adjusted for age, gender (male or female; imputed as natal sex for non-binary individuals given low numbers), and digit span, with models relating parameters to transdiagnostic factor scores additionally adjusted for distancing status.

Figure 2. Model comparison, parameter distributions, and raw training and test phase performance. a. Difference in numerical fit metrics between the dual and single learning rate models fit to training data alone, or training-plus-test, by distancing group. ELPD is the expected log posterior density (presented here as the difference between the dual and single learning rate models, where positive differences indicate a better model), and LOOIC is the leave-one-out information criterion (lower indicates a better model). b–c. Distributions of individual-level posterior means for learning parameters from the fits to training (b) and test (c) data. d. Raw training performance (cumulative probability of choosing the higher-probability symbol A/C/E), by group and stimulus pair, lagged by twenty trials (i.e. block-lagged, as each pair is presented twenty times per block. e. Raw test phase performance (% correct, where ‘correct’ is choosing the option in each pair which was most often correct during training) by test type, plus test phase performance on individual training pairs and novel pairs including symbols C or E.

An exploratory analysis was then run to assess at what stage differences in training performance emerged between distanced and non-distanced participants. To do so, the dual learning rate Q-learning model was fit separately in each group as before to trials from increasing numbers of training blocks (i.e. block one, block one to two, block one to three, etc.). The inclusion of all previous trials in each model is necessary as the beginning of training is the only time we can assume Q-values for all symbols to be zero. Subsequently, Bayesian GLMs were run to assess the evidence for group differences in learning parameters at each stage as before.

Results

After applying exclusion criteria, a total of 935 participants (49.1% distanced) were included in analyses (Table 1). Dual learning rate models had consistently higher ELPD and lower LOOIC than corresponding single learning rate models (Fig. 2a), indicating better estimated out-of-sample predictive accuracy, though we report results from all models, as they are largely comparable. Across all non-excluded participants, mean compulsive behaviour factor scores estimated from psychiatric questionnaires were comparable to a similarly-large online sample (2.05 v. 2.00; t(2286.1) = 1.24, p = 0.22) (Gillan et al., Reference Gillan, Kosinski, Whelan, Phelps and Daw2016); mean scores for the anxiety/depression factor were slightly higher (2.76 v. 2.59; t(2287) = 5.04, p < 0.001); and mean scores for social withdrawal markedly higher (1.41 v. 1.06; t(2225.7) = 10.2, p < 0.001) (Fig. 1d).

Table 1. Demographic characteristics of the sample, by group.

Group differences in raw training and test phase performance, and affect ratings

We assessed whether cognitive distancing affected learning during the task. When comparing the performance of the groups over the course of training, quantified by the mean proportion of times participants chose the symbol most likely to be correct in each pair in each training block, we found strong evidence that the groups differed in accuracy across training trials (multivariate block-stratified Kruskal–Wallis $\lambda _3^2$ = 18.6, p = 0.0003), with distanced individuals appearing to perform better (Fig. 2d). Though post-hoc testing indicated some weak evidence that distanced individuals performed on average 1.40% better on the easiest ‘AB’ pairs across training blocks (reward probabilities 0.8/0.2; block-stratified Kruskal–Wallis $\lambda _1^2$ = 4.17, Holm–Bonferroni corrected p = 0.041), the evidence was stronger for more ‘difficult’ pairs, where the symbols' reward probabilities were closer together (Fig. 2d, e). Specifically, distanced participants performed on average 2.65% better on medium-difficulty ‘CD’ trials (reward probabilities 0.7/0.3; block-stratified Kruskal–Wallis $\lambda _1^2$ = 11.1, Holm–Bonferroni corrected p = 0.003), and on average 3.09% better on the most difficult ‘EF’ pairs in each block (reward probabilities 0.6/0.4; block-stratified Kruskal–Wallis $\lambda _1^2$ = 11.2, Holm–Bonferroni corrected p = 0.003).

In the test phase, there was only trend-level evidence that the groups differed in performance over the four groups of stimuli (multivariate Kruskal–Wallis $\lambda _2^4$ = 7.73, p = 0.10). Post-hoc testing suggested that while there was little evidence of any group difference in performance on test pairs including the most-likely correct ‘A’ symbol or the least-likely ‘B’ symbol (‘chooseA’ or ‘avoidB’ respectively), there was some weak evidence that distanced individuals also performed on average 5.0% better when tested on difficult novel pairs (those excluding the symbols most or least associated with reward), though the evidence for this was trend-level after multiple comparisons adjustment (Kruskal–Wallis $\lambda _1^2$ = 4.87, Holm–Bonferroni corrected p = 0.11; Fig. 2e).

Distancing also had subtle effects on self-reported affect throughout training. Linear mixed effects models were used to relate average affect ratings per block to group, block, and their interaction (adjusting for age, gender, and digit span). All affect ratings decreased over time, but there was evidence for a block-by-distancing interaction for happiness and engagement (but not confidence), such that both happiness and engagement declined by an estimated 0.44 and 0.41 points (out of 100) less per block in distanced participants (${{\hat{\beta }} \over {SE( {\hat{\beta }} ) }}$ = 2.95, 95% CI (0.14–0.73); and ${{\hat{\beta }} \over {SE( {\hat{\beta }} ) }}$ = 2.22, 95% CI (0.05–0.78) for happiness and engagement respectively).

Associations between learning parameters and transdiagnostic symptom dimensions

Individual-level posterior mean parameter values from single or dual learning rate Q-learning models fit to training alone or training-plus-test (parameters from fits to training-plus-test denoted prime) were related to the three transdiagnostic factors using Bayesian GLMs adjusted for age, gender, digit span and distancing status.

Results from the GLMs suggested there was limited evidence for an association between any Q-learning model parameter and scores on either the anxiety/depression or social withdrawal factors (Fig. 3a, b). An exception was the negative association between the anxiety/depression factor score and the inverse temperature parameter β (from the dual learning rate model fit to training alone), such that unit increases in this score were associated with lower β, indicating lower sensitivity to value differences between symbols in those scoring higher on this transdiagnostic symptom dimension (Fig. 3a; posterior mean coefficient = −0.08). Still, the evidence for this was very weak, with the 95% highest density interval (HDI) wide and including zero (−0.22, 0.05).

Figure 3. Associations between reinforcement learning parameters and transdiagnostic psychiatric symptom dimensions. a–d. Coefficient posterior distributions from Bayesian GLMs (adjusted for age, gender, digit span and distancing status) reflecting the estimated percentage change in the learning rate parameter or the estimated mean change in the inverse temperature for a unit increase in anxiety/depression (a), social withdrawal (b), and compulsive behaviour (c–d) transdiagnostic factor scores. Parameters were derived from single (only compulsive behaviour presented here, C) or dual learning rate (a–b & d) Q-learning models fit to training alone or training-plus-test (parameters from fits to training-plus-test denoted prime). In all plots, boxplot boxes denote 95% HDI, and lines denote 99% HDI.

There was also little evidence of any association between β and anxiety/depression from models fit to training-plus-test, nor of associations between learning rate parameters from any model and any of the three factor scores. That said, we did find evidence that increases in compulsive behaviour factor scores were associated with lower inverse temperature values (β and β ; Fig. 3c, d). The strongest evidence for this came from the single learning rate models (Fig. 3c) [β: 95% HDI = (−0.33, −0.03); β : 95% HDI = (−0.28, 0.004)]; evidence from the dual learning rate model was consistent but weaker (Fig. 3d) [β: 95% HDI = (−0.26, 0.002); β : 95% HDI = (−0.22, 0.05)].

Associations between learning parameters and cognitive distancing

To quantify potential differences in Q-learning model parameters between groups, we compared individual-level posterior means between distanced and non-distanced individuals using adjusted Bayesian GLMs.

There was weak but consistent evidence that distanced participants had higher inverse temperature values (β), suggesting less stochastic choices throughout training, with the mean difference estimated at 0.17 from both models [95% HDI = (−0.05, 0.39) and (−0.01, 0.37) for single and dual learning rate models, respectively]. Distanced participants also showed increased sensitivity to recent feedback, as indicated by estimated 15.3% higher α values [single learning rate model; 95% HDI for multiplier = (1.01, 1.32); Fig. 3a]. Notably, the dual learning rate model showed this increased sensitivity was specific to negative feedback trials (Fig. 3b), affecting loss learning rates [estimated multiplier for α loss = 1.19, 95% HDI = (1.0002, 1.42)] but not positive learning rates [estimated multiplier for α reward = 1.004, 95% HDI = (0.90, 1.13)].

In line with results from models fit to training alone, we found weak evidence for differences in inverse temperature parameters from models fit to training-plus-test (β ), with distanced participants having an estimated 0.16 [95% HDI = (−0.05, 0.37)] and 0.18 higher β [95% HDI = (−0.02, 0.38)] values from single and dual learning rate models respectively. There was also evidence for differences in sensitivity to feedback from the single learning rate model additionally fit to test phase choices [estimated multiplier for α  = 1.15, 95% HDI = (0.99, 1.33); Fig. 3a]. Results from the dual learning rate model confirmed this specificity to negative feedback, with distanced participants estimated to have 24.6% higher loss learning rates at end of training [estimated multiplier for $\alpha _{loss}^{\prime}$ = 1.25; 95% HDI = (1.05, 1.48); Fig. 3b].

Temporal emergence of differences in learning parameters

Although we found consistent evidence of increases in loss learning rates in distanced individuals, particularly towards the end of training, there was no evidence of a group difference in test performance on pairs that included the ‘B’ symbol (Kruskal–Wallis $\lambda _1^2$ = 0.17, Holm–Bonferroni corrected p = 1; Fig. 2e) – the symbol with the lowest reward probability – which would be expected in the context of a higher loss learning rate (Frank et al., Reference Frank, Moustafa, Haughey, Curran and Hutchison2007). To investigate this, we fit dual learning rate models to increasing numbers of training blocks. As each subsequent model includes all previous trials, the results from these models cannot be interpreted directly as the best-fitting parameters for each participant over individual training blocks. However, given that learning parameters are assumed fixed in each model, estimates from models including later trials should be indicative of the underlying (dynamic) parameter values later in training. Indeed, as would be expected, reward and loss learning rate estimates decreased in both groups with the inclusion of later trials, while inverse temperatures increased (Fig. 4c).

Figure 4. Model-derived comparisons between distanced and non-distanced participants. a–b. Coefficient posterior distributions from Bayesian GLMs (adjusted for age, gender, and digit span) reflecting the estimated percentage change in the learning rate parameter α (a) or α reward/α loss (b), and the estimated mean change in the inverse temperature β, comparing distanced and non-distanced participants. Parameters were estimated from Q-learning models with single (a) or dual (b) learning rates, fit to training alone or training-plus-test (parameters from fits to training-plus-test denoted with prime). c. Individual-level posterior mean parameter estimates for α reward, α loss, and β, for models including increasing numbers of training blocks. d. Parameter differences between distanced and non-distanced participants, estimated from dual learning rate models fit to increasing numbers of training blocks (sixty trials per block). In all plots, boxplot boxes denote 95% HDIs, and lines denote 99% HDIs.

We then compared the individual-level parameter values in the distanced and non-distanced groups with adjusted GLMs as before. We found that group differences in loss learning rates emerged as the task progressed (95% HDI for its multiplier between groups included 0 for all fits to fewer than all six blocks: Fig. 4d), with distanced participants maintaining a higher loss sensitivity relative to non-distanced participants as the task progressed (Fig. 4c). In contrast, there was strong evidence that distanced participants had consistently higher inverse temperature parameter values throughout training, with all 95% HDI excluding 0 except for the final fit to all six blocks, and some evidence that they had 10.9% lower positive learning rates over the first three training blocks [i.e. block one to three; estimated multiplier for α reward = 0.89; 95% HDI = (0.81, 0.98); Fig. 4d].

Discussion

Learning to predict rewards and avoid punishments is essential to adaptive behaviour. Disruptions in this fundamental process have been found across psychiatric disorders (Lee, Reference Lee2013; Maia & Frank, Reference Maia and Frank2011) and may be a target of psychological therapy (Brown et al., Reference Brown, Zhu, Solway, Wang, McCurry, King-Casas and Chiu2021). In a large online sample representative of the UK population in terms of age, gender, and psychiatric history, we found that a common psychotherapeutic strategy, cognitive distancing, enhanced performance in a reinforcement learning task. Our results indicate two computational mechanisms may underpin the therapeutic effects of cognitive distancing: enhanced integration, potentially coupled with subsequent exploitation of previously reinforced choices; and, with time or practice, an adaptively enhanced ability to update decisions following negative feedback.

We found that cognitive distancing was associated with heightened inverse temperatures, from the start of the task. The magnitude and evidence for group differences waned over the course of training, though we note that this is consistent with the observed increases in evidence for higher learning rates in the distancing group. Higher inverse temperatures are indicative of clearer representations of differences in true (latent) values between choice options (Pedersen & Frank, Reference Pedersen and Frank2020), and may have enabled more deterministic choosing of the ‘better’ option in each pair, over those with less reward-certainty. Higher inverse temperature values can also be interpreted as a bias towards exploitation of current task knowledge, v. exploration of uncertain options (Pedersen & Frank, Reference Pedersen and Frank2020). Notably, a reduction in inverse temperature in patients with major depressive disorder has been found in previous studies using similar tasks and computational models (Huys et al., Reference Huys, Pizzagalli, Bogdan and Dayan2013; Pike & Robinson, Reference Pike and Robinson2022), and we did find that higher levels of anxiety/depression symptomology were associated with slightly lower inverse temperatures, albeit with very weak evidence (Fig. 3a). Taken with evidence that cognitive distancing is an effective therapeutic strategy for depression (Kross et al., Reference Kross, Gard, Deldin, Clifton and Ayduk2012), our results tentatively suggest that cognitive distancing interventions may improve depressive symptoms by increasing exploitation of rewarding outcomes. This is a key hypothesis emerging from our results, which will require confirmatory evidence from future longitudinal studies.

Distanced participants also adaptively altered reward and loss learning throughout the task, possibly adjusting to changing task dynamics. Initially, higher inverse temperatures compared to non-distanced participants suggest they more quickly developed preferences for certain symbols. However, deterministic preferences may not be the best strategy for the entire task: initial impressions can be wrong, especially for the harder-to-distinguish stimuli. Notably, towards the end of the task, distanced participants became more sensitive to negative feedback, and showed weaker evidence for inverse temperature differences, which is consistent with more exploratory behaviour later in the task. At this stage, losses may be more informative: assuming preferences are largely correct for the easier pairs, negative feedback will be rarer, and primarily experienced when choosing between the harder-to-distinguish pairs. Previous work shows that participants adjust dual learning rates independently, increasing learning rate when one type of feedback becomes more informative (Pulcu & Browning, Reference Pulcu and Browning2017). By increasing loss sensitivity, and perhaps testing out less preferable options, distanced participants may have been able to discern values of harder-to-distinguish pairs more accurately by the end of training, enabling better performance when subsequently tested on these pairs.

A higher loss learning rate seems therapeutically counterintuitive, given several mental health disorders are marked by heightened punishment sensitivity (Elliott, Sahakian, Herrod, Robbins, & Paykel, Reference Elliott, Sahakian, Herrod, Robbins and Paykel1997; Jappe et al., Reference Jappe, Frank, Shott, Rollin, Pryor, Hagman and Davis2011; Pike & Robinson, Reference Pike and Robinson2022). Critically, however, aberrant loss learning may represent a ‘catastrophic response to perceived failure’ (Elliott et al., Reference Elliott, Sahakian, Herrod, Robbins and Paykel1997; Roiser & Sahakian, Reference Roiser and Sahakian2013). In contrast, a high loss learning rate in distanced participants was accompanied by better performance. It is notable that there was limited evidence for increased learning rates in distanced relative to non-distanced participants unless later training blocks were included; indeed, there was evidence that distanced individuals had lower reward learning rates over the first half of the task. This suggests that practice may be critical to realising the beneficial effects on aspects of learning. Speculatively, our results suggest that persistent use of cognitive distancing might drive symptom change by improving patients' ability to learn from negative experiences, and apply that learning to more adaptive behaviours. Clinically, this suggests the mechanism underlying distancing may be more effective engagement with negative information, rather than reduced engagement with negative information.

Our finding of weak negative associations between inverse temperatures and compulsive behaviour factor scores is consistent with previous work reporting a (positive) association between compulsive behaviour and deficits in goal-directed control, characterised by the (in)ability to integrate information over time (Gillan et al., Reference Gillan, Kosinski, Whelan, Phelps and Daw2016). However, we only found very limited evidence for a similar negative association between inverse temperature parameters and anxiety/depression symptom scores. This result is at odds with a considerable body of literature reporting dysfunctional reinforcement learning across psychopathologies (Maia & Frank, Reference Maia and Frank2011), especially depression (Halahakoon et al., Reference Halahakoon, Kieslich, O'Driscoll, Nair, Lewis and Roiser2020). That said, most previous work has compared clinical populations to healthy controls, which may result in better detection of a true effect due to greater symptomatic delineation between groups. Alternatively, together with publication bias in the field (Halahakoon et al., Reference Halahakoon, Kieslich, O'Driscoll, Nair, Lewis and Roiser2020), true effects may be smaller than those reported by small-scale clinical studies (Gelman & Carlin, Reference Gelman and Carlin2014). Constraints on our study design and length due to its online nature may also have restricted our ability to detect associations for three reasons. Firstly, ‘loss’ trials in our task lacked actual punishment, which is experienced differently to the absence of reward (Huys et al., Reference Huys, Pizzagalli, Bogdan and Dayan2013). Secondly, we assessed symptoms through self-report questionnaires in the general population which may not be comparable to more severe clinical populations. Thirdly, we took a fully transdiagnostic approach to psychopathology (Dalgleish, Black, Johnston, & Bevan, Reference Dalgleish, Black, Johnston and Bevan2020; Wise et al., Reference Wise, Robinson and Gillan2022), which precluded us from additionally deriving categorical psychiatric measures and investigating any potential taxonic associations with learning parameters, as complete symptom questionnaires were not included for many key mental health disorders.

We note several limitations. Though motivated by discrepancies between test phase performance and preregistered computational analyses, it should be noted that the increasing-blocks analyses were exploratory, and inherently limited in that we could not estimate group differences in parameter values over individual blocks. The effects we report are generally small, perhaps reflecting the subtlety of our distancing manipulation. Note, however, that recent work finds linguistic distancing in therapy has a small but crucial clinical effect (Nook, Hull, Nock, & Somerville, Reference Nook, Hull, Nock and Somerville2022). These effects may also be partially mediated by expectancy (e.g. ‘placebo effects’), which some have argued play a central role in the beneficial effects of psychotherapy (Enck & Zipfel, Reference Enck and Zipfel2019). Without consensus as to the active ingredients of psychological therapy, the potential contribution of expectancy is a limitation of any psychological therapy (Weimer, Colloca, & Enck, Reference Weimer, Colloca and Enck2015). As noted previously, we also found only very limited evidence for often-reported negative associations between inverse temperatures and anxiety/depression psychopathology (Pike & Robinson, Reference Pike and Robinson2022), weakening the interpretation that distancing may act to resolve this. Lastly, we did not collect information on participants' previous experience of psychological therapies, preventing us from controlling for potential advantageous effects of prior exposure to distancing or similar psychotherapeutic techniques.

In this study, by using a relatively simple learning paradigm with an established computational model, we were able to measure the effects of cognitive distancing on well-understood choice behaviour. This echoes clinical reports that distancing causes an adaptive shift in depressed people's processing of negative experiences (Kross et al., Reference Kross, Gard, Deldin, Clifton and Ayduk2012). Our work demonstrates the utility of computational approaches for reverse-translational studies to illuminate the mechanisms of existing treatments, in turn enabling improved augmentation and potentially personalisation of treatment for mental health disorders.

Supplementary material

The supplementary material for this article can be found at https://doi.org/10.1017/S0033291723001587

Acknowledgments

The authors would like to thank Dr Becky Gilbert for her assistance with jsPsych programming.

Author contributions

Conceptualization, C.L.N.; Methodology, C.L.N., Q.D., S.M., T.S., R.P.L., D.A.P., T.D.; Investigation, Q.D., S.M., C.L.N.; Project Administration, Q.D., S.M., C.L.N., Writing – Original Draft, Q.D., S.M., C.L.N.; Formal Analysis, Q.D., S.M.; Software, Q.D., Writing – Review & Editing, C.H., D.A.P., T.D., T.S., R.P.L.; Funding Acquisition, C.L.N., T.D., Supervision: C.L.N., T.D.

Financial support

This study was funded by an AXA Research Fund Fellowship awarded to C.L.N. (G102329) and the Medical Research Council (MC_UU_00030/5, MC_UU_00030/12) and partly supported by the National Institute for Health Research Cambridge Biomedical Research Centre. Q.D. is supported by a Wellcome Trust PhD studentship. R.P.L. is supported by a Royal Society Wellcome Trust Henry Dale Fellowship (206691) and is a Lister Institute Prize Fellow. C.H. was supported by fellowships from the UKRI (ES/R010781/1) and Australian Research Council (DE200100043).

Conflict of interest

Over the past 3 years, D.A.P. has received consulting fees from Albrights Stonebridge Group, Boehringer Ingelheim, Compass Pathways, Concert Pharmaceuticals, Engrail Therapeutics, Neumora Therapeutics (former BlackThorn Therapeutics), Neurocrine Biosciences, Neuroscience Software, Otsuka Pharmaceuticals, and Takeda Pharmaceuticals; honoraria from the American Psychological Association and the Psychonomic Society (for editorial work) as well as Alkermes, and research funding from Brain and Behavior Research Foundation, Dana Foundation, Millennium Pharmaceuticals, National Institute of Mental Health, and Wellcome Leap. In addition, he has received stock options from Compass Pathways, Engrail Therapeutics, Neumora Therapeutics, Neuroscience Software. All other authors report no financial relationships with commercial interest.

Open code and data

All data, task, and analysis code are shared openly to encourage replication and extension of our findings. Our results can be reproduced step-by-step via an R package and Jupyter notebooks – see the Github repository (https://github.com/qdercon/pstpipeline) for more details. Our study preregistration can be found on OSF (https://osf.io/fd4qu). An earlier version of this manuscript was pre-printed on PsyArXiv (https://psyarxiv.com/jmnek).

Footnotes

*

These authors contributed equally to this work.

References

Ahn, W.-Y., Haines, N., & Zhang, L. (2017). Revealing neurocomputational mechanisms of reinforcement learning and decision-making with the hBayesDM package. Computational Psychiatry, 1(0), 24. https://doi.org/10.1162/cpsy_a_00002.CrossRefGoogle ScholarPubMed
Ahn, W.-Y., Krawitz, A., Kim, W., Busemeyer, J. R., & Brown, J. W. (2011). A model-based fMRI analysis with hierarchical Bayesian parameter estimation. Journal of Neuroscience, Psychology, and Economics, 4(2), 95110. https://doi.org/10.1037/a0020684.CrossRefGoogle ScholarPubMed
Brown, V. M., Zhu, L., Solway, A., Wang, J. M., McCurry, K. L., King-Casas, B., & Chiu, P. H. (2021). Reinforcement learning disruptions in individuals with depression and sensitivity to symptom change following cognitive behavioral therapy. JAMA Psychiatry, 78(10), 11131122. https://doi.org/10.1001/JAMAPSYCHIATRY.2021.1844.CrossRefGoogle ScholarPubMed
Dalgleish, T., Black, M., Johnston, D., & Bevan, A. (2020). Transdiagnostic approaches to mental health problems: Current status and future directions. Journal of Consulting and Clinical Psychology, 88(3), 179. https://doi.org/10.1037/CCP0000482.CrossRefGoogle ScholarPubMed
de Leeuw, J. R. (2015). jsPsych: A JavaScript library for creating behavioral experiments in a Web browser. Behavior Research Methods, 47(1), 112. https://doi.org/10.3758/S13428-014-0458-Y.CrossRefGoogle Scholar
Delgado, M. R., Gillis, M. M., & Phelps, E. A. (2008). Regulating the expectation of reward via cognitive strategies. Nature Neuroscience, 11(8), 880. https://doi.org/10.1038/NN.2141.CrossRefGoogle ScholarPubMed
Elliott, R., Sahakian, B. J., Herrod, J. J., Robbins, T. W., & Paykel, E. S. (1997). Abnormal response to negative feedback in unipolar depression: Evidence for a diagnosis specific impairment. Journal of Neurology, Neurosurgery & Psychiatry, 63(1), 7482. https://doi.org/10.1136/JNNP.63.1.74.CrossRefGoogle ScholarPubMed
Enck, P., & Zipfel, S. (2019). Placebo effects in psychotherapy: A framework. Frontiers in Psychiatry, 10, 456. https://doi.org/10.3389/fpsyt.2019.00456.CrossRefGoogle ScholarPubMed
Frank, M. J., Moustafa, A. A., Haughey, H. M., Curran, T., & Hutchison, K. E. (2007). Genetic triple dissociation reveals multiple roles for dopamine in reinforcement learning. Proceedings of the National Academy of Sciences of the United States of America, 104(41), 1631116316. https://doi.org/10.1073/pnas.0706111104.CrossRefGoogle ScholarPubMed
Frank, M. J., Seeberger, L. C., & O'Reilly, R. C. (2004). By carrot or by stick: Cognitive reinforcement learning in parkinsonism. Science (New York, N.Y.), 306(5703), 19401943. https://doi.org/10.1126/SCIENCE.1102941.CrossRefGoogle ScholarPubMed
Gelman, A., & Carlin, J. (2014). Beyond power calculations: Assessing type S (Sign) and type M (Magnitude) errors. Perspectives on Psychological Science, 9(6), 641651. https://doi.org/10.1177/1745691614551642.CrossRefGoogle ScholarPubMed
Gelman, A., Carlin, J. B., Stern, H. S., Dunson, D. B., Vehtari, A., & Rubin, D. B. (2013). Bayesian Data Analysis (3rd ed.). New York: CRC Press.CrossRefGoogle Scholar
Gillan, C. M., Kosinski, M., Whelan, R., Phelps, E. A., & Daw, N. D. (2016). Characterizing a psychiatric symptom dimension related to deficits in goal-directed control. eLife, 5, e11305. https://doi.org/10.7554/eLife.11305.CrossRefGoogle ScholarPubMed
Goodrich, B., Gabry, J., Ali, I., & Brilleman, S. (2020). rstanarm: Bayesian applied regression modeling via Stan. R package version 2.21.1 https://mc-stan.org/rstanarm.Google Scholar
Halahakoon, D. C., Kieslich, K., O'Driscoll, C., Nair, A., Lewis, G., & Roiser, J. P. (2020). Reward-processing behavior in depressed participants relative to healthy volunteers: A systematic review and meta-analysis. JAMA Psychiatry, 77(12), 12861295. https://doi.org/10.1001/jamapsychiatry.2020.2139.CrossRefGoogle ScholarPubMed
Hales, C. A., Houghton, C. J., & Robinson, E. S. J. (2017). Behavioural and computational methods reveal differential effects for how delayed and rapid onset antidepressants effect decision making in rats. European Neuropsychopharmacology, 27(12), 12681280. https://doi.org/10.1016/J.EURONEURO.2017.09.008.CrossRefGoogle ScholarPubMed
Hamidian, S., Omidi, A., Mousavinasab, S. M., & Naziri, G. (2016). The effect of combining mindfulness-based cognitive therapy with pharmacotherapy on depression and emotion regulation of patients with dysthymia: A clinical study. Iranian Journal of Psychiatry, 11(3), 166.Google ScholarPubMed
Huys, Q. J. M., Maia, T. V., & Frank, M. J. (2016). Computational psychiatry as a bridge from neuroscience to clinical applications. Nature Neuroscience, 19(3), 404413. https://doi.org/10.1038/nn.4238.CrossRefGoogle ScholarPubMed
Huys, Q. J. M., Pizzagalli, D. A., Bogdan, R., & Dayan, P. (2013). Mapping anhedonia onto reinforcement learning: A behavioural meta-analysis. Biology of Mood & Anxiety Disorders, 3(1), 116. https://doi.org/10.1186/2045-5380-3-12.CrossRefGoogle ScholarPubMed
Insel, C., Reinen, J., Weber, J., Wager, T. D., Jarskog, L. F., Shohamy, D., & Smith, E. E. (2014). Antipsychotic dose modulates behavioral and neural responses to feedback during reinforcement learning in schizophrenia. Cognitive, Affective and Behavioral Neuroscience, 14(1), 189201. https://doi.org/10.3758/S13415-014-0261-3/TABLES/6.CrossRefGoogle ScholarPubMed
Jappe, L. M., Frank, G. K. W., Shott, M. E., Rollin, M. D. H., Pryor, T., Hagman, J. O., … Davis, E. (2011). Heightened sensitivity to reward and punishment in anorexia nervosa. International Journal of Eating Disorders, 44(4), 317324. https://doi.org/10.1002/EAT.20815.CrossRefGoogle ScholarPubMed
Kross, E., Gard, D., Deldin, P., Clifton, J., & Ayduk, O. (2012). “Asking why” from a distance: Its cognitive and emotional consequences for people with major depressive disorder. Journal of Abnormal Psychology, 121(3), 559.CrossRefGoogle ScholarPubMed
Lee, D. (2013). Decision making: From neuroscience to psychiatry. Neuron, 78(2), 233. https://doi.org/10.1016/J.NEURON.2013.04.008.CrossRefGoogle ScholarPubMed
Maia, T. V., & Frank, M. J. (2011). From reinforcement learning models to psychiatric and neurological disorders. Nature Neuroscience, 14(2), 154162. https://doi.org/10.1038/NN.2723.CrossRefGoogle ScholarPubMed
Mennin, D. S., Ellard, K. K., Fresco, D. M., & Gross, J. J. (2013). United we stand: Emphasizing commonalities across cognitive-behavioral therapies. Behavior Therapy, 44(2), 234248.CrossRefGoogle ScholarPubMed
Neacsiu, A. D., Eberle, J. W., Kramer, R., Wiesmann, T., & Linehan, M. M. (2014). Dialectical behavior therapy skills for transdiagnostic emotion dysregulation: A pilot randomized controlled trial. Behaviour Research and Therapy, 59, 4051. https://doi.org/10.1016/J.BRAT.2014.05.005.CrossRefGoogle ScholarPubMed
Nook, E. C., Hull, T. D., Nock, M. K., & Somerville, L. H. (2022). Linguistic measures of psychological distance track symptom levels and treatment outcomes in a large set of psychotherapy transcripts. Proceedings of the National Academy of Sciences of the United States of America, 119(13), e2114737119. https://doi.org/10.1073/pnas.2114737119.CrossRefGoogle Scholar
Palan, S., & Schitter, C. (2018). Prolific.ac – A subject pool for online experiments. Journal of Behavioral and Experimental Finance, 17, 2227. https://doi.org/10.1016/J.JBEF.2017.12.004.CrossRefGoogle Scholar
Papa, A., Boland, M., & Sewell, M. T. (2012). Emotion regulation and CBT. In O'Donohue, W. T. & Fisher, J. E. (Eds.), Cognitive behavior therapy: Core principles for practice (pp. 273323). Hoboken, NJ: John Wiley & Sons, Inc.CrossRefGoogle Scholar
Pedersen, M. L., & Frank, M. J. (2020). Simultaneous hierarchical Bayesian parameter estimation for reinforcement learning and drift diffusion models: A tutorial and links to neural data. Computational Brain and Behavior, 3(4), 458471. https://doi.org/10.1007/s42113-020-00084-w.CrossRefGoogle ScholarPubMed
Pike, A. C., & Robinson, O. J. (2022). Reinforcement learning in patients with mood and anxiety disorders vs control individuals: A systematic review and meta-analysis. JAMA Psychiatry, 79(4), 313322. https://doi.org/10.1001/jamapsychiatry.2022.0051.CrossRefGoogle ScholarPubMed
Pulcu, E., & Browning, M. (2017). Affective bias as a rational response to the statistics of rewards and punishments. eLife, 6, e27879. https://doi.org/10.7554/eLife.27879.CrossRefGoogle Scholar
Roiser, J. P., & Sahakian, B. J. (2013). Hot and cold cognition in depression. CNS Spectrums, 18(3), 139149. https://doi.org/10.1017/S1092852913000072.CrossRefGoogle ScholarPubMed
Siegle, G. J., Carter, C. S., & Thase, M. E. (2006). Use of fMRI to predict recovery from unipolar depression with cognitive behavior therapy. American Journal of Psychiatry, 163(4), 735738. https://doi.org/10.1176/AJP.2006.163.4.735/ASSET/IMAGES/LARGE/Q329F1.JPEG.CrossRefGoogle ScholarPubMed
Sloan, E., Hall, K., Moulding, R., Bryce, S., Mildred, H., & Staiger, P. K. (2017). Emotion regulation as a transdiagnostic treatment construct across anxiety, depression, substance, eating and borderline personality disorders: A systematic review. Clinical Psychology Review, 57, 141163. https://doi.org/10.1016/J.CPR.2017.09.002.CrossRefGoogle ScholarPubMed
Stan Development Team, (2021). Stan modelling language users guide and reference manual, v2.28.1. https://mc-stan.orgGoogle Scholar
Stansfeld, S., Clark, C., Bebbington, P., King, M., Jenkins, R., & Hinchliffe, S. (2016). Common mental disorders. In McManus, S., Bebbington, P., Jenkins, R., & Brugha, T. (Eds.), Mental health and wellbeing in England: Adult psychiatric morbidity survey 2014 (pp. 3768). Leeds: NHS Digital.Google Scholar
Staudinger, M. R., Erk, S., Abler, B., & Walter, H. (2009). Cognitive reappraisal modulates expected value and prediction error encoding in the ventral striatum. NeuroImage, 47(2), 713721. https://doi.org/10.1016/J.NEUROIMAGE.2009.04.095.CrossRefGoogle ScholarPubMed
Staudinger, M. R., Erk, S., & Walter, H. (2011). Dorsolateral prefrontal cortex modulates striatal reward encoding during reappraisal of reward anticipation. Cerebral Cortex, 21(11), 25782588. https://doi.org/10.1093/CERCOR/BHR041.CrossRefGoogle ScholarPubMed
Sutton, R. S., & Barto, A. G. (1998). Reinforcement learning: An introduction. Cambridge: MIT Press.Google Scholar
Valton, V, Wise, T., & Robinson, O. J.. (2020). Recommendations for Bayesian hierarchical model specifications for case-control studies in mental health. arXiv. https://doi.org/10.48550/arXiv.2011.01725.CrossRefGoogle Scholar
Vehtari, A., Gelman, A., & Gabry, J. (2016). Practical Bayesian model evaluation using leave-one-out cross-validation and WAIC. Statistics and Computing, 27(5), 14131432. https://doi.org/10.1007/S11222-016-9696-4.CrossRefGoogle Scholar
Vehtari, A., Gelman, A., Simpson, D., Carpenter, B., & Burkner, P. C. (2021). Rank-normalization, folding, and localization: An improved R-hat for assessing convergence of MCMC. Bayesian Analysis, 16(2), 667718. https://doi.org/10.1214/20-BA1221.CrossRefGoogle Scholar
Volman, I., Pringle, A., Verhagen, L., Browning, M., Cowen, P. J., & Harmer, C. J. (2020). Lithium modulates striatal reward anticipation and prediction error coding in healthy volunteers. Neuropsychopharmacology, 46(2), 386393. https://doi.org/10.1038/s41386-020-00895-2.CrossRefGoogle ScholarPubMed
Watkins, C. J. C. H., & Dayan, P. (1992). Q-learning. Machine Learning, 8(3), 279292. https://doi.org/10.1007/BF00992698.CrossRefGoogle Scholar
Weimer, K., Colloca, L., & Enck, P. (2015). Placebo effects in psychiatry: Mediators and moderators. The Lancet Psychiatry, 2(3), 246257. https://doi.org/10.1016/S2215-0366(14)00092-3.CrossRefGoogle ScholarPubMed
Wise, T., & Dolan, R. J. (2020). Associations between aversive learning processes and transdiagnostic psychiatric symptoms in a general population sample. Nature Communications, 11(1), 4179. https://doi.org/10.1038/s41467-020-17977-w.CrossRefGoogle Scholar
Wise, T., Robinson, O., & Gillan, C. (2022). Identifying transdiagnostic mechanisms in mental health using computational factor modeling. Biological Psychiatry, 93(8), 690703. https://doi.org/10.1016/j.biopsych.2022.09.034.CrossRefGoogle ScholarPubMed
Zorowitz, S., Solis, J., Niv, Y., & Bennett, D. (2021). Inattentive responding can induce spurious associations between task behavior and symptom measures. PsyArXiv. https://doi.org/10.31234/osf.io/rynhk.CrossRefGoogle Scholar
Figure 0

Figure 1. Reinforcement learning task design, transdiagnostic factor derivation, and comparison of factor scores to previous samples. a. The probabilistic selection task (PST) in the present study consisted of six blocks of sixty trials (training phase) where participants were instructed to choose between Hiragana characters presented as three pairs (AB, CD, EF; twenty of each per block), and given feedback (‘Correct!’ or ‘Incorrect.’). One character in each of the pairs was consistently more likely to be correct (reward probabilities of 0.8/0.2, 0.7/0.3, and 0.6/0.4 for A/B, C/D, and E/F respectively. 497 participants (49.9%) were randomised to a self-distancing intervention, and additionally received a prompt to ‘Distance yourself…’ with the fixation cross at the start of each training trial. The training phase was followed by a sixty-trial test phase without feedback where the twelve other possible character combinations were added (i.e. four of each of the fifteen pairs). All participants saw the same six characters, but the pairs themselves were randomised for each participant, and the order of the pairs was counterbalanced across trials. One of three affect questions was asked after each trial, with unlimited time to answer. b. Multi-target lasso regression with five-fold cross-validation was used to predict the three transdiagnostic dimension factor scores from different subsets of the 209 questions in the original dataset (Gillan et al., 2016, study 2). 78 questions were found to predict the three factor scores with high predictive accuracy (colours correspond to heatmap labels). c. Heatmap of five-fold cross-validated multi-target lasso regression coefficient weights for each of the included 78 questions across eight psychiatric questionnaires used to predict the three transdiagnostic symptom dimensions (i.e. weights for all other questions fixed at zero; see online Supplementary Methods for full details). d. Comparison between the factor score distributions for the n = 935 non-excluded participants in the present study (darker colours), and those previously obtained by Gillan et al. (2016) in n = 1413 participants (lighter colours). Inset plots show the predictive validity of the subset of 78 questions in predicting these scores in the original dataset.

Figure 1

Figure 2. Model comparison, parameter distributions, and raw training and test phase performance. a. Difference in numerical fit metrics between the dual and single learning rate models fit to training data alone, or training-plus-test, by distancing group. ELPD is the expected log posterior density (presented here as the difference between the dual and single learning rate models, where positive differences indicate a better model), and LOOIC is the leave-one-out information criterion (lower indicates a better model). b–c. Distributions of individual-level posterior means for learning parameters from the fits to training (b) and test (c) data. d. Raw training performance (cumulative probability of choosing the higher-probability symbol A/C/E), by group and stimulus pair, lagged by twenty trials (i.e. block-lagged, as each pair is presented twenty times per block. e. Raw test phase performance (% correct, where ‘correct’ is choosing the option in each pair which was most often correct during training) by test type, plus test phase performance on individual training pairs and novel pairs including symbols C or E.

Figure 2

Table 1. Demographic characteristics of the sample, by group.

Figure 3

Figure 3. Associations between reinforcement learning parameters and transdiagnostic psychiatric symptom dimensions. a–d. Coefficient posterior distributions from Bayesian GLMs (adjusted for age, gender, digit span and distancing status) reflecting the estimated percentage change in the learning rate parameter or the estimated mean change in the inverse temperature for a unit increase in anxiety/depression (a), social withdrawal (b), and compulsive behaviour (c–d) transdiagnostic factor scores. Parameters were derived from single (only compulsive behaviour presented here, C) or dual learning rate (a–b & d) Q-learning models fit to training alone or training-plus-test (parameters from fits to training-plus-test denoted prime). In all plots, boxplot boxes denote 95% HDI, and lines denote 99% HDI.

Figure 4

Figure 4. Model-derived comparisons between distanced and non-distanced participants. a–b. Coefficient posterior distributions from Bayesian GLMs (adjusted for age, gender, and digit span) reflecting the estimated percentage change in the learning rate parameter α (a) or αreward/αloss (b), and the estimated mean change in the inverse temperature β, comparing distanced and non-distanced participants. Parameters were estimated from Q-learning models with single (a) or dual (b) learning rates, fit to training alone or training-plus-test (parameters from fits to training-plus-test denoted with prime). c. Individual-level posterior mean parameter estimates for αreward, αloss, and β, for models including increasing numbers of training blocks. d. Parameter differences between distanced and non-distanced participants, estimated from dual learning rate models fit to increasing numbers of training blocks (sixty trials per block). In all plots, boxplot boxes denote 95% HDIs, and lines denote 99% HDIs.

Supplementary material: File

Dercon et al. supplementary material

Dercon et al. supplementary material
Download Dercon et al. supplementary material(File)
File 1.1 MB