3,362
Views
5
CrossRef citations to date
0
Altmetric
Articles

“Bigger” or “better”: the roles of magnitude and valence in “affective bias”

ORCID Icon & ORCID Icon
Pages 633-642 | Received 30 Nov 2018, Accepted 23 Aug 2019, Published online: 09 Sep 2019

ABSTRACT

Negative affective biases are thought to be a key symptom driving and upholding many psychiatric disorders. When presented with ambiguous information, anxious individuals, for example, tend to anticipate lower rewards than asymptomatic individuals (Aylward et al., Citation2019. Translating a rodent measure of negative bias into humans: the impact of induced anxiety and unmedicated mood and anxiety disorders. Psychological Medicine). The assumption is that this is because anxious individuals assume “worse” outcomes. However, predictions are often made about high and low rewards, so it is not clear whether the bias is due to the valence (the “worse” option) or just magnitude (the lower number). We therefore explored the roles of valence and magnitude in a translational measure of negative affective bias. We adapted a two-alternative forced choice (2AFC) “reward-reward” task into a “punishment-punishment” paradigm, and followed up with “high reward-high punishment” and “low reward-high punishment” variants. The results from the “punishment-punishment” paradigm – a bias towards higher punishments in healthy controls – suggest that it is outcome magnitude that is important. However, this is qualified by the other variants which indicate that both valence and magnitude are important. Overall, our results temper the assumption that negative affective biases observed in tasks using numeric outcomes are solely as a result of subjective outcome valence.

Negative affective bias refers to the phenomenon whereby patients with mood disorders, such as anxiety and depression, but also many other psychiatric disorders, tend to prioritise emotionally negative or unfavourable information or outcomes (Mathews & MacLeod, Citation1994). This bias can reveal itself in many ways. Anxious and depressed patients appear “pessimistic” in tasks which require them to anticipate the probability of a reward (Pizzagalli, Iosifescu, Hallett, Ratner, & Fava, Citation2008); identify a higher proportion of unfavourable words than neutral words (Foa & McNally, Citation1986; Powell & Hemsley, Citation1984); have longer response latencies to respond to words with emotionally-negative connotations (Gotlib & McCann, Citation1984; Watts, McKenna, Sharrock, & Trezise, Citation1986); and are able to more rapidly identify attentional probes when they appear in the region of a negative word (Broadbent & Broadbent, Citation1988; MacLeod, Mathews, & Tata, Citation1986).

Although many cognitive tasks used to investigate affective states utilise verbal or visual stimuli (e.g. faces), a large number use numeric outcomes. This may take the form of a “reward-reward” (R-R) two-alternative forced choice (2AFC) task, where the participant is required to estimate whether an ambiguous tone is closer in frequency to a tone associated with a higher reward or one associated with a lower reward (Aylward, Hales, Robinson, & Robinson, Citation2019). The prediction is that those afflicted by a negative affective bias will tend to anticipate lower rewards more often than higher rewards. The problem with this paradigm, however, is that is rests on the assumption that lower rewards are chosen because they are intrinsically lower in value than the higher rewards, rather than simply a lower magnitude. For example, Brilot, Asher, and Bateson (Citation2010) explored the association between stereotypic behaviour and “negative affect” in starlings, using food rewards of differing sizes. The fact that birds which somersaulted more were more likely to anticipate a low reward was taken to mean that somersaulting was symptomatic of a negative affective state. Hales, Robinson, and Houghton (Citation2016) also found a preference for the low reward when rats were subject to either acute drug-induced anxiety or chronic restraint stress, again with the assumption that this indicated a negative affective state. Most recently, Aylward et al. (Citation2019) used a similar 2AFC task and found that humans suffering from anxiety mimic the choice behaviour of rodents undergoing anxiogenic manipulation. The translational nature of the task, between animals and humans, is useful, however it has not been demonstrated that this choice bias is due to valence (i.e. the subjective, positive or negative, value of the outcome) rather than magnitude (i.e. the absolute size of the outcome, whether a reward or a punishment). Although valence and magnitude are equivalent in tasks which offer two different monetary rewards, in the real world, a “bigger” car is not always the “better” car, for instance.

To differentiate between these hypotheses, we expanded the investigation to include a “punishment-punishment” (P-P) or “negative domain” 2AFC task, where participants must choose whether they anticipate a larger or a smaller monetary loss. Presumably, a bias towards both larger rewards and punishments (i.e. a “magnitude bias”) could signify a bias towards more “extreme” outcomes in general. This is not implausible: the environment in which we live is noisy, and the brain must be able to filter out relatively unimportant information whilst being alerted to “important” inputs, regardless of whether they are perceived to be “positive” or “negative”. One theory addressing such a phenomenon is known as the “salience bias” or “perceptual salience”, and is well-documented in the literature (Bordalo, Gennaioli, & Shleifer, Citation2012; Tversky, Slovic, & Kahneman, Citation1982).

However, we predicted that outcome valence would be the deciding factor, and as a bias towards “positive” outcomes has been demonstrated in healthy individuals using different tasks (Anderson, Hardcastle, Munafò, & Robinson, Citation2011; Erickson et al., Citation2005), we expected to see a bias towards the lowest punishment in an unselected sample. Contrary to predictions, the effect appeared to be driven by outcome magnitude (and those who anticipated higher gains also anticipated higher losses), so we then explored bias on follow-up “high reward-high punishment” and “low reward-high punishment” task variations. We predicted that if the effect was entirely due to magnitude, we should see no bias towards either outcome on the first task (where magnitudes are matched), and a bias towards losses on the second task (because the gain has a lower magnitude). We then performed a full replication based on effect size estimates from the initial studies. Finally, we isolated the negative domain task from experiment 1 and replicated a third time in the absence of reward stimuli, to demonstrate that effects are not specific to the original (task-switching) context.

Method

Study design

We used the online behavioural science platform Gorilla (https://gorilla.sc/) to conduct the tasks. We used, and modified, an existing R-R 2AFC task in our assessment of cognitive bias. 2AFC tasks have originally been conducted using an auditory platform for stimulus presentation, although it has been demonstrated that they can be translated onto a visual platform whilst retaining the same sensitivity to the affective bias. For example, a repeated measures ANOVA found no effect of task on the anticipation of favourable outcome (F(2102) = 1.357, p = 0.262) (Jaber Ansari, Citation2017), which is the key measure of affective bias on this task. Using a visual platform (where participants must decide whether a shape is closer in size to one associated with a higher or lower reward) is more appropriate for online platforms as it removes any confounding factors associated with speaker or headphone availability.

Participants

Experiment 1 involved 45 participants. We powered experiment 2 (N = 67) to detect an effect size d = 0.347 to match the effect size found in experiment 1, for the same measure as experiment 1, at p < 0.05 with 80% power. After these experiments, we replicated our analyses in a larger sample size for both experiment 1 (N = 122) and experiment 2 (N = 130). This allowed us to detect a small-medium effect size d = 0.3 at p < 0.05 with 90% power. Statistical inference was run on the two experiments separately. However, to get more accurate estimates of effect sizes we pooled these results across experiment 1 (N = 167) and experiment 2 (N = 197). Finally, we isolated the negative domain task from experiment 1 and replicated this task again, this time using 156 participants on the basis of the average pooled effect sizes we observed across experiments 1 and 2.

Participants were recruited online in the Amazon Mechanical Turk (MTurk) marketplace, and completed a consent form prior to taking part (UCL ethics reference 6199/001). Aside from access to a computer or laptop with an internet connection, there were no specific exclusion criteria for the study. Participants were able to withdraw from the study at any time by closing their browser, and randomisation of task order proceeded according to participant recruitment number and Gorilla's built-in randomisation procedure. Our data are available at https://osf.io/uxcd7/.

Details of the tasks

A schematic of both experiment 1 and experiment 2 is shown in .

Figure 1. A schematic of the method used. The feedback outcomes are shown for each task.

Figure 1. A schematic of the method used. The feedback outcomes are shown for each task.

Experiment 1

Experiment 1 consisted of an R-R task and a P-P task. Participants completed both tasks in a randomly-allocated order. Both main tasks were preceded by a practice block. In the practice block, participants were presented with a random series of big (400 × 400) and small (200 × 200) shapes, and had to press the “z” and “m” keys on the keyboard to identify whether a shape was big or small.

In the R-R task, responding correctly to the shape resulted in a big (£4) or small (£1) reward, depending on the size of the shape. An incorrect or absent response resulted in a £20 loss. Participants started the task with no money and were told to try and win as much money as possible. The Hales et al. (Citation2016) and Aylward et al. (Citation2019) studies both used values of 1 and 4 as rewards, and as biases in choice behaviour were able to be detected using such values, we have used the same ones in the current study.

In the P-P task, responding correctly to the shape resulted in a big (-£4) or small (-£1) loss, depending on the size of the shape. An incorrect or absent response led to a £20 loss, in order to discourage abstinence from responding. Participants were told they would start with £800 and were instructed to try and minimise their losses.

Stimuli were presented for 1000 ms, followed by a 750 ms fixation period. Therefore, the participants had 1750 ms in total to respond. Feedback on the response was then presented for 850 ms. In the practice block, 10 big shapes and 10 small shapes were presented.

The main task followed the same structure as the practice block. However, as well as the “unambiguous” 400 × 400 and 200 × 200 shapes, participants were shown “ambiguous” mid-size (300 × 300) shapes, which randomly corresponded to either outcome. In the main task, there were 120 trials, with 40 big shapes, 40 small shapes and 40 mid-size shapes being randomly presented.

Experiment 2

We ran a second experiment to build on the results we attained in experiment 1, and further tease apart the effects of valence and magnitude. Keeping the basic structure of both tasks the same, we changed the outcomes in the positive domain task to a £4 win and a £4 loss (4, −4), and those in the negative domain to a £1 win and a £4 loss (1, −4). This resulted in a “high reward-high punishment” task, and a “low reward-high punishment” task. As before, participants lost £20 if they gave an incorrect answer.

Statistical analyses

Statistical analyses were run using JASP (0.8.5.1). Choice behaviour was assessed by calculating the proportion of ambiguous stimuli that were interpreted as a “favourable outcome” (p(mid = favourable)). This was analysed using a two-tailed one-sample t-test with a test value of 0.5, and a within-subject t-test. A valence bias would cause p(mid = favourable) to be significantly above 0.5 in all experiments.

Correlation analyses were run on p(mid = favourable) between tasks. A valence bias would result in a positive correlation between the anticipation of the favourable outcome in each task. Independent t-tests were also run to compare choice behaviour across the two experiments (i.e. between the (4, 1) and (4, −4), and (4, 1) and (1, −4) tasks), in order to see whether increasing the magnitude of the unfavourable outcome also increased the bias towards it. If bias was driven by magnitude, we would expect this to be the case.

We also collected accuracy and reaction time measures for each outcome on each task.

In our replications we used one-tailed instead of two-tailed t-tests whenever this was justified based on clear predictions from previous results. In the dataset pooled across initial and replication studies, we do not calculate inferential statistics, only effect sizes.

Bayesian statistics were also used; Bayesian inference compares the likelihood of observing the given data under the alternative hypothesis (H1) with the likelihood of observing it under the null hypothesis (H0). BF10 refers to the evidence for H1 relative to H0, and can be graded, for example, as anecdotal (1–3), moderate (3–10), or strong (10–30). BF+0 and BF−0 refer to one-tailed t-tests where the hypothesis is that the bias is more than or less than 0.5 respectively. The Bayesian credible interval (BCI) is reported for p(mid = favourable), and is the range of values within which the true value lies with 95% probability. In all cases we used the JASP default Cauchy prior for these Bayesian analyses.

In both experiments, circular and triangular stimuli were counterbalanced across tasks to minimise learning being carried over between tasks. We counterbalanced the shape size and reward/loss contingencies, and the “z” and “m” key response and stimulus size contingencies, across participants (these were consistent across the two tasks for each participant). An ANOVA found a significant effect of the counterbalancing condition upon choice behaviour in the (4, 1) (F(151) = 3.127, p < 0.001), (−4, −1) (F(146) = 3.495, p < 0.001), (4, −4) (F(181) = 4.944, p < 0.001), and (1, −4) (F(175) = 3.474, p < 0.001) tasks. However, since the counterbalancing was matched for each participant across tasks, counterbalancing cannot account for any differences in task performance within each experiment.

Results

Experiment 1

Choice behaviour

Initial sample

In the positive domain task, participants were more likely to anticipate a high reward (i.e. 4 over 1; t(44) = 2.580, p = 0.013, d = 0.385, BF10 = 3.061, BCI [0.518, 0.643]), and in the negative domain task they were more likely to anticipate a high punishment (i.e. −4 over −1; t(41) = −2.103, p = 0.042, d = −0.325, BF10 = 1.217, BCI [0.379, 0.498]). There was a significant difference in bias towards the favourable outcome between the two tasks (t(41) = 2.674, p = 0.005, d = 0.413, BF+0 = 7.499), but when the bias score was reversed in the P-P task so that the extent of the bias towards 4 and −4 in the two tasks could be compared, there was no significant difference between choice probability across the two tasks (t(41) = 0.881, d = 0.136, p = 0.192); and Bayesian analysis favoured the null model (BF+0 = 0.385), i.e. that the bias measure is equivalent between 4 and −4. In other words, participants favoured the larger (rather than more favourable) of the two options in both cases. This is presented in .

Figure 2. Raincloud plot showing the proportion of favourable responses to the mid-shape on the (4, 1) (N = 167), (−4, −1) (N = 162), (4, −4) (N = 197), and (1, −4) (N = 191) tasks, for the initial (blue) and replication (orange) samples. The dashed grey line represents a bias measure of 0.5, and the dashed coloured lines in each data set correspond to the mean line for that set. Significance levels for the bias measures are shown for each sample (* p < 0.05, ***p < 0.001 for test against 0.5; ns = not significant). Effect sizes are shown for the pooled sample for each task, and for the pooled differences in the bias measure between tasks.

Figure 2. Raincloud plot showing the proportion of favourable responses to the mid-shape on the (4, 1) (N = 167), (−4, −1) (N = 162), (4, −4) (N = 197), and (1, −4) (N = 191) tasks, for the initial (blue) and replication (orange) samples. The dashed grey line represents a bias measure of 0.5, and the dashed coloured lines in each data set correspond to the mean line for that set. Significance levels for the bias measures are shown for each sample (* p < 0.05, ***p < 0.001 for test against 0.5; ns = not significant). Effect sizes are shown for the pooled sample for each task, and for the pooled differences in the bias measure between tasks.

Replication sample

As before, participants were more likely to anticipate a high reward (t(121) = 1.866, p = 0.032, d = 0.169, BF+0 = 1.042, BCI [0.498, 0.565]). However, in the negative domain task there was no significant bias in either direction (t(119) = −0.923, p = 0.179, d = −0.084, BF−0 = 0.251, BCI [0.446, 0.520]). There was no significant difference in bias towards the favourable outcome between the two tasks (t(114) = 1.461, p = 0.073, d = 0.136, BF+0 = 0.537). Again, however, there was no significant difference between the bias towards 4 and that towards −4 (t(114) = 0.854, p = 0.198, d = 0.080, BF+0 = 0.236). This is presented in .

Pooled samples

p(mid = favourable) was above 0.5 in both the positive domain (0.545 (SD 0.194), t(166) = 2.983, d = 0.231) and below 0.5 in the negative domain (0.471 (SD 0.199), t(161) = −1.824, d = −0.143).

All data collected for the initial, replication and pooled samples in experiment 1, including accuracy and reaction time measures, are displayed in .

Table 1. Summary of the key measures in experiment 1. For each measure, the results from the initial, replication and pooled samples are shown.

Isolated negative domain task

Here, we presented the (−1, −4) (negative domain) task on its own, in order to remove the effects of any reward stimuli. As in the initial sample, on full-powered second replication participants were more likely to anticipate a high punishment than a low punishment in the negative task (t(155) = −2.196, p = 0.030, d = −0.176, BF10 = 0.920, BCI [0.441, 0.497]). The bias measure from the isolated negative domain task is demonstrated in .

Figure 3. Raincloud plot, using the same layout as , showing the proportion of favourable responses to the mid-shape on a second replication of the (−4, −1) (N = 156) task on its own (i.e. isolated from any reward stimuli).

Figure 3. Raincloud plot, using the same layout as Figure 2, showing the proportion of favourable responses to the mid-shape on a second replication of the (−4, −1) (N = 156) task on its own (i.e. isolated from any reward stimuli).

Correlations

There was a negative correlation between p(mid = favourable) across both domains in the initial (r = −0.598, p < 0.001, BF10 > 30), replication (r = −0.523, p < 0.001, BF10 > 30) and pooled (r = −0.548) samples, suggesting that those who showed a bias towards high rewards also showed a bias towards high punishments and vice versa. This correlation is shown in .

Figure 4. Correlation between the proportion of high outcome responses to mid-shapes in the positive and negative domains, in both the initial sample (blue) and replicated sample (orange). The negative correlation indicates that individuals who chose the most favourable option in the positive domain chose the least favourable option in the negative domain, indicating that they chose on the basis of magnitude rather than valence. The dashed grey line is the identity.

Figure 4. Correlation between the proportion of high outcome responses to mid-shapes in the positive and negative domains, in both the initial sample (blue) and replicated sample (orange). The negative correlation indicates that individuals who chose the most favourable option in the positive domain chose the least favourable option in the negative domain, indicating that they chose on the basis of magnitude rather than valence. The dashed grey line is the identity.

Experiment 2

Choice behaviour

Initial sample

In the (4, −4) task, there was a significant bias towards the reward (t(66) = 4.301, p < 0.001, d = 0.525, BF10 > 30, BCI [0.553, 0.644]). In the (1, −4) task, there was no significant bias towards either outcome (t(64) = 1.415, p = 0.162, d = 0.175, BF10 = 0.351, BCI [0.486, 0.584]). In this case, the bias towards the reward was significantly higher in the (4, −4) task than it was in the (1, −4) task (t(63) = 2.307, p = 0.012, d = 0.288, BF10 = 1.599). This is presented in

Replication sample

There was a significant bias towards the reward in both the (4, −4) task (t(129) = 2.345, p = 0.010, d = 0.206, BF+0 = 2.694, BCI [0.507, 0.583]) and the (1, −4) task (t(125) = 2.076, p = 0.040, d = 0.185, BF10 = 0.788, BCI [0.502, 0.573]). There was no significant difference between the bias in the (4, −4) task and in the (1, −4) task (t(122) = 0.470, p = 0.320, d = 0.042, BF+0 = 0.151). This is presented in .

Pooled samples

There was a numeric bias towards the reward in both the (4, −4) task (0.563 (SD 0.209), t(196) = 4.228, d = 0.301) and the (1, −4) task (0.536 (SD 0.200), t(190) = 2.519, d = 0.182), albeit with a considerably reduced effect size.

All data collected for the initial, replication and pooled samples in experiment 2, including accuracy and reaction time measures, are displayed in .

Table 2. Summary of the key measures in experiment 2. For each measure, the results from the initial, replication and pooled samples are shown.

Correlations

There was a significant positive correlation between p(mid = favourable) across both tasks in initial (r = 0.384, p = 0.002, BF10 = 18.65) replication (r = 0.588, p < 0.001, BF10 > 30) and pooled (r = 0.521) samples, meaning that those who had a bias towards the reward in one task had a similar bias in the other.

Between-experiment analysis

Initial sample

Independent t-tests revealed that changing the outcomes from (4, 1) to (4, −4) had no significant effect on p(mid = favourable) (t(110) = −0.474, p = 0.318, d = −0.091). Bayesian analysis favoured the null model (BF−0 = 0.303).

Changing the outcomes from (4, 1) to (1, −4) also had no significant effect on p(mid = favourable) (t(108) = 1.160, p = 0.124, d = 0.225). Bayesian analysis favoured the null model (BF+0 = 0.645), so we could accept the null hypothesis. Therefore, although there was no significant outcome bias on the (1, −4) task, its results were not significantly different to the (4, 1) task.

Replication sample

Changing the outcomes from (4, 1) to (4, −4) had no significant effect on p(mid = favourable) (t(250) = −0.517, p = 0.303, d = −0.065). Bayesian analysis favoured the null model (BF−0 = 0.217), so the null hypothesis was accepted.

Changing the outcomes from (4, 1) to (1, −4) also had no significant effect on p(mid = favourable) (t(246) = −0.232, p = 0.408, d = −0.030). Bayesian analysis favoured the null model (BF−0 = 0.168).

The overall bias measures from the pooled studies, for both experiments 1 and 2, are shown in .

Discussion

In experiment 1, counter to predictions, those who were more likely to anticipate a high reward were also more likely to anticipate a high punishment. These results do not support the hypothesis that the interpretation bias is due to outcome valence; rather, the bias appears to be due to the absolute magnitude of the outcome, whether that outcome is positive or negative. In experiment 2, we therefore attempted to probe this further using a task comparing a gain and a loss of 4, which should demonstrate no bias if the effect was purely magnitude, and a task comparing a gain of 1 and a loss of 4, which should demonstrate a bias towards the loss if the effect was purely magnitude. However, in both cases a significant bias was seen towards the favourable (gain) condition, which indicates an effect of valence.

The R-R task in experiment 1 is consistent with the “positive” bias found among healthy individuals in previous tasks. However, the P-P version of the task contradicts this. Specifically, the individuals who were more likely to select the high reward option were also more likely to select the high punishment option. To try to explain this, the results could be approached from two different perspectives. The first perspective posits that valence, or the subjective value of an outcome, is the key modulator of bias. Results from the R-R (4, 1) task and both tasks in experiment two ((4, −4) and (1, −4)) fit this perspective, because there was a bias towards the most favourable outcome in all cases. Also, the results from the between-experiment analyses suggested that increasing the magnitude of the unfavourable outcome did not increase the bias towards it. The negative domain task may therefore contradict this because the task was simply misinterpreted; the unintuitive “damage limitation” paradigm of this task may have caused some confusion whereby participants thought that the aim of the task was to lose as much money as possible; notably, we did not see the same effect in the first replication of the negative domain task.

However, the second perspective is that outcome magnitude (the absolute size of the outcome, whether positive or negative) is at least partially responsible for the results. This perspective is supported by correlation analysis in experiment 1, where those who favoured higher rewards also favoured higher losses, as well as the initial P-P task where there was an overall bias towards the higher loss. Perhaps more importantly, when we isolated the negative domain task such that participants did not also complete the original positive domain task, we again saw a bias towards the higher loss. This reduces the likelihood that the results from the initial sample were a result of confusion between the P-P and R-R tasks. Similarly, the considerably reduced effect size of the “favourable” bias on the (1, −4) task in the pooled data (and lack of significant bias in the initial sample) relative to the (4, −4) task suggests that the higher magnitude of the loss relative to the gain may have “dragged down” the “favourable” bias. However, it should be noted that this difference is not significant in the replication sample. Moreover, if the effect was purely magnitude then we would have demonstrated no overall bias in the (4, −4) version and a bias towards losses in the (1, −4) version. Thus, there appears to be a combined effect of magnitude and valence which may change depending on the outcome values used. The effect that we see in the P-P task may therefore be because when the task is reframed in the loss domain (i.e. it is only possible to lose), participants shift to a decision-making strategy that makes them warier of larger losses and more likely to anticipate them, as opposed to taking advantage of any large rewards on offer, as seen in the R-R task. Of course, there still remains the possibility that the results are due to a misunderstanding of task instructions, but the results from the isolated task in a large sample demonstrate that this is not driven by the presence of a reward task, or confusion due to task-switching.

This study is important due to the extensive use of 2AFC tasks to assess cognitive affective bias in research. Tasks of this nature are popular because they are translational between species; the impact of a treatment upon task performance in other species can serve as a good predictor of its effects in humans. When these tasks follow a “reward-reward” paradigm, there is often no distinction between a bias towards a “more valuable” reward, and simply a “larger” reward. The results from our study do not definitively support the idea that subjective valence alone produces the biases observed in emotional 2AFC tasks, and may suggest a possible influence of outcome magnitude. This implies that caution may be warranted when interpreting the results of 2AFC tasks as an assessment of cognitive affective bias, with care taken to control for any possible effects of outcome magnitude upon the bias.

In terms of future directions, one area of interest might be to explore the impact of multipliers on the numbers; for example, is the effect exacerbated for 10 and 40 relative to 1 and 4? In order to reduce the risk of confusion on behalf of the participants whilst still isolating the effects of valence and magnitude, it may be sensible to compare biases across 2AFC tasks which use outcome combinations of (1, 2), (1, 3), (1, 4) and (1, 5), rather than contrasting two wins of different magnitudes against two losses of different magnitudes. It could also be interesting to move away from the 2AFC design and explore choice behaviour when using outcomes of 4, 1, −1 and −4 in the same task. A similar paradigm may also be used in other tasks adapted to probe affective biases which use non-numeric outcomes or stimuli. For example, the emotional Stroop task requires subjects to name the colour of an emotionally-valenced word, with the rationale being that longer response latencies signify an increased attentional bias to the word. Response times of healthy participants to “weakly positive” versus “strongly negative” words in a task such as this could help to delineate the effects of the “valence” and the “magnitude” of a stimulus.

One potential limitation of our study stems from the online method of data collection, where it is difficult to standardise the surroundings and behaviour of participants whilst they complete the cognitive tasks. Although it has been shown that online data collection on MTurk leads to no more issues with task engagement than data collection in the lab (Thomas & Clifford, Citation2017), collecting data in person is the only feasible way to guarantee that external factors are kept constant between subjects and, if the effects are due to confusion about task instructions, is a better way to asses this possibility. Also, although prior work (Aylward et al., Citation2019) has linked such tasks to affective biases in psychopathology, it is important to note that this sample was not screened and we have no way of directly linking this effect to clinical symptoms. Future work should explore the relationship between these effects and clinically relevant symptoms. It should be noted that we had no a priori predictions about the influence of demographic variables on our effects of interest and as such did not acquire this data. Future work may address this limitation by exploring the impact of such variables on magnitude and valence bias.

Conclusions

Behavioural experiments of this nature are an important first step towards understanding exactly how mood disorders and their treatments may operate, as they provide a rudimentary correlate of the way the brain processes information to interpret the world around it. However, these experiments will only prove to be useful if the conclusions we draw from them are valid, and this necessitates exploring the various factors that may affect decision-making in such a setting, including outcome magnitude. Only with a robust understanding of the cognitive traits exhibited by those with affective disorders can meaningful progress be made regarding their treatment and interventions.

Declaration of contribution

Oliver Robinson conceived of the experiment and wrote the data pre-processing script. Jack Love created the task testing scripts, and completed data collection and data analysis. Jack Love wrote the paper and Oliver Robinson provided feedback. The authors declare no conflicts of interest.

Acknowledgements

This research was funded by A Medical Research Foundation Equipment Competition Grant (C0497, Principal Investigator OJR), and a Medical Research Council Career Development Award to OJR (MR/K024280/1).

Disclosure statement

OJR's current MRC senior fellowship is partially in collaboration with Cambridge Cognition (who plan to provide in kind contribution) and he is running an investigator initiated trial with medication donated by Lundbeck (escitalopram and placebo, no financial contribution). He holds an MRC-Proximity to discovery award with Roche (who provide in kind contributions and sponsor travel) regarding work on heart-rate variability and anxiety. He has also completed consultancy work on affective bias modification for Peak and online CBT for IESO digital health. He is a member of the committee of the British Association of Psychopharmacology. These disclosures are made in the interest of full transparency and do not constitute a conflict of interest with the current work. JL reports no conflict of interest.

Additional information

Funding

This work was supported by Medical Research Council [grant number MR/K024280/1]; Medical Research Foundation [grant number C0497].

References

  • Anderson, M., Hardcastle, C., Munafò, M., & Robinson, E. (2011). Evaluation of a novel translational task for assessing emotional biases in different species. Cognitive, Affective, & Behavioral Neuroscience, 12(2), 373–381. doi: 10.3758/s13415-011-0076-4
  • Aylward, J., Hales, C., Robinson, E., & Robinson, O. (2019). Translating a rodent measure of negative bias into humans: The impact of induced anxiety and unmedicated mood and anxiety disorders. Psychological Medicine, 1–10. doi:10.1017/S0033291718004117.
  • Bordalo, P., Gennaioli, N., & Shleifer, A. (2012). Salience theory of choice under risk. The Quarterly Journal of Economics, 127(3), 1243–1285. doi: 10.1093/qje/qjs018
  • Brilot, B., Asher, L., & Bateson, M. (2010). Stereotyping starlings are more ‘pessimistic’. Animal Cognition, 13(5), 721–731. doi: 10.1007/s10071-010-0323-z
  • Broadbent, D., & Broadbent, M. (1988). Anxiety and attentional bias: State and trait. Cognition & Emotion, 2(3), 165–183. doi: 10.1080/02699938808410922
  • Erickson, K., Drevets, W., Clark, L., Cannon, D., Bain, E., Zarate, C., … Sahakian, B. (2005). Mood-congruent bias in affective Go/No-Go performance of unmedicated patients with major depressive disorder. American Journal of Psychiatry, 162(11), 2171–2173. doi: 10.1176/appi.ajp.162.11.2171
  • Foa, E., & McNally, R. (1986). Sensitivity to feared stimuli in obsessive-compulsives: A dichotic listening analysis. Cognitive Therapy and Research, 10(4), 477–485. doi: 10.1007/BF01173299
  • Gotlib, I., & McCann, C. (1984). Construct accessibility and depression: An examination of cognitive and affective factors. Journal of Personality and Social Psychology, 47(2), 427–439. doi: 10.1037/0022-3514.47.2.427
  • Hales, C., Robinson, E., & Houghton, C. (2016). Diffusion modelling reveals the decision making processes underlying negative judgement bias in rats. PLoS One, 11(3), e0152592. doi: 10.1371/journal.pone.0152592
  • Jaber Ansari, H. (2017). Developing an online platform for the assessment of negative affective bias (Unpublished master’s thesis). University College London, UK.
  • MacLeod, C., Mathews, A., & Tata, P. (1986). Attentional bias in emotional disorders. Journal of Abnormal Psychology, 95(1), 15–20. doi: 10.1037/0021-843X.95.1.15
  • Mathews, A., & MacLeod, C. (1994). Cognitive approaches to emotion and emotional disorders. Annual Review of Psychology, 45(1), 25–50. doi: 10.1146/annurev.ps.45.020194.000325
  • Pizzagalli, D., Iosifescu, D., Hallett, L., Ratner, K., & Fava, M. (2008). Reduced hedonic capacity in major depressive disorder: Evidence from a probabilistic reward task. Journal of Psychiatric Research, 43(1), 76–87. doi: 10.1016/j.jpsychires.2008.03.001
  • Powell, M., & Hemsley, D. (1984). Depression: A breakdown of perceptual defence? British Journal of Psychiatry, 145(4), 358–362. doi: 10.1192/bjp.145.4.358
  • Thomas, K., & Clifford, S. (2017). Validity and mechanical turk: An assessment of exclusion methods and interactive experiments. Computers in Human Behavior, 77, 184–197. doi: 10.1016/j.chb.2017.08.038
  • Tversky, A., Slovic, P., & Kahneman, D. (1982). Judgment under uncertainty: Heuristics and biases. Cambridge: Cambridge University Press.
  • Watts, F., McKenna, F., Sharrock, R., & Trezise, L. (1986). Colour naming of phobia-related words. British Journal of Psychology, 77(1), 97–108. doi: 10.1111/j.2044-8295.1986.tb01985.x