Hostname: page-component-78c5997874-fbnjt Total loading time: 0 Render date: 2024-11-19T08:28:58.112Z Has data issue: false hasContentIssue false

Methods in causal inference. Part 2: Interaction, mediation, and time-varying treatments

Published online by Cambridge University Press:  01 October 2024

Joseph A. Bulbulia*
Affiliation:
Victoria University of Wellington, Wellington, New Zealand

Abstract

The analysis of ‘moderation’, ‘interaction’, ‘mediation’ and ‘longitudinal growth’ is widespread in the human sciences, yet subject to confusion. To clarify these concepts, it is essential to state causal estimands, which requires the specification of counterfactual contrasts for a target population on an appropriate scale. Once causal estimands are defined, we must consider their identification. I employ causal directed acyclic graphs and single world intervention graphs to elucidate identification workflows. I show that when multiple treatments exist, common methods for statistical inference, such as multi-level regressions and statistical structural equation models, cannot typically recover the causal quantities we seek. By properly framing and addressing causal questions of interaction, mediation, and time-varying treatments, we can expose the limitations of popular methods and guide researchers to a clearer understanding of the causal questions that animate our interests.

Type
Methods Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
Copyright © The Author(s), 2024. Published by Cambridge University Press

Introduction

The young Charles Darwin was a keen fossil hunter and amateur geologist. In August 1831, he accompanied the geologist Adam Sedgwick to the Glyderau mountain range in northern Wales.

We spent many hours in Cwm Idwal, … but neither of us saw a trace of the wonderful glacial phenomena all around us; we did not notice the plainly scored rocks, the perched boulders, the lateral and terminal moraines. Yet these phenomena are so conspicuous that … a house burnt down by fire did not tell its story more plainly than did this valley. If it had still been filled by a glacier, the phenomena would have been less distinct than they now are. (Darwin, Reference Darwin and Darwin1887: 25)

This ‘striking instance of how easy it is to overlook phenomena, however conspicuous’ (Darwin, Reference Darwin and Darwin1887: 25) is cited in cultural evolution to emphasise the importance of theory for organising observations (Wilson, Reference Wilson, Bulbulia, Sosis, Harris, Genet, Genet and Wyman2008). However, the importance of theory to scientific discovery carries broader relevance: it applies to the statistical methods scientists routinely apply to the data they collect. Without a clear framework that relates statistical models to observations, the understanding we seek from our data remains elusive.

Across many human sciences, we apply statistical models to data and report ‘moderation’, ‘interaction’, ‘mediation’ and ‘longitudinal growth’. How are we to interpret the results of these models? It is often unclear. The confidence with which investigators report findings does not make interpretation any clearer. The problem is that investigators are typically interested in causal questions. Answering a causal question requires a careful workflow that begins by defining the causal quantities and the target population of interest. We specify a causal quantity, or estimand, as the contrast between counterfactual outcomes under two (or more) clearly defined interventions. Mathematical proofs establish that we can consistently estimate the size of this counterfactual difference using data, provided certain assumptions are met (discussed below). The subsequent steps in a causal inferential workflow involve assessing the credibility of these necessary assumptions, constructing appropriate estimators, and obtaining relevant data. Only after these steps should statistical analysis be attempted. Without adhering to a causal inferential workflow, the associations derived from statistical models will reflect true causal relationships only by chance, regardless of the sophistication of our statistical methods (Westreich & Greenland, Reference Westreich and Greenland2013).

There is good news. Advances in the health sciences, computer science and economics have fostered a common vocabulary and established robust workflows, enabling investigators to formulate causal questions that can be addressed with data. These developments facilitate the evaluation of the necessary assumptions for obtaining consistent estimates, the construction of valid estimators, and the application of statistical models at the culmination of the workflow. This conceptual framework, grounded in mathematical proofs, empowers investigators to effectively clarify, communicate, and evaluate causal questions using observational data. The consensus that has emerged in causal inference over the past several decades is, in my view, as transformative for the human sciences as the theory of glaciation was for geology or Darwin's theory of evolution was for biology. By reframing questions of interaction, mediation and time-varying treatments as causal inquiries, I aim to clarify the framework's interest, relevance and power.

Several excellent resources clarify workflows for causal inference, guiding researchers from stating causal questions to communicating results (Hernán & Robins, Reference Hernán and Robins2024; Laan et al., Reference Laan, Coyle, Hejazi, Malenica, Phillips and Hubbard2023; Montgomery et al., Reference Montgomery, Nyhan and Torres2018; Morgan & Winship, Reference Morgan and Winship2014; Neal, Reference Neal2020; Pearl, Reference Pearl2009; Tibshirani et al., Reference Tibshirani, Athey, Sverdrup and Wager2024; T. J. VanderWeele, Reference VanderWeele2015).

Here, my ambition is focussed.

Part 1 considers how to formulate causal questions when our interest lies in comparing effect magnitudes between groups (effect modification).

Part 2 considers how to formulate causal questions to assess the joint effects of two independent interventions (interaction).

Part 3 considers how to formulate causal questions to assess the joint effects of two dependent interventions (mediation analysis).

Part 4 considers how to formulate causal questions when evaluating two or more sequential treatments of the same kind (time-varying treatments).

I begin with a brief introduction to key concepts and terminology.

Fundamental assumptions for causal inference

Consider indicators A and Y measuring states of the world. Let A denote the ‘treatment’ or ‘exposure’ and Y the ‘outcome’. For unit i, we say that A i causes Y i if changing A i from one level, say A i = a*, to another level, A i = a, leads to a different outcome for Y i. We assume that A i occurs before Y i. To compare these outcomes, we use the notation $Y_i( \tilde{a})$, which represents the outcome for unit i under the treatment level $A_i = \tilde{a}$. To determine whether $Y_i( \tilde{a})$ quantitatively differs under two treatment levels on the difference scale, we would compute the contrast Y i(a*) − Y i(a). If Y i(a*) − Y i(a) ≠ 0, we would say there is a causal effect of A on Y for individual i.

Note that, for any given treatment of A for unit i, we can only observe one level of treatment. Therefore, we refer to Y i(a*) − Y i(a) ≠ 0 as a counterfactual contrast, or equivalently, as a contrast of potential outcomes. Because an individual may only receive one of two treatments at any given time, individual causal effects cannot generally be observed. However, when certain assumptions are satisfied, we may compute average treatment effects by aggregating individual observations under different treatment conditions. For a binary treatment, the difference in the average of the potential outcomes under two different treatment levels for the population from which a sample is drawn may be expressed as the difference in mean outcomes: 𝔼[Y(1)] − 𝔼[Y(0)] or equivalently as the average of the differences of the potential outcomes: 𝔼[Y(1) − Y(0)].

This counterfactual contrast represents the quantity obtained from an ideally conducted randomised controlled trial, where any common cause of the treatment and outcome would occur only by chance. Three fundamental assumptions for computing average treatment effects. Although these are typically satisfied in well-conducted randomised controlled trial, they may not be hold for observational or ‘real world’ data:

  1. (1) Causal consistency –The treatment level received by each unit matches the treatment defined by the protocol, remaining consistent within the treatment arms to be compared. There must be at least two arms for comparison.

  2. (2) (Conditional) exchangeability – Potential confounders that might affect the outcomes under different treatment conditions are balanced across all arms. In randomised controlled trials this balance is generally achieved by randomisation. In observational studies it must be addressed using methods of covariate adjustment.

  3. (3) Positivity (or Overlap) – For every combination of covariates in the target population, there is a non-zero probability of receiving each treatment condition. In other words, every unit with a given set of covariates must have some chance of being observed in all treatment arms being compared. This ensures that all subgroups in the population have exposure to each treatment condition, enabling comparisons across those subgroups. Violations of positivity occur when certain subgroups have no overlap in treatment conditions (e.g., a covariate perfectly predicts treatment assignment), which can lead to biased or unstable estimates of causal effects (refer to Westreich & Cole, Reference Westreich and Cole2010; Bulbulia et al., Reference Bulbulia, Afzali, Yogeeswaran and Sibley2023).

Note that real-world experiments may fail to meet these assumptions (Bulbulia, Reference Bulbulia2024c; Hernán & Robins, Reference Hernán and Robins2017). Practical limitations, such as imperfect randomisation, non-compliance, or measurement error, can lead to violations of consistency, exchangeability, or positivity. Our interest here is specifically restricted to causal inference using observational or ‘real world’ data, where these assumptions are often particularly challenging to satisfy.

Schematic workflow for inferring causal effects from real-world data before stating statistical estimators and performing statistical analysis

In causal inference, we do not apply statistical models to data until after we have stated a causal question and considered whether and how the question may be identified from the data. We take the following steps before considering a statistical estimator or estimation:

  1. (1) State a well-defined treatment – clearly define the treatment (or equivalently exposure) that states the hypothetical intervention to which all population members will be exposed. For example, the treatment ‘weight loss’ is a vaguely stated intervention because there are many ways one can lose weight – exercise, diet, depression, cancer, amputation and others (Hernán & Robins, Reference Hernán and Robins2024). The intervention ‘weight loss by at least 30 minutes of vigorous exercise each data’ is more clearly defined (Hernán et al., Reference Hernán, Alonso, Logan, Grodstein, Michels, Willett and Robins2008).

  2. (2) State a well-defined outcome – specify the outcome measure so that a causal contrast is interpretable. For example, ‘well-being’ is arguably a vaguely stated outcome. However, the outcome, ‘psychological distress measured one year after the intervention using the Kessler-6 distress scale’ (Kessler et al., Reference Kessler, Andrews, Colpe, Hiripi, Mroczek, Normand and Zaslavsky2002) is more clearly defined.

  3. (3) Clarify the target population – define the population to whom the results will generalise. The eligibility criteria for a study will define the source population from which units in the study are sampled. However, sampling from the source population may yield a study population that differs from the source population in variables that modify the effects of treatment (Bulbulia, Reference Bulbulia2024b; Dahabreh et al., Reference Dahabreh, Robins, Haneuse and Hernán2019; Dahabreh & Hernán, Reference Dahabreh and Hernán2019; Stuart et al., Reference Stuart, Ackerman and Westreich2018). Investigators may also seek to generalise beyond the source population, which requires additional assumptions and may require additional knowledge (Bareinboim & Pearl, Reference Bareinboim and Pearl2013; Dahabreh & Hernán, Reference Dahabreh and Hernán2019; Deffner et al., Reference Deffner, Rohrer and McElreath2022; Pearl & Bareinboim, Reference Pearl and Bareinboim2022; Westreich et al., Reference Westreich, Edwards, Lesko, Stuart and Cole2017).

  4. (4) Evaluate whether treatment groups, conditional on measured covariates, are exchangeable – the potential outcomes must be independent of treatments conditional on measured covariates (Angrist & Pischke, Reference Angrist and Pischke2009; Hernán & Robins, Reference Hernán and Robins2024; Morgan & Winship, Reference Morgan and Winship2014; Neal, Reference Neal2020).

  5. (5) Ensure treatments to be compared satisfy causal consistency – the versions of treatments over which a causal effect is estimated must be independent of the potential outcomes to be compared conditional on measured covariates (Hernán & Robins, Reference Hernán and Robins2024; T. J. VanderWeele & Hernán, Reference VanderWeele and Hernán2012, Reference VanderWeele and Hernán2013).

  6. (6) Check if the positivity assumption is satisfied – there must be a non-zero probability of receiving each treatment level at every level of covariate required to satisfy the conditional exchangeability assumption, and if there are many versions of treatment, to satisfy the causal consistency assumption (Westreich & Cole, Reference Westreich and Cole2010).

  7. (7) Ensure that the measures relate to the scientific questions at hand – in particular, evaluate structural features of measurement error bias (Bulbulia, Reference Bulbulia2024b; Hernán & Robins, Reference Hernán and Robins2024; T. J. VanderWeele & Hernán, Reference VanderWeele and Hernán2012).

  8. (8) Consider strategies to ensure the study group measured at the end of the study represents the target population – if the study population differs in the distribution of variables that modify the effect of a treatment on the outcome at both the beginning and end of treatment, the study will be biased when there is a treatment effect; as such, investigators must develop strategies to address attrition, non-response and structural sources of bias from measurement error (Bulbulia, Reference Bulbulia2024b; Hernán et al., Reference Hernán, Hernández-Díaz and Robins2004; Hernán & Robins, Reference Hernán and Robins2017; Hernán, Reference Hernán2017).

  9. (9) Clearly communicate the reasoning, evidence, and decision-making that inform steps 1–8 – provide transparent and thorough documentation of the decisions in steps 1–8. This includes detailing investigators’ causal assumptions and any disagreements about these assumptions (Ogburn & Shpitser, Reference Ogburn and Shpitser2021).

Conventions used in this article

Table 1 reports our variables. Table 2 describes our graphical conventions. Here, we use two types of graphical tools to clarify causal questions: causal directed acyclic graphs and single world intervention graphs. (Refer to supplementary materials S1 for a glossary for commonly used causal inference terms.)

Table 1. Terminology

Table 2. Elements of Causal Graphs

Throughout, for clarity, we repeat our definitions and graphical conventions as they are used. To begin, we define the following:

  • Node – or equivalently a ‘variable’, denotes properties or characteristics of units within a population. In causal directed acyclic graphs, we draw nodes with respect to features in a target population, which is the population for whom we seek causal inferences (Suzuki et al., Reference Suzuki, Shinozaki and Yamamoto2020). A time-indexed node, X t, allows us to index measurements within time intervals t ∈ 1…T, denoting relative chronology. If relative timing is not known, we may use X ϕt. The directions of arrows on a causal directed acyclic graph imply causation, and causation implies temporal order.

  • Arrow () – denotes a causal relationship from the node at the base of the arrow (a ‘parent’) to the node at the tip of the arrow (a ‘child’). In causal directed acyclic graphs, we refrain from drawing an arrow from treatment to outcome to avoid asserting a causal path from A to Y. Our purpose is to ascertain whether causality can be identified for this path. All other nodes and paths, including the absence of nodes and paths, are typically assumed.

  • Boxed variable $\fbox{X}$ – denotes conditioning or adjustment for X.

Judea Pearl demonstrated that causal dependencies in a directed acyclic graph could be evaluated using observable probability distributions according to rules known as ‘d-separation’ (Pearl, Reference Pearl1995, Reference Pearl2009).

The rules, presented in Table 3, are as follows:

  1. (1) Fork rule B and C are independent when conditioning on A: ($B\coprod C{\rm \mid }A$).

  2. (2) Chain rule – conditioning on B blocks the path between A and C: ($A\coprod C{\rm \mid }B$).

  3. (3) Collider rule A and B are independent until conditioning on C, which introduces dependence (A B|C.

From d-separation, Pearl derived a ‘backdoor adjustment theorem’, which provides a general identification algorithm given structural assumptions encoded in a causal directed acyclic graph Pearl (Reference Pearl2009): in a causal directed acyclic graph (causal DAG), a set of variables L satisfies the backdoor adjustment theorem relative to the treatment A and the outcome Y if L blocks every path between A and Y that contains an arrow pointing into A (a backdoor path). Formally, L must:

  1. (1) not be a descendant of A; and

  2. (2) block all backdoor paths from A to Y.

If L satisfies these conditions, the causal effect of A on Y is identified by conditioning on $\fbox{L}$ (Pearl, Reference Pearl2009). In what follows, I will assume that readers are familiar with causal directed acyclic graphs. Accessible introductions to causal directed acyclic graphs can be found in Pearl (Reference Pearl2009), Barrett (Reference Barrett2021), McElreath (Reference McElreath2020), Neal (Reference Neal2020), Hernán and Robins (Reference Hernán and Robins2024) and Bulbulia (Reference Bulbulia2024a). (For an introduction to single world intervention graphs, used below, refer to Richardson & Robins, Reference Richardson and Robins2013a, b.)

Table 3. Five elementary causal structures in a causal directed acyclic graph

Key: , a directed edge, denotes causal association. The absence of an arrow denotes no causal association. Rules of d-separation: In a causal diagram, a path is ‘blocked' or ‘d-separated’ if a node along it interrupts causation. Two variables are d-separated if all paths connecting them are blocked or if there are no paths linking them, making them conditionally independent. Conversely, unblocked paths result in ‘d-connected’ variables, implying statistical association. Refer to Pearl (Reference Pearl1995). Note that ‘d' stands for ‘directional’.

Implication: ${\cal G}$ denotes a causal directed acyclic graph (causal DAG). P denotes a probability distribution function. Pearl proved that independence in a causal DAG ($B\coprod C\vert A$)$_{\cal G}$ implies probabilistic independence ($B\coprod C\vert A$)P; likewise if ($B\coprod C\vert A$)P holds in all distributions compatible with ${\cal G}$, it follows that ($B\coprod C\vert A$)$_{\cal G}$ (refer to Pearl Reference Pearl2009, p. 61.) We read causal graphs to understand the implications of causality for relationships in observable data. However, reading causal structures from data is more challenging because the relationships in observable data are typically compatible with more than one (and typically many) causal graphs.

Part 1: interaction as ‘effect modification’

We have said that in causal inference, we must explicitly define our causal question before applying statistical models to data. What question might the analysis of interaction answer? In causal inference, we think of interaction in two ways:

  1. (1) Interaction as effect modification from a single intervention – we want to understand how an intervention varies in its effect across the strata of the target population in which we are interested. For example, we might ask: does the one-year effect of attending weekly religious service differ among people born in Australia compared with people born in Egypt? Note that here we do not imagine intervening on birthplace.

  2. (2) Interaction as joint intervention – we want to understand whether the combined effects of two treatments administered together differ from the separate effect of each treatment acting alone. For example, we might ask: does the one-year effect of attending weekly religious service and the one-year effect of being at least one standard deviation above population average wealth differ from not attending any religious service and being at the population average in wealth? Here there are two interventions that might act individually, separately, or in concert.

Part 1 considers interaction as effect modification. Readers who do not wish to use the ‘effect modification’ may prefer the term ‘moderation’.

Effect modification

First, we define the ‘sharp-null hypothesis’ as the hypothesis that there is no effect of the exposure on the outcome for any unit in the target population. Unless the sharp-null hypothesis is false, there may be effect modification (Bulbulia, Reference Bulbulia2024b). Clearly, the variability of causal effect modification cannot be assessed from descriptive measures of individuals in one's sample. Such heterogeneity must be modelled (refer to Tibshirani et al., Reference Tibshirani, Athey, Sverdrup and Wager2024; Vansteelandt & Dukes, Reference Vansteelandt and Dukes2022). Alternatively, we might seek to compare whether the effects of treatments vary by strata within the target population. For example we may ask whether effects vary by culture group membership, gender or another grouping variable.

Table 4 describes conventions to clarify how to ask a causal question of effect modification. We assume no confounding of the treatment on the outcome and that A has been randomised (i.e. ). As such, we will not use causal directed acyclic graphs to evaluate a treatment effect. We will assume .

Table 4. Graphical conventions we use for representing effect modification

To sharpen focus on our interest in effect modification, we will not draw a causal arrow from the direct effect modifier F to the outcome Y. This convention is specific to this article. (Refer to Hernán & Robins, Reference Hernán and Robins2024: 126–127, for a discussion of ‘non-causal’ arrows.)

In Table 5 G 1, we represent that F is a direct effect modifier for the effect of A on Y. The open arrow indicates that we are not attributing causality to F. Because our estimand does not involve intervening on Z, there is no need to close its backdoor paths. Note that if F were to affect A, we could still estimate the effect modification of A on Y because F has no causal interpretation. However, if A were to cause F, and F were to cause Y, then by the chain rule (recall Table 2 G 4), conditioning on F would bias the effect estimate of A on Y.

Table 5. Effect Modification

A denotes the treatment;

Y denotes the outcome;

U denotes an unmeasured confounder;

F denotes a direct-effect modifier;

{G, B} denote indirect effect modifiers of ;

asserts causality;

$\fbox{X}$ indicates conditioning on variable X;

indicates variables conditioned upon is an effect modifier (direct or indirect) of ;

indicates effect-modifier F is not conditioned upon;

Observation 1: classifying an indirect-effect modifier depends on the structure of causation and model specification.

Observation 2: Whether we condition on F or not, differences in the distribution of effect-modifiers within the sample population compared to the target population, specifically where these effect-modifiers interact with the causal path from , may introduce target validity bias. Because target-validity bias is indifferent to conditioning on F, we may represent effect modification without reference to whether F is included in the model:

In Table 5 G 2, we represent that F is an unobserved direct effect modifier of A to Y. When the distribution of direct effect modifiers F differs between two populations and effect modification is non-linear, marginal treatment effects between populations will generally differ and will not easily transport from one population to another. The concept of an average treatment effect has no meaning without a population over which the effect marginalises. This point, although obvious, has profound implications when investigators seek to assess whether their research generalises; refer to Hernán and Robins (Reference Hernán and Robins2024) and Bulbulia (Reference Bulbulia2024b). For example, if the study population differs in the distribution of features that modify a treatment effect, and no correction is applied, effect estimates will be biased for the target population in at least one measure of effect (Bulbulia, Reference Bulbulia2024b; Greenland, Reference Greenland2009; Lash et al., Reference Lash, Rothman, VanderWeele and Haneuse2020)

We present two candidate effect modifiers in Table 5 G 3. Notice that whether a variable is an effect modifier also depends on which other variables are included in the model. Here, F is a direct effect modifier and G, a descendant of F, is an indirect effect modifier. Suppose we were interested in whether treatment effects vary (on the difference scale) within levels of F. For example, let F denote childhood deprivation, G denote educational achievement, A denote a government educational initiative and Y denote recycling. If we were to condition on F, we would not observe effect modification by education G for the effect of the government initiative A on recycling behaviour Y : $\fbox{F}$ blocks the path .

We present the same causal structure in Table 5 G 4. However, we do not condition on the direct effect modifier F, but rather condition only on G, the indirect effect modifier. In this scenario, we would find that the effectiveness of the government initiative A on recycling behaviour Y varies by educational achievement G. Thus, we would observe G as an effect modifier because this path is open: .

In Table 5 G 5, suppose we add another variable to our model, depression, denoted by B. We imagine B to be a stable trait or that investigators measured childhood depression (that is, B precedes G). Suppose we do not condition on the direct effect modifier F (childhood deprivation), but we condition on educational attainment (G) and depression (B). In this graph, G is a collider of F and B. Thus, conditioning on G (but not F) opens a path from . The investigators would find evidence for effect modification by depression on the effectiveness of the government intervention A on recycling (Y). However, they should not interpret this result to mean that if levels of depression were to change within the population the treatment effect would change. B is not causally related to Y in this scenario. Here, association is not causation.

In Table 5 G 6, we will not find evidence for effect modification for B and G because conditioning on F blocks the flow of information that was open in G 4 and G 5. This again underscores the relativity of effect modification to (1) the structure of causality in the world and (2) an investigator's statistical modelling strategy.

These examples reveal the power – and simplicity – of causal diagrams to ‘transform the obvious’. Using causal directed acyclic graphs and Pearl's rules of d-separation, it is clear that the analysis of effect modification cannot be conducted without reference to an assumed causal order and an explicit statement about which variables within that order investigators have included in their statistical models (T. J. VanderWeele, Reference VanderWeele2012). Investigators and policymakers may make incorrect inferences and policy decisions if they do not understand the relevance of effect modification to such parameters. It is important to remember that when evaluating evidence for effect modification, we are not assessing the effects of intervening on variables other than the treatment. Instead, we qualitatively evaluate whether treatment effects vary across subgroups. For more on effect modification, refer to T. J. VanderWeele (Reference VanderWeele2012), T. J. VanderWeele and Robins (Reference VanderWeele and Robins2007) and Suzuki et al. (Reference Suzuki, Mitsuhashi, Tsuda and Yamamoto2013).

Example showing scale dependence of effect modification

Suppose investigators are interested in whether treatment varies across levels of another variable, an effect modifier. We next illustrate how causal inferences about the presence or absence of effect modification depend on the scale that is used to measure the contrast. We show that an effect modifier on the ratio scale may not be an effect modifier on the difference scale, and vice versa.

Recall individual treatment effects are not observed. Assume a binary treatment is randomised, and we have A = a ∈ {0, 1}. Investigators are interested in comparing the magnitude of this treatment effect across two levels of F = f ∈ {0, 1}.

We define the average treatment effects for each group under each intervention as follows:

$$\mathbb{E}[ Y{\rm \mid }A = 0, \;F = 1] = \mu _{01}, \;\quad \mathbb{E}[ Y{\rm \mid }A = 1, \;F = 1] = \mu _{11}$$
$$\mathbb{E}[ Y{\rm \mid }A = 0, \;F = 0] = \mu _{00}, \;\quad \mathbb{E}[ Y{\rm \mid }A = 1, \;F = 0] = \mu _{10}$$

The treatment effect for each group on the difference scale (absolute scale) is given by:

$${\rm AT}{\rm E}_{F = 1} = \mu _{11}-\mu _{01}$$
$${\rm AT}{\rm E}_{F = 0} = \mu _{10}-\mu _{00}$$

The treatment effect on the ratio scale (relative scale) for each group is:

$${\rm R}{\rm R}_{F = 0} = \displaystyle{{\mu _{10}} \over {\mu _{00}}}$$
$${\rm R}{\rm R}_{F = 1} = \displaystyle{{\mu _{11}} \over {\mu _{01}}}$$

We say that there is effect modification on the difference scale if:

We say that there is effect modification on the ratio scale if:

We have stated each causal question in relation to well-defined causal contrast and population, here defined by membership in F.

Imagine we obtain the following estimates from our study:

Outcomes under A = 0:

  • μ 00 = 5

  • μ 01 = 15

Outcomes under A = 1:

  • μ 10 = 10

  • μ 11 = 20

Next, we calculate the treatment effects on the difference and ratio scales for each group:

Difference scale

$${\rm AT}{\rm E}_{F = 0} = \mu _{10}-\mu _{00} = 10-5 = 5$$
$${\rm AT}{\rm E}_{F = 1} = \mu _{11}-\mu _{01} = 20-15 = 5$$

Both groups have the same treatment effect on the difference scale, ATEF=0 = ATEF=1 = 5. investigators conclude there is no evidence for effect modification on the difference scale.

Ratio scale:

$${\rm R}{\rm R}_{F = 0} = \displaystyle{{\mu _{10}} \over {\mu _{00}}} = \displaystyle{{10} \over 5} = 2.00$$
$${\rm R}{\rm R}_{F = 1} = \displaystyle{{\mu _{11}} \over {\mu _{01}}} = \displaystyle{{20} \over {15}}\approx 1.33$$

The treatment effect on the ratio scale is different between the two groups, RRF=0 = 2 ≠ RRF=1 ≈ 1.33. Hence, investigators find evidence for effect modification on the ratio scale.

The discrepancy in evidence for effect modification depending on the scale we choose arises because the two scales measure different aspects of the treatment effect: the absolute difference in outcomes vs. the relative change in outcomes. Parallel considerations apply to the analysis of interaction, where we imagine a joint intervention. For this reason, it is important to state the causal effect scale of interest in advance of estimation (Bulbulia, Reference Bulbulia2024a). We next consider interaction as a joint intervention.

Part 2: interaction

Introducing single world intervention graphs

When evaluating evidence for interaction, we must assess whether the combined effects of two treatments differ from the unique effects of each treatment relative to a baseline where neither treatment is administered. Understanding multiple interventions can be facilitated by using Single World Intervention Graphs (SWIGs) (Richardson & Robins, Reference Richardson and Robins2013a).

Single world intervention graphs employ Pearl's rules of d-separation but offer additional benefits by graphically representing the complex factorisations required for identification, presenting distinct interventions in separate graphs. The first advantage is greater precision and clarity: SWIGs allow us to consider identification conditions for each counterfactual outcome individually. Such precision is useful because identification conditions may differ for one, but not another, of the treatments to be compared. Node-splitting also makes it easier to determine identification conditions that are obscured in causal directed acylic graphs, for an example refer to supplementary materials S2. The second advantage is that single world intervention graphs unify the potential outcomes framework with Pearl's structural causal model framework: any causal relationship that can be represented in a causal directed acyclic graph can also be represented in a SWIG (Richardson & Robins, Reference Richardson and Robins2013b).

Single world intervention graphs work by node-splitting

We create a single world intervention graph by ‘node-splitting’ at each intervention such that the random variable that is intervened upon is presented on one side and the level at which the random variable is fixed is presented on the other.

Consider a template graph Table 6 G. Applying node-splitting to A involves creating separate graphs for each value of A to be contrasted.

  1. (1) SWIG for A = 0 – denoted as ${\cal G}_{A = 0}$, this graph shows the hypothetical scenario where A is set to 0.

  2. (2) SWIG for A = 1 – denoted as ${\cal G}_{A = 1}$, this graph shows the hypothetical scenario where A is set to 1.

Table 6. Single World Interventions Graphs $( {{\cal G}_{3-4}} ) $ present separate causal diagrams for each treatment to be contrasted. A Single World Intervention Template $( {{\cal G}_2} ) $ is a ‘graph value function’ that produces the individual counterfactual graphs (Richardson & Robins, Reference Richardson and Robins2013a). On the other hand, causal directed acyclic graphs, such as $( {{\cal G}_1} ) $, require positing interventional distributions. The formalism underpinning these interventional distributions is mathematically equivalent to formalism underpinning the potential outcomes framework, assuming the errors of the underlying structural causal models that define the nodes on which interventions occur are independent (Richardson & Robins, Reference Richardson and Robins2013a). Single World Intervention Graphs (SWIGs), however, permit the comparison of distinct interventions in our causal diagram without requiring that the non-parametric structural equations that correspond to nodes on a causal graph have independent error structures. This is useful when attempting to identify the causal effects of sequential treatments, refer to supplementary materials S2

Key: ${\cal G}_ 1$ is a causal DAG: Y denotes the outcome; A denotes treatment; U denotes an unmeasured confounder. denotes an assumed causal path. $\fbox{L}$ denotes conditioning on measured variables. Conditioning on L is sufficient to block the backdoor path between A and Y. We do not draw a path to underscore d-separation. Counterfactuals are not drawn on a causal DAG. We assume all nodes may be intervened upon. ${\cal G}_ 2$ This represents a Single World Intervention Template (SWIT). The rules of d-separation remain; however, we factorise counterfactuals separately for each treatment by ’node-splitting’ at an intervention. This distinguishes the random intervention variable from the fixed intervention, and descendants under fixed interventions are relabelled using potential outcomes notation. For example, is represented directly on our causal diagram.

To factorise potential outcomes, we must create separate Single World Intervention Graphs (SWIGs). Templates are ‘graph-valued functions’ that allow us to do this. Here, template ${\cal G}_ 2$ produces SWIG ${\cal G}_ 3$ and SWIG ${\cal G}_ 4$. We use the convention to denote closed confounding paths after conditioning on measured covariates and fixing interventions.

Assuming the structure of the world is faithfully captured in both causal DAG ${\cal G}_ 1$ and the SWIGs ${\cal G}_{3-4}$, we find $Y( a = 0) \coprod A{\rm \mid }L$ and $Y( a = 1) \coprod A{\rm \mid }L$$Y( a = 0) \coprod A{\rm \mid }L$

In these graphs, the node corresponding to the treatment A is split, relabelled with the random and fixed component, and then each node that follows is labelled with the fixed component until the next intervention. Here, $Y( \tilde{a})$is the only variable to follow A and it is relabelled either Y(0) or Y(1) corresponding to whether A = 1 orA = 0; hence $Y( \tilde{a})$ is relabelled as either Y(0) or Y(1). Note that we do not place both Y(0) and Y(1) on the same SWIG because the variables are not jointly observed. Hence, we evaluate identification for $Y( 0) \coprod A = 0\vert L$ and $Y( 1) \coprod A = 1\vert L$ separately.

Interaction as a joint intervention

We now use Single World Intervention Graphs (SWIGs) to clarify the concept of causal interaction as a joint intervention. Consider two treatments, denoted as A and B, and a single outcome, Y. Causal interaction as a joint interaction examines whether the combined effect of A and B on Y (denoted as Y(a, b)) differs from what would be expected from their individual effects. Specifically, we assess whether the effect of both treatments together is equal to, greater than, or less than the sum of their individual effects on the additive scale.

First, we obtain the expected outcomes when the entire target population is treated at each level of the treatments to be compared. These potential outcomes are illustrated in Table 7:

  • Table 7 G 1 – neither treatment A nor treatment B is given.

  • Table 7 G 2 – both treatment A and treatment B are given.

  • Table 7 G 3 – treatment A is given, and treatment B is not given.

  • Table 7 G 4 – treatment A is not given, and treatment B is given.

Table 7.

Key: Y denotes the outcome; A denotes treatment; denotes an assumed causal path; denotes blocked path by conditioning on L A and L B

Single World Intervention Graphs show counterfactual states of the world under different interventions

To evaluate each intervention, we split a node into its random and deterministic components and represent descendants under fixed interventions using potential outcomes notation.

We require four counterfactual outcomes to evaluate interaction: joint treatment: $\mathbb{E}[Y(1,1)]$; joint treatment: $\mathbb{E}[Y(a = 1,b = 1)]$; Only treat with A: $\mathbb{E}[Y(a = 1,b = 0)]$; Only treat with B: $\mathbb{E}[Y(a = 0,b = 1)]$; Treat neither with A nor with B: $\mathbb{E}[Y(a = 0,b = 0)]$.

By comparing these expected outcomes, we can determine the presence and nature of causal interaction between treatments A and B with respect to the outcome Y.

Example

Consider the effect of beliefs in big Gods (exposure A) and a culture's monumental architecture (exposure B) on social complexity (outcome Y). Both interventions have equal status; we are not investigating effect modification of one by the other. The interventions must be well defined. We must state, understand and obtain measures for the quantities ‘big Gods’, ‘monumental architecture’ and ‘social complexity’ at specified time intervals after the interventions are first observed.

We need to state a population and scale to assess the individual and combined effects of A and B. Suppose our population consists of societies of primary urban genesis (Wheatley, Reference Wheatley1971). Suppose further we seek evidence for causal interaction on the additive (difference) scale. Evidence for interaction would be present if the following inequality were to hold. Where,

  • 𝔼[Y(1, 1)]– mean outcome for those jointly exposed to both treatments, big Gods and big architecture;

  • 𝔼[Y(1, 0)] – mean outcome for those exposed only to the treatment of big Gods;

  • 𝔼[Y(0, 1)] – mean outcome for those exposed only to the treatment of big architecture; and

  • 𝔼[Y(0, 0)] – mean outcome for those exposed to neither treatment of big Gods nor of big architecture.

Evidence for interaction on the additive scale if

$$( \underbrace{{\mathbb{E}[ Y( 1, \;1) ] }}_{{{\rm joint\ exposure}}}-\underbrace{{\mathbb{E}[ Y( 0, \;0) ] }}_{{{\rm neither\ exposed}}}) -[ (\! \underbrace{{\mathbb{E}[ Y( 1, \;0) ] }}_{{{\rm only\ A\ exposed}}}-\underbrace{{\mathbb{E}[ Y( 0, \;0) ] }}_{{{\rm neither\ exposed}}}) + ( \underbrace{{\mathbb{E}[ Y( 0, \;1) ] }}_{{{\rm only\ B\ exposed}}}-\underbrace{{\mathbb{E}[ Y( 0, \;0) ] }}_{{{\rm neither\ exposed}}}) ] \ne 0$$

Simplifying:

$$\underbrace{{\mathbb{E}[ Y( 1, \;1) ] }}_{{{\rm joint\ exposure}}}-\underbrace{{\mathbb{E}[ Y( 1, \;0) ] }}_{{{\rm only\ A\ exposed}}}-\underbrace{{\mathbb{E}[ Y( 0, \;1) ] }}_{{{\rm only\ B\ exposed}}} + \underbrace{{\mathbb{E}[ Y( 0, \;0) ] }}_{{{\rm neither\ exposed}}}\ne 0$$

A positive value indicates evidence for a synergistic (super-additive) interaction. A negative value indicates evidence for a sub-additive interaction. A value close to zero implies no interaction on the additive scale.

Table 7 presents each counterfactual intervention. To identify these causal effects, we need to adjust for all confounders of the relationships between A, B, and Y. This includes any variables that influence both A and Y, both B and Y, both A and B, or all three variables simultaneously.

As with effect modification, evidence for causal interaction may differ depending on the measurement scale one chooses to assess it (T. J. VanderWeele, Reference VanderWeele2012; T. J. VanderWeele & Knol, Reference VanderWeele and Knol2014). For most policy settings, the additive scale is recommended because it directly relates to differences in outcome levels, which are often more actionable (see T. J. VanderWeele & Knol, Reference VanderWeele and Knol2014).

Note that if A and B potentially influence each other over time, we would need to collect longitudinal data and estimate causal effects using mediation analysis. Indeed, if there has been a co-evolution of religious culture, monumental architecture, and social complexity – as archaeologists have long reported (De Coulanges, Reference De Coulanges1903; Wheatley, Reference Wheatley1971) – mediation analysis may be more appropriate. However, the requirements for causal mediation analysis are more stringent than those for causal interaction analysis, which we will consider next.

Part 3: causal mediation analysis

In 1992, Robins and Greenland clarified the objectives of interpretable causal mediation analysis: to decompose the total effect into natural direct and indirect effects within a set of hypothetical interventions, contrasting their counterfactual outcomes (J. M. Robins & Greenland, Reference Robins and Greenland1992). This landmark paper has been to mediation analysis what On The 0rigin of Species has been to evolutionary biology (Darwin, Reference Darwin1859). However, mediation analysis in the human sciences remains rife with confusion. The primary source of this confusion is the application of statistical models to data without first defining the causal quantities of interest. Associations derived from statistical mediation analysis do not necessarily imply causation and are typically uninterpretable. This section considers how to formulate causal questions in mediation analysis.

Defining a mediation analysis estimand

To understand causal mediation, we deconstruct the total effect into natural direct and indirect effects. Again, the total effect of treatment A on outcome Y is defined as the difference between potential outcomes when the treatment is applied vs. when it is not. The estimand for the total (or average, or ‘marginal’) treatment effect is given by:

$${\rm Total\;Treatment\;Effect} = \mathbb{E}[ Y( 1) ] -\mathbb{E}[ Y( 0) ] $$
$$\mathbb{E}[ Y( 1) ] = \mathbb{E}[ Y( 1, \;M( 1) ) ] $$

The total effect can be further decomposed into direct and indirect effects, addressing questions of mediation. The potential outcome Y(1), considering the mediator, expands to:

This considers the effect of the exposure A = 1 and the mediator at its natural value when A = 1. Similarly, the potential outcome 𝔼[Y(0)], considering the mediator, expands to:

$$\mathbb{E}[ Y( 0) ] = \mathbb{E}[ Y( 0, \;M( 0) ) ] $$

This quantity denotes the effect of exposure A = 0 and the mediator at its natural value when A = 0.

Next, we clarify our estimand by decomposing the total effect into the natural direct effect and the natural indirect effect.

The natural direct effect is the effect of the treatment on the outcome while maintaining the mediator at the level it would have been if the treatment had not been applied:

$$ \text{Natural Direct Effect} = {\textcolor{blue}{{\mathbb{E}}[Y(1, M(0))]}} - {\mathbb{E}}[Y(0, M(0))] $$

Here, the counterfactual quantities not directly realised in the data are highlighted in blue: 𝔼[Y(1, M(0))]. Notice we add this term to the potential outcomes when A = 0, recalling 𝔼[Y(0, M(0))] = Y(0).

The natural indirect effect is the effect of the exposure on the outcome that is mediated. To obtain these quantities, we compare the potential outcome Y under treatment, where the mediator assumes its natural level under treatment, with the potential outcome when the mediator assumes its natural value under no treatment:

$$ \text{Natural Indirect Effect} = {\mathbb{E}}[Y(1, M(1))] - \textcolor{blue}{{\mathbb{E}}[Y(1, M(0))]} $$

Here, the counterfactual quantities not directly realised in the data are again highlighted in blue: 𝔼[Y(1, M(0))]. Notice we subtract this term from the potential outcomes when A = 1, recalling 𝔼[Y(1, M(1))] = 𝔼[Y(1)].

By rearranging this decomposition, we find that the total effect is the sum of the natural direct effect and natural indirect effect. This is shown by adding and subtracting the term 𝔼[Y(1, M(0))] in our equation:

$$ \text{Total Effect (TE)} = \underbrace{\bigg\{{\mathbb{E}}[Y(1, M(1))] - \textcolor{blue}{{\mathbb{E}}[Y(1, M(0))]}\bigg\}}_{\text{Natural Indirect Effect (NIE)}} + \underbrace{\bigg\{\textcolor{blue}{{\mathbb{E}}[Y(1, M(0))]} - {\mathbb{E}}[Y(0, M(0))]\bigg\}}_{\text{Natural Direct Effect (NDE)}} $$

Table 8 presents a conceptual challenge for causal mediation analysis. Suppose we randomise a binary treatment A ∈ {0, 1}. Although randomising A does not ensure that there is no confounding of the mediator/outcome path, we assume no unmeasured confounding for either the treatment or the mediator. (We will relax this assumption in the next section.)

Table 8. In causal mediation, the quantities that we require to obtain natural direct and indirect effects, namely $\mathbb{E}[Y(1,M(0))]$, cannot be experimentally observed because we cannot treat someone and observe the level of their mediator if they were not treated

Key: Y denotes the outcome, A denotes treatment, and M denotes the mediator. ${\cal R}$ denotes randomisation into treatment, and indicates an assumed causal path. For each intervention being contrasted, we split a node into its random and deterministic components and use potential outcomes notation to represent descendants under fixed interventions.

Table 8 G 1 is a single world intervention Template (SWIT), which generates SWIGs for each condition.

Table 8 G 2 presents counterfactual outcomes for condition A = 0; here, the natural value of M is M(a = 0), and the counterfactual outcome is given by Y(0, M(0)).

Table 8 G 3 presents counterfactual outcomes for condition A = 1; here, the natural value of M is M(a = 1), and the counterfactual outcome is given by Y(1, M(1)).

These SWIGs clarify that we cannot identify natural direct and indirect effects from observations on individual units under treatment because 𝔼[Y(1, M(0))] is not observable (Shi et al., Reference Shi, Choirat, Coull, VanderWeele and Valeri2021; Steen et al., Reference Steen, Loeys, Moerkerke and Vansteelandt2017; Valeri et al., Reference Valeri, Lin and VanderWeele2014; T. J. VanderWeele, Reference VanderWeele2015; T. VanderWeele & Vansteelandt, Reference VanderWeele and Vansteelandt2014; Vansteelandt et al., Reference Vansteelandt, Bekaert and Lange2012). Expressing these quantities requires a counterfactual framework. Here, we see that a counterfactual formulation of mediation analysis has made the familiar strange. However, under assumptions, we can sometimes recover natural direct and indirect effects from data (T. J. VanderWeele, Reference VanderWeele2015), given that our interest is in contrasts obtained for the target population, not for individuals, where we assume no causal effects are directly observed.

Assumptions of causal mediation analysis

Table 9 G 1 presents a Single World Intervention Template (SWIT) that specifies the assumptions required for inferring natural direct and indirect effects. This template highlights that, when estimating natural mediated effects, we only intervene on the treatment. Therefore, we must infer the mediated effect of the treatment under the condition that the mediator is set to the level it would naturally take under the control condition.

Table 9. Assumptions of Causal Mediation Analysis

Key: Y denotes the outcome; A denotes treatment; M denotes the mediator; L is the measured confounder of the path A → Y; Q is the measured confounder of the path A → M; N is the measured confounder of the path M → Y ; denotes a causal path; denotes a blocked backdoor path.

The assumptions of causal mediation

1. No unmeasured treatment-outcome confounder, hence we condition on L;

2. No unmeasured treatment-mediator confounder, hence we condition on Q;

3. No unmeasured mediator-outcome confounder, hence we condition on Z;

4. No mediator-outcome confounder affected by the exposure — a problem because we must not condition on Z to estimate the natural indirect effect.

Randomisation only guaranteed assumptions 1 and 2 ${\rm {\cal G}1}$: Z is a confounder of the M to Y path. We must, therefore, condition on Z. However Z is also a mediator of the A to M path. We must, therefore, not condition on Z. In this setting, there is no identification of natural indirect and direct effects of A through M because Z is both a confounder and an intermediary mediator. We may, however, estimate natural direct and indirect for the combined mediated effect of M and Z. We may use random draws from distributions of $A \to Z$ and A to M to obtain analogues of natural direct and indirect effects. ${\cal G}_ 2$: If assumptions 1 and 3 are satisfied, controlled direct effects can be identified in the presence of intermediary confounding. This requires fixing the mediator to some level M = m and evaluating the direct (unmediated) effect of treatment at this level of the mediator. However, in this setting, there is no longer a decomposition of the total effect into the natural direct and natural indirect effects for a controlled direct effect.

Additionally, Table 9 G 1 also clarifies the assumptions needed for inferring controlled direct effects, where the mediator is fixed to a level specified by the investigators. In this scenario, we obtain causal contrasts by fixing variables to specific states.

Consider the hypothesis that cultural beliefs in ‘big Gods’ influence social complexity, with political authority mediating. Assuming that we have well-defined interventions and outcomes, what requirements are necessary to decompose this causal effect into natural direct and indirect effects?

  1. (1) No unmeasured exposure-outcome confounder – this requirement is expressed as $Y( a, \;m) \coprod A\vert L$. After accounting for the covariates in set L, there must be no unmeasured confounders influencing cultural beliefs in big Gods (A) and social complexity (Y). For example, if our study examines the causal effect of cultural beliefs in big Gods on social complexity, and the covariates in L include factors such as geographic location and historical context, we need to ensure that these covariates effectively block any confounding paths between A and Y. The relevant path in Table 9 G 1 is the confounder of the path A → Y.

  2. (2) No unmeasured mediator-outcome confounder – this requirement is expressed as $Y( a, \;m) \coprod M\vert Z$. After controlling for the covariate set Z, we must ensure that no other unmeasured confounders affect political authority (M) and social complexity (Y). For instance, if trade networks affect political authority and social complexity, we must account for trade networks to block the path linking our mediator and outcome. The relevant path in Table 9 G 1 is the confounder of the path M → Y.

  3. (3) No unmeasured exposure-mediator confounder – this requirement is expressed as:$M( a) \coprod A\vert Q$. After controlling for the covariate set Q, we must ensure that no additional unmeasured confounders affect cultural beliefs in big Gods (A) and political authority (M). For example, the capability to construct large ritual theatres may influence the belief in big Gods and the level of political authority. If we have indicators for this technology measured before the emergence of big Gods (these indicators being Q), we must assume that accounting for Q closes the backdoor path between the exposure and the mediator. The relevant path in Table 9 G 1 is the confounder of the path A → M.

  4. (4) No mediator-outcome confounder affected by the exposure – this assumption requires that there are no unmeasured confounders of the mediator-outcome relationship that are themselves affected by the exposure A. Such confounders cannot be adjusted for without introducing bias. The relevant path in Table 9 G 1 involves confounders of the $M \rightarrow Y$ path that are influenced by A. Satisfying assumption 4 imposes considerable demands on causal mediation analysis. When the exposure influences a confounder of the mediator and outcome, we face a dilemma. Without adjusting for this confounder, a backdoor path between the mediator and the outcome remains open, introducing bias. In this setting, we cannot recover the natural direct and indirect effects from observational data. We may may need to settle for investigating controlled direct effects, estimate jointly mediated effects of Z and M together, or consider alternative estimands as suggested by VanderWeele and others (Díaz et al., Reference Díaz, Williams and Rudolph2023b; J. M. Robins & Richardson, Reference Robins and Richardson2010; T. J. VanderWeele et al., Reference VanderWeele, Vansteelandt and Robins2014; T. J. VanderWeele, Reference VanderWeele2015; T. J. VanderWeele & Tchetgen Tchetgen, Reference VanderWeele and Tchetgen Tchetgen2017; Vo et al., Reference Vo, Williams, Liu, Rudolph and Díaz2024).

Notice that even when assumptions 1–4 are satisfied, natural direct effect estimates and natural indirect effect estimates require conceptualising a counterfactual that is never directly observed on any individual, namely: Y(1, M(0)). Such effects are only identified in distribution (refer to T. J. VanderWeele, Reference VanderWeele2015).

Controlled direct effects

Consider another identification challenge, as described in template Table 9 G 1. Suppose we aim to understand the effect of a stringent pandemic lockdown, A, on psychological distress, Y, focusing on trust in government, M, as a mediator. Further, suppose that pandemic lockdowns may plausibly influence attitudes towards the government through pathways that also affect psychological distress. For instance, people might trust the government more when it provides income relief payments, which may also reduce psychological distress.

Under the rules of d-separation, conditioning on income relief payments (Z) could block necessary causal paths. If we adjust for Z, we might block part of the effect of A on M and Y that operates through Z, potentially biasing the estimation of the natural indirect effect. The paths and are blocked. If we do not adjust for Z, Z acts as an unmeasured confounder of the $M \rightarrow Y$ relationship since Z influences both M and Y. The path . remains open, introducing bias.

In such a scenario, it would not be feasible to consistently decompose the total effect of the exposure (pandemic lockdowns) on the outcome (psychological distress) into natural indirect and direct effects. However, if all other assumptions were to hold, we might obtain an unbiased estimate for the controlled direct effect of pandemic lockdowns on psychological distress as a fixed level of government trust.

Table 9 G 2 presents the weaker assumptions required to identify a controlled direct effect. We might examine the effect of the pandemic lockdown if we could intervene and set everyone's trust in government to, say, one standard deviation above the baseline, compared with fixing trust in government to the average level at baseline. We might use modified treatment policies (described below) that specify interventions as functions of the data. For instance, we might investigate interventions that ‘shift only those whose mistrust of government was below the mean level of trust at baseline and compare these potential outcomes with those observed’. Asking and answering precisely formulated causal questions such as this might lead to clearer policy advice, especially in situations where policymakers can influence public attitudes towards the government; see Williams and Díaz (Reference Williams and Díaz2021), Díaz et al. (Reference Díaz, Williams, Hoffman and Schenck2023a) and Hoffman et al. (Reference Hoffman, Schenck, Satlin, Whalen, Pan, Williams and Díaz2022, Reference Hoffman, Salazar-Barreto, Rudolph and Díaz2023).

In any case, I hope this discussion of causal mediation analysis clarifies that it would be unwise to simply examine the coefficients obtained from statistical structural equation models and interpret them as meaningful. To answer any causal question, we must first state it, with respect to clearly defined counterfactual contrasts and a target population. Once we state our causal question, we find have no guarantees that the coefficients from statistical models are straightforwardly interpretable (T. J. VanderWeele, Reference VanderWeele2015).

For those interested in statistical estimators for causal mediation analysis, I recommend visiting the CMAverse website: https://bs1125.github.io/CMAverse/articles/overview.html (accessed 12 December 2023). This excellent resource provides comprehensive documentation, software and practical examples, including sensitivity analyses. Next, we will consider more complex scenarios that involve feedback between treatments and confounders across multiple time points – settings in which traditional statistical methods also fail to provide valid causal inferences.

Part 4: time-fixed and time-varying sequential treatments (treatment strategies, modified treatment policies)

Our discussion of causal mediation analysis focused on how effects from two sequential treatments – the initial treatment and that of a mediator affected by the treatment – may combine to affect an outcome.

This concept can be expanded to investigate the causal effects of multiple sequential exposures, referred to as ‘treatment regimes’, ‘treatment strategies’ or ‘modified treatment policies’. Researchers often use longitudinal growth and multi-level models in many human sciences, where longitudinal data are collected. How may we identify such effects? What do they mean?

As before, to answer a causal question, we must first clearly state it. This involves specifying the counterfactual contrast of interest, including the treatments to be compared, the scale on which the contrast will be computed and the target population for whom inferences are valid. Without such clarity, our statistical models are often uninterpretable.

Worked example: does marriage affect happiness?

Richard McElreath considers whether marriage affects happiness and provides a simulation to clarify how age structure complicates causal inferences (McElreath, Reference McElreath2020: 123–144). We expand on this example by first clearly stating a causal question and then considering how time-varying confounding invalidates the use of standard estimation methods such as multi-level regression.

Let A t = 1 denote the state of being married at time t and A t = 0 denote the state of not being married, where t ∈ {0, 1, τ} and τ is the end of the study. $Y_\tau$ denotes happiness at the end of study. For simplicity, assume that this concept is well defined and measured without error.

Tables 10 and 11 reveal four treatment strategies and six causal contrasts that we may estimate for each treatment strategy combination.

Table 10. Four fixed treatment regimens in time-series data where exposure varies

Table 11. Six causal contrasts in time-series data where exposure varies

To answer our causal question, we need to:

  1. (1) Specify treatments – define the treatment strategies being compared (e.g. always married vs. never married).

  2. (2) Define the contrast – state the counterfactual contrast of interest (e.g. 𝔼[Y(1, 1) − Y(0, 0)]).

  3. (3) Identify the population – specify the population for which the inferences are valid (e.g. adults aged 20–40).

Time-varying confounding with treatment-confounder feedback

Table 12 G 1 and Table 12 G 2 represent two subsets of possible confounding structures for a treatment regime conducted over two intervals. Covariates in L t denote measured confounders, and U denotes unmeasured confounders. A t denotes the treatment, ‘Marriage Status,’ at time t. Y denotes ‘Happiness’ measured at the end of the study.

Table 12. Single World Intervention Graph for Sequential Treatments

Key: Y represents the outcome, A t the treatment, U unmeasured confounders, and L t measured confounders ensuring $A_t\coprod Y {\rm \mid }L_{t-1}$. Researchers aim to determine if $Y( \bar{a}) \coprod {\overline {A_t} } {\rm \mid }\overline {H_t} $, with A t as a sequence of treatments and $H_t = ( \bar{A}_{t-1}, \;\bar{L}_t) $ indicating past treatment and confounder histories. Counterfactual contrasts are predefined for static treatments and adaptive for dynamic treatments. indicates causal paths; shows confounding paths; In Single World Intervention Graphs, $A = \tilde{a}; \;\tilde{a}\in { 0, \;1} $; highlights dependencies from time-varying treatments (modified treatment policies) in dynamic Single World Intervention Graphs (or templates). ${\cal G}$1–2 are causal directed acyclic graphs. All paths are omitted. indicates conditioning on a confounder, which opens a backdoor path from treatment to outcome. ${\cal G}$3–5, employ a Single World Intervention Template (SWIT), we apply d-separation to each intervention to be contrasted. We split each node into random and deterministic components, with descendants labelled as potential outcomes under treatment.

Consider the structure of confounding presented in Table 12 G 1. To close the backdoor path from A 1 to Y, we must condition on L 0. To close the backdoor path from A 2 to Y, we must condition on L 2. However, L 2is a collider of treatment A 1 and unmeasured confounders, such that conditioning on L 2 opens a backdoor path between A 1 and Y. This path is highlighted in red: .

If Table 12 G 1 faithfully represents causality, it might seem that we cannot obtain valid inferences for any of the six causal contrasts we have defined. Indeed, using standard methods, we could not obtain valid causal inferences. However, J. Robins (Reference Robins1986) first described a consistent estimation function that can be constructed where there is time-varying confounding (Hernán et al., Reference Hernán, Hernández-Díaz and Robins2004; J. Robins & Hernán, Reference Robins and Hernán2008).

Table 12 G 3 presents a single world intervention template that clarifies how identification may be obtained in fixed treatment regimes where there is time-varying confounding as observed in Table 12 G 1. When constructing a single world intervention graph (or template), we obtain factorisations for counterfactual outcomes under a specific treatment regime by employing ‘node-splitting,’ such that all nodes following an intervention are relabelled as counterfactual states under the preceding intervention. After node-splitting, a fixed intervention is no longer a random variable. Thus, under fixed treatment regimes, the counterfactual states that follow an intervention are independent of the states that occur before node-splitting if there are no backdoor paths into the random partition of the node that has been split.

If all backdoor paths are closed into the random partitions of the nodes on which interventions occur, we can graphically verify that the treatment is independent of the counterfactual outcome for that intervention node. Where there are multiple interventions, we ensure sequential exchangeability at the following node – which we likewise split and relabel – by closing all backdoor paths between the random portion of the following treatment node. We have sequential independence if, for each intervention node, all backdoor paths are closed (refer to J. M. Robins & Richardson, Reference Robins and Richardson2010; Richardson & Robins, Reference Richardson and Robins2013b; Richardson & Robins, Reference Richardson and Robins2023).

The single world intervention template Table 12 G 3 makes it clear that sequential identification may be obtained. A 1 is d-separated from Y by conditioning on L 0; A 2 is d-separated from Y by conditioning on L 2.

Notice that we cannot estimate the combined effect of a treatment strategy over A 1 and A 2 by employing regression, multi-level regression, statistical structural equation models or propensity score matching. However, special estimators may be constructed. (Refer to J. Robins, Reference Robins1986, J. Robins & Hernán, Reference Robins and Hernán2008; Van Der Laan & Rose, Reference Van Der Laan and Rose2011; Díaz et al., Reference Díaz, Hejazi, Rudolph and Der Laan2021; for recent reviews of special estimators refer to Hernán & Robins, Reference Hernán and Robins2024; Chatton et al., Reference Chatton, Le Borgne, Leyrat, Gillaizeau, Rousseau, Barbin and Foucher2020; Van Der Laan & Rose, Reference Van Der Laan and Rose2018; Chatton & Rohrer, Reference Chatton and Rohrer2024.)

Time-varying confounding without treatment-confounder feedback

Consider how we may have time-varying confounding in the absence of treatment-confounder feedback. Suppose we are interested in computing a causal effect estimate for a two-treatment ‘marriage’ intervention on ‘happiness’. We assume that all variables are well defined, that ‘marriage’ can be intervened upon, that we have specified a target population, and that our questions are scientifically interesting. Here, we focus on the challenges in addressing certain causal questions with time-varying confounding without treatment confounder feedback. Table 12 G 1 presents such a structure of time-varying confounding.

Let U AL denote an over-confident personality, an unmeasured variable that is causally associated with decisions to marry early and with income. We do not suppose that U AL affects happiness (Y). Taken in isolation, U AL is not a confounder of the of the $A \rightarrow Y$ path.

Let L t denote income. Suppose that income affects whether one stays married. For example, suppose it takes wealth to divorce. We can sharpen focus on risks of time-varying confounding by assuming that income itself does not affect happiness.

Let U AY denote a common cause of income, L 2, and of happiness at the end of the study, Y. An example of such a confounder might be educational opportunity, which affects both income and happiness. Table 12 G 2 presents the confounding structure for this problem in its simplest form. To declutter, we remove the baseline node, L 0, which we assume to be measured.

Notice that in this example there is no treatment–confounder feedback. We have not imagined that marriage affects wealth or even that wealth affects happiness. Nevertheless, there is confounding. To obtain valid causal inference for the effect of A 2 on Y, we must adjust for L 2 [otherwise: ]. However, L 2is a collider of U AL and U AY. In this setting, adjusting for L 2 opens the path:

We have confounding without treatment-confounder feedback (refer to Hernán & Robins, Reference Hernán and Robins2024).

Table 12 G 4 clarifies that sequential exchangeability can be obtained in the fixed treatment regime. To estimate the effect of A 2 on Y, we must condition on L 2. When estimating the effect of A 1 on Y, all backdoor paths are closed because $L_2(\!\! {{\tilde{\,\,\mathbf a_1}}}\!)$ is a collider and . Note that because a single world intervention template does not represent the joint distributions of more than one treatment, treatment sequence or time-varying treatment, to evaluate the conditional independences we must specify the interventions of interest for A 1 and A 2. That is, we would need to evaluate at least two single world intervention graphs that correspond to the treatment level we wish to contrast. Note further that because there is time-varying confounding, we cannot use standard estimators such as multi-level regressions or structural equation models. Estimation must employ special stepwise methods (refer, for example, to Díaz et al., Reference Díaz, Hejazi, Rudolph and Der Laan2021; Williams and Díaz, Reference Williams and Díaz2021).

Confounding under dynamic treatment strategies (modified treatment policies)

Suppose we are interested in the population average effect of divorce on happiness if divorce is only permitted for those with incomes above 1.5 times the average income. This is a dynamic treatment strategy because the treatment assignment depends on an individual's income, which might vary over time.

Define the treatment policy $g^\phi ( {\cdot} )$ as remaining married for at least two additional years beyond baseline:

$$ A_t^{+}({\mathbf{g}}^\phi) = {\mathbf{g}}^\phi (A_{t}) = \left\{ \hskip -1pc \matrix{&{a_{1} = 1} \cr &{a_{2} = 1}} $$

This regime is identified. The setting is identical to Table 12 G 2 although no unmeasured variables and no arrow from A 1 to L 2.

However, every causal contrast requires a comparison of at least two interventions.

Define the treatment policy $g^\lambda ( {\cdot} )$ as divorce only if one's personal wealth is at least 50% greater than average and one would have divorced in the absence of intervention; otherwise, enforce marriage:

$$ A_t^{+}({\mathbf{g}}^{\lambda}) = {\mathbf{g}}^{\lambda}(A_{t}) = \left\{ \matrix{{a_{1} = 0 &\unicode{x2003} \text{if income} \unicode{x003E} 1.5 \times \mu_{\text{income}} \;\&\; A_1 = 0 }\cr{a_{2} = 0 &\unicode{x2003} \text{if income} \unicode{x003E} 1.5 \times \mu_{\text{income}} \;\&\; A_2 = 0}\cr \hskip -10pc {a_{t} = 1 & \unicode{x2003}\text{otherwise}} $$

Notice that for the treatment policy $\mathbf{g}^\lambda(\cdot)$, treatment is computed as a function of income at both the natural value of A t and of wealth L t. Again, to declutter our graphs, we leave L at baseline off the graph, noting that adjustment at baseline does not change the confounding structure.

Template Table 12 G 5 presents this confounding structure. To convey the dependence of the fixed node on covariate history under treatment, we use Richardson and Robins’ (Reference Richardson and Robins2013a) conventions and draw a dashed line () to indicate paths from the variables on which the time-varying treatment regime depends on the deterministic portion of the intervention node. This strategy clarifies that setting the treatment level requires information about prior variables, including the ’natural value of the treatment’ in the absence of any intervention (Young et al., Reference Young, Hernán and Robins2014).

The reason for noting these dependencies on a Single World Intervention Graph is that such dependencies impose stronger identification assumptions. At every time t, A t must be conditionally independent not only of the potential outcome at the end of the study but also of any future variable that might lead to a non-causal association between future treatments and the potential outcome at the end of the study. We clarify this additional requirement for strong sequential exchangeability in the next section.

Identification of dynamic time-varying treatment strategies using an extension of Robin's dynamic G-formula

Supplementary materials S3 describes Richardson and Robins extension of J. Robins (Reference Robins1986) dynamic g-formula. Essentially, the algorithm can be stated as follows:

  • Step 1 – where $\mathbf{g}(\cdot)$ is a modified treatment policy, identify all the variables that influence the outcome $Y(\mathbf{g})$, excluding those that are current or past treatment variables or covariates.

  • Step 2 – for each treatment at time t, check if the treatment is independent of the variables identified in Step 1, after accounting for past covariates and treatments, in each single world intervention graph where the treatment values are fixed. This step amounts to removing the dotted green arrows from the dynamic single world intervention graph in Table 12 G 5, and doing so gives us Table 12 G 4. For each time point, we recover a set of future counterfactual variables that includes the potential outcome for the treatment regime under consideration, $Y(\mathbf{g})$, and other variables that the treatment might affect, including the natural value of future treatments. All backdoor paths must be closed to each member of this set of counterfactual variables. We call the more stringent assumptions required for identification in time-varying treatments (or equivalently longitudinal modified treatment policies; Díaz et al., Reference Díaz, Williams, Hoffman and Schenck2023a) strong sequential exchangeability, where:

    1. (1) ${\mathbb{Z}}_t(\mathbf{a}^*)$ denotes the subset of vertices in $\mathcal{G}(\mathbf{a}^*)$ corresponding to ${\mathbb{Z}}_t({\mathbf{g}})$;

    2. (2) $A_t({\mathbf{a}}^*) = {a}^*_t$ denotes the specific value of the treatment variable at time t under the intervention ${\mathbf{a}}^*$;

    3. (3) $\bar{\mathbb{L}}_t({\mathbf{a}}^*)$${\mathbf{a}}^*$ denotes the set of covariates up to time t under the intervention $\mathbf{a}^*$; and

    4. (4) $\bar{\mathbb{A}}_{t-1}(\mathbf{a}^*)$ denotes the set of past treatment variables up to time t − 1 under the intervention ${\mathbf{a}}^*$.

Applying Richardson and Robins’ (Reference Richardson and Robins2013a) dynamic extended g-formula, we obtain the following sets of future variables for which each current treatment must be independent:

$$ \eqalign{{\mathbb{Z}}({\mathbf{g}}) &= \{A_1, L_1({\mathbf{g}}), A_1({\mathbf{g}}), A_2({\mathbf{g})}, Y({\mathbf{g}})\} \cr {\mathbb{Z}}_1({\mathbf{g}}) &= \{A_1({\mathbf{g}}), L_1({\mathbf{g}}), Y({\mathbf{g}})\}\cr {\mathbb{Z}}_2({\mathbf{g}}) &= \{Y({\mathbf{g}})\}} $$

Having determined which variables must remain conditionally independent of each treatment in a sequence of dynamic treatments to be compared, we then consider whether strong sequential exchangeability holds. We do this by inspecting template Table 12 G 4 (recall this is Table 12 G 5 without the dashed green arrows). On inspection of G 4 (the dynamic SWIG without dashed green arrows), we discover that this dynamic treatment strategy is not identified because we have the following open backdoor path:

We also have:

Strong sequential exchangeability fails for A 1. We might consider lowering our sights and estimating a fixed or time-varying treatment strategy that can be identified.

Note that certain time-varying treatment strategies impose weaker assumptions than time-fixed strategies. For example, with a continuous intervention, we might consider intervening only if the observed treatment does not reach a specific threshold, such as:

$$ {\mathbf{g}}^{\phi} (A_i) =\left\{\! \matrix{{\mu_A & \unicode{x2003}\text{if } A_i \unicode{x003C} \mu_A}\cr{\!\!\!\!A_i & \unicode{x2003}\text{otherwise}}} $$

This is a weaker intervention than setting everyone whose natural value of treatment is above this threshold to precisely the threshold's value:

$$ {\mathbf{g}}^{\lambda} (A_i) = \left\{\! \matrix{{\mu_A & \unicode{x2003}\text{if } A_i \neq \mu_A}\cr{\!\!\!\!A_i & \unicode{x2003}\text{otherwise}}} \\ $$

Whereas $\mathbf{g}^{\lambda}$ sets everyone in the population to the same treatment level, $\mathbf{g}^{\phi}$sets only those below a certain threshold to a fixed level but does not estimate treatment effects for those above (Hoffman et al., Reference Hoffman, Salazar-Barreto, Rudolph and Díaz2023). We can also write stochastic treatment functions (Díaz et al., Reference Díaz, Hejazi, Rudolph and Der Laan2021; Muñoz & Van Der Laan, Reference Muñoz and Van Der Laan2012; T. VanderWeele & Vansteelandt, Reference VanderWeele and Vansteelandt2014; Young et al., Reference Young, Hernán and Robins2014); see supplementary materials S4.

Of course, the details of every problem must be developed in relation to the scientific context and the practical questions that address gaps in present science. However, causal inference teaches us that the questions we ask – seemingly coherent and tractable questions such as whether marriage makes people happy – demand considerable scrutiny to become interpretable. When such questions are made interpretable, causal inference often reveals that answers may elude us, regardless of the quality and abundance of our data, or even if we randomise interventions. Modest treatment functions, however, might be more credible and useful for many scientific and practical questions. Such functions often cannot be estimated using the models routinely taught in the human sciences, such as multi-level modelling and statistical structural equation modelling.

Conclusions

Philosophical interests in causality are ancient. Democritus once declared, ‘I would rather discover one cause than gain the kingdom of Persia’ (Freeman, Reference Freeman1948). Hume provided a general account of causality by referencing counterfactuals: ‘where, if the first object had not been, the second never would have existed’ (Hume, Reference Hume1902). However, it was not until Jerzy Neyman's master's thesis that a quantitative analysis of causality was formalised (Splawa-Neyman, Reference Splawa-Neyman1990). Remarkably, Neyman's work went largely unnoticed until the 1970s, when Harvard statistician Donald Rubin formalised what became known as the ‘Rubin Causal Model’ (also the Rubin–Neyman Causal Model) (Holland, Reference Holland1986; Rubin, Reference Rubin1976).

In 1986, Harvard statistician James Robins extended the potential outcomes framework to time-varying treatments, laying the foundation for powerful new longitudinal data science methods (J. Robins, Reference Robins1986). Judea Pearl introduced directed acyclic graphs, making identification problems transparent and accessible to non-specialists (Pearl, Reference Pearl1995). Robins and Richardson extended Pearl's graphical models to evaluate counterfactual causal contrasts on graphs, building on Robins’ earlier work. Concurrently, the causal revolution in economics opened new, fertile frontiers in causal data sciences. By the early 2000s, targeted learning frameworks were being developed (Van Der Laan & Rose, Reference Van Der Laan and Rose2011), along with causal mediation analysis methods (Díaz et al., Reference Díaz, Williams and Rudolph2023b; Pearl, Reference Pearl2009; J. M. Robins & Greenland, Reference Robins and Greenland1992; Rudolph et al., Reference Rudolph, Williams and Diaz2024; T. J. VanderWeele, Reference VanderWeele2015; T. VanderWeele & Vansteelandt, Reference VanderWeele and Vansteelandt2014; Vansteelandt et al., Reference Vansteelandt, Bekaert and Lange2012) and techniques for analysing time-varying treatments (Muñoz & Van Der Laan, Reference Muñoz and Van Der Laan2012; Richardson & Robins, Reference Richardson and Robins2013a, Reference Richardson and Robins2023; J. Robins, Reference Robins1986; J. M. Robins et al., Reference Robins, Greenland and Hu1999; J. Robins & Hernán, Reference Robins and Hernán2008; Shpitser et al., Reference Shpitser, Richardson and Robins2022; Young et al., Reference Young, Hernán and Robins2014).

Readers should note that the causal inference literature contains vigorous debates at the horizons of discovery. However, there is a shared consensus about the foundations of causal inference and a common conceptual and mathematical vocabulary within which to express disagreements and accumulate progress – a hallmark of a productive science. Old debates resolve and new debates arise, the hallmark of a vibrant science.

Despite the progress and momentum of the causal revolution in certain human sciences, many areas have yet to participate and benefit. The demands for researchers to acquire new skills, coupled with the intensive requirement for data collection, have significant implications for research design, funding and the accepted pace of scientific publishing. To foster essential changes in causal inference education and practice, the human sciences need to shift from a predominantly output-focused, correlation-reporting culture to a slow, careful, creative culture that promotes retraining and funds time-series data collection. Such investments are worthwhile. Much as Darwin's theory transformed the biological sciences from speculative taxonomy, causal inference is slowly but steadily transforming the human sciences from butterfly collections of correlations to causal inferential sciences capable of addressing the causal questions that animate our curiosities.

Supplementary material

To view supplementary material for this article, please visit https://doi.org/10.1017/ehs.2024.32

Acknowledgements

I am grateful to Dr Inkuk Kim for checking previous versions of this manuscript and offering feedback, to two anonymous reviewers and the editors, Charles Efferson and Ruth Mace, for their constructive feedback. Any remaining errors are my own.

Competing interest

The author declares no conflicts of interest

Financial support

This work is supported by a grant from the Templeton Religion Trust (TRT0418) and RSNZ Marsden 3721245, 20-UOA-123; RSNZ 19-UOO-090. I also received support from the Max Planck Institute for the Science of Human History. The Funders had no role in preparing the manuscript or deciding to publish it.

Research transparency and reproducibility

No data were used in this paper.

References

Angrist, J. D., & Pischke, J.-S. (2009). Mostly harmless econometrics: An empiricist's companion. Princeton University Press.Google Scholar
Bareinboim, E., & Pearl, J. (2013). A general algorithm for deciding transportability of experimental results. Journal of Causal Inference, 1(1), 107134.Google Scholar
Barrett, M. (2021). Ggdag: Analyze and create elegant directed acyclic graphs. https://CRAN.R-project.org/package=ggdagGoogle Scholar
Bulbulia, J. A. (2024a). Methods in causal inference. Part 1: Causal diagrams and confounding. Evolutionary Human Sciences, 6. https://doi.org/10.1017/ehs.2024.35Google Scholar
Bulbulia, J. A. (2024b). Methods in causal inference. Part 3: Measurement error and external validity threats. Evolutionary Human Sciences, 6. https://doi.org/10.1017/ehs.2024.33Google Scholar
Bulbulia, J. A. (2024c). Methods in causal inference. Part 4: Confounding in experiments. Evolutionary Human Sciences, 6. https://doi.org/10.1017/ehs.2024.34Google Scholar
Bulbulia, J. A., Afzali, M. U., Yogeeswaran, K., & Sibley, C. G. (2023). Long-term causal effects of far-right terrorism in New Zealand. PNAS Nexus, 2(8), pgad242.Google Scholar
Chatton, A., & Rohrer, J. M. (2024). The causal cookbook: Recipes for propensity scores, g-computation, and doubly robust standardization. Advances in Methods and Practices in Psychological Science, 7(1), 25152459241236149.Google Scholar
Chatton, A., Le Borgne, F., Leyrat, C., Gillaizeau, F., Rousseau, C., Barbin, L., …, Foucher, Y. (2020). G-computation, propensity score-based methods, and targeted maximum likelihood estimator for causal inference with different covariates sets: a comparative simulation study. Scientific Reports, 10(1), 9219. https://doi.org/10.1038/s41598-020-65917-xGoogle Scholar
Dahabreh, I. J., & Hernán, M. A. (2019). Extending inferences from a randomized trial to a target population. European Journal of Epidemiology, 34(8), 719722. https://doi.org/10.1007/s10654-019-00533-2Google Scholar
Dahabreh, I. J., Robins, J. M., Haneuse, S. J., & Hernán, M. A. (2019). Generalizing causal inferences from randomized trials: Counterfactual and graphical identification. arXiv Preprint arXiv:1906.10792.Google Scholar
Darwin, C. (1859). On the origin of species. Facsimile of the first edition.Google Scholar
Darwin, C. (1887). The life and letters of charles darwin, volume i (Darwin, F., Ed.). D. Appleton. https://charles-darwin.classic-literature.co.uk/the-life-and-letters-of-charles-darwin-volume-i/Google Scholar
De Coulanges, F. (1903). La cité antique: Étude sur le culte, le droit, les institutions de la grèce et de rome. Hachette.Google Scholar
Deffner, D., Rohrer, J. M., & McElreath, R. (2022). A causal framework for cross-cultural generalizability. Advances in Methods and Practices in Psychological Science, 5(3), 25152459221106366. https://doi.org/10.1177/25152459221106366Google Scholar
Díaz, I., Hejazi, N. S., Rudolph, K. E., & Der Laan, M. J. van (2021). Nonparametric efficient causal mediation with intermediate confounders. Biometrika, 108(3), 627641.Google Scholar
Díaz, I., Williams, N., Hoffman, K. L., & Schenck, E. J. (2023a). Nonparametric causal effects based on longitudinal modified treatment policies. Journal of the American Statistical Association, 118(542), 846857. https://doi.org/10.1080/01621459.2021.1955691Google Scholar
Díaz, I., Williams, N., & Rudolph, K. E. (2023b). Efficient and flexible mediation analysis with time-varying mediators, treatments, and confounders. Journal of Causal Inference, 11(1), 20220077. https://doi.org/doi:10.1515/jci-2022-0077Google Scholar
Freeman, K. (1948). Ancilla to the pre-socratic philosophers (reprint edition). Harvard University Press.Google Scholar
Greenland, S. (2009). Commentary: Interactions in epidemiology: Relevance, identification, and estimation. Epidemiology, 20(1), 1417.Google Scholar
Hernán, M. A. (2017). Invited commentary: Selection bias without colliders. American Journal of Epidemiology, 185(11), 10481050. https://doi.org/10.1093/aje/kwx077Google Scholar
Hernán, M. A., & Robins, J. M. (2017). Per-protocol analyses of pragmatic trials. New England Journal of Medicine, 377(14), 13911398.Google Scholar
Hernán, M. A., & Robins, J. M. (2024). Causal inference: What if? Taylor & Francis. https://www.hsph.harvard.edu/miguel-hernan/causal-inference-book/Google Scholar
Hernán, M. A., Hernández-Díaz, S., & Robins, J. M. (2004). A structural approach to selection bias. Epidemiology, 15(5), 615625. https://www.jstor.org/stable/20485961Google Scholar
Hernán, M. A., Alonso, A., Logan, R., Grodstein, F., Michels, K. B., Willett, W. C., …, Robins, J. M. (2008). Observational studies analyzed like randomized experiments: An application to postmenopausal hormone therapy and coronary heart disease. Epidemiology, 19(6), 766. https://doi.org/10.1097/EDE.0b013e3181875e61Google Scholar
Hoffman, K. L., Schenck, E. J., Satlin, M. J., Whalen, W., Pan, D., Williams, N., & Díaz, I. (2022). Comparison of a target trial emulation framework vs cox regression to estimate the association of corticosteroids with COVID-19 mortality. JAMA Network Open, 5(10), e2234425. https://doi.org/10.1001/jamanetworkopen.2022.34425Google Scholar
Hoffman, K. L., Salazar-Barreto, D., Rudolph, K. E., & Díaz, I. (2023). Introducing longitudinal modified treatment policies: A unified framework for studying complex exposures. https://doi.org/10.48550/arXiv.2304.09460Google Scholar
Holland, P. W. (1986). Statistics and causal inference. Journal of the American Statistical Association, 81(396), 945960.Google Scholar
Hume, D. (1902). Enquiries concerning human understanding: And concerning the principles of morals. Clarendon Press.Google Scholar
Kessler, R. C., Andrews, G., Colpe, L. J., Hiripi, E., Mroczek, D. K., Normand, S.-L. T., …, Zaslavsky, A. M. (2002). Short screening scales to monitor population prevalences and trends in non-specific psychological distress. Psychological Medicine, 32(6), 959976. https://doi.org/10.1017/S0033291702006074Google Scholar
Laan, M. van der, Coyle, J., Hejazi, N., Malenica, I., Phillips, R., & Hubbard, A. (2023). Targeted learning in R: Causal data science with the tlverse software ecosystem. Collins Foundation Press. https://tlverse.org/tlverse-handbook/index.htmlGoogle Scholar
Lash, T. L., Rothman, K. J., VanderWeele, T. J., & Haneuse, S. (2020). Modern epidemiology. Wolters Kluwer. https://books.google.co.nz/books?id=SiTSnQEACAAJGoogle Scholar
McElreath, R. (2020). Statistical rethinking: A Bayesian course with examples in R and Stan. CRC press.Google Scholar
Montgomery, J. M., Nyhan, B., & Torres, M. (2018). How conditioning on posttreatment variables can ruin your experiment and what to do about It. American Journal of Political Science, 62(3), 760775. https://doi.org/10.1111/ajps.12357Google Scholar
Morgan, S. L., & Winship, C. (2014). Counterfactuals and causal inference: Methods and principles for social research (2nd ed.). Cambridge University Press. https://doi.org/10.1017/CBO9781107587991Google Scholar
Muñoz, I. D., & Van Der Laan, M. (2012). Population intervention causal effects based on stochastic interventions. Biometrics, 68(2), 541549.Google Scholar
Neal, B. (2020). Introduction to causal inference from a machine learning perspective. Course lecture notes (draft). https://www.bradyneal.com/Introduction_to_Causal_Inference-Dec17_2020-Neal.pdfGoogle Scholar
Ogburn, E. L., & Shpitser, I. (2021). Causal modelling: The two cultures. Observational Studies, 7(1), 179183. https://doi.org/10.1353/obs.2021.0006Google Scholar
Pearl, J. (1995). Causal diagrams for empirical research. Biometrika, 82(4), 669688.Google Scholar
Pearl, J. (2009). Causality. Cambridge University Press.Google Scholar
Pearl, J., & Bareinboim, E. (2022). External validity: From do-calculus to transportability across populations. In Probabilistic and causal inference: The works of judea pearl (pp. 451–482).Google Scholar
Richardson, T. S., & Robins, J. M. (2013a). Single world intervention graphs: A primer. https://core.ac.uk/display/102673558Google Scholar
Richardson, T. S., & Robins, J. M. (2013b). Single world intervention graphs: A primer. Second UAI Workshop on Causal Structure Learning, Bellevue, Washington. https://citeseerx.ist.psu.edu/document?repid=rep1&type=pdf&doi=07bbcb458109d2663acc0d098e8913892389a2a7Google Scholar
Richardson, T. S., & Robins, J. M. (2023). Potential outcome and decision theoretic foundations for statistical causality. Journal of Causal Inference, 11(1), 20220012.Google Scholar
Robins, J. (1986). A new approach to causal inference in mortality studies with a sustained exposure period – Application to control of the healthy worker survivor effect. Mathematical Modelling, 7(9-12), 13931512.Google Scholar
Robins, J. M., & Greenland, S. (1992). Identifiability and exchangeability for direct and indirect effects. Epidemiology, 3(2), 143155.Google Scholar
Robins, J., & Hernán, M. (2008). Estimation of the causal effects of time-varying exposures. In Handbook of Modern Statistical Methods. Chapman & Hall/CRC, 553599.Google Scholar
Robins, J. M., & Richardson, T. S. (2010). Alternative graphical causal models and the identification of direct effects. Causality and Psychopathology: Finding the Determinants of Disorders and Their Cures, 84, 103158.Google Scholar
Robins, J. M., Greenland, S., & Hu, F.-C. (1999). Estimation of the causal effect of a time-varying exposure on the marginal mean of a repeated binary outcome. Journal of the American Statistical Association, 94(447), 687700. https://doi.org/10.1080/01621459.1999.10474168Google Scholar
Rubin, D. B. (1976). Inference and missing data. Biometrika, 63(3), 581592. https://doi.org/10.1093/biomet/63.3.581Google Scholar
Rudolph, K. E., Williams, N. T., & Diaz, I. (2024). Practical causal mediation analysis: Extending nonparametric estimators to accommodate multiple mediators and multiple intermediate confounders. Biostatistics, kxae012. https://doi.org/10.1093/biostatistics/kxae012Google Scholar
Shi, B., Choirat, C., Coull, B. A., VanderWeele, T. J., & Valeri, L. (2021). CMAverse: A suite of functions for reproducible causal mediation analyses. Epidemiology, 32(5), e20e22.Google Scholar
Shpitser, I., Richardson, T. S., & Robins, J. M. (2022). Multivariate counterfactual systems and causal graphical models. In Probabilistic and causal inference: The works of Judea Pearl (pp. 813–852). Association for Computing Machinery (ACM)Google Scholar
Splawa-Neyman, J. (1990). On the application of probability theory to agricultural experiments. Essay on principles. Section 9. (1923). Statistical Science, 5(4), 465472. (Original work published 1923)Google Scholar
Steen, J., Loeys, T., Moerkerke, B., & Vansteelandt, S. (2017). Medflex: An R package for flexible mediation analysis using natural effect models. Journal of Statistical Software, 76, 146.Google Scholar
Stuart, E. A., Ackerman, B., & Westreich, D. (2018). Generalizability of randomized trial results to target populations: Design and analysis possibilities. Research on Social Work Practice, 28(5), 532537.Google Scholar
Suzuki, E., Mitsuhashi, T., Tsuda, T., & Yamamoto, E. (2013). A counterfactual approach to bias and effect modification in terms of response types. BMC Medical Research Methodology, 13(1), 117.Google Scholar
Suzuki, E., Shinozaki, T., & Yamamoto, E. (2020). Causal diagrams: Pitfalls and tips. Journal of Epidemiology, 30(4), 153162. https://doi.org/10.2188/jea.JE20190192Google Scholar
Tibshirani, J., Athey, S., Sverdrup, E., & Wager, S. (2024). Grf: Generalized random forests. https://github.com/grf-labs/grfGoogle Scholar
Valeri, L., Lin, X., & VanderWeele, T. J. (2014). Mediation analysis when a continuous mediator is measured with error and the outcome follows a generalized linear model. Statistics in Medicine, 33(28), 48754890.Google Scholar
Van Der Laan, M. J., & Rose, S. (2011). Targeted learning: Causal inference for observational and experimental data. Springer. https://link.springer.com/10.1007/978-1-4419-9782-1Google Scholar
Van Der Laan, M. J., & Rose, S. (2018). Targeted learning in data science: Causal inference for complex longitudinal studies. Springer International. http://link.springer.com/10.1007/978-3-319-65304-4Google Scholar
VanderWeele, T. J. (2012). Confounding and effect modification: Distribution and measure. Epidemiologic Methods, 1(1), 5582. https://doi.org/10.1515/2161-962X.1004Google Scholar
VanderWeele, T. J. (2015). Explanation in causal inference: Methods for mediation and interaction. Oxford University Press.Google Scholar
VanderWeele, T. J., & Hernán, M. A. (2012). Results on differential and dependent measurement error of the exposure and the outcome using signed directed acyclic graphs. American Journal of Epidemiology, 175(12), 13031310. https://doi.org/10.1093/aje/kwr458Google Scholar
VanderWeele, T. J., & Hernán, M. A. (2013). Causal inference under multiple versions of treatment. Journal of Causal Inference, 1(1), 120.Google Scholar
VanderWeele, T. J., & Knol, M. J. (2014). A tutorial on interaction. Epidemiologic Methods, 3(1), 3372.Google Scholar
VanderWeele, T. J., & Robins, J. M. (2007). Four types of effect modification: A classification based on directed acyclic graphs. Epidemiology (Cambridge, Mass.), 18(5), 561568. https://doi.org/10.1097/EDE.0b013e318127181bGoogle Scholar
VanderWeele, T. J., & Tchetgen Tchetgen, E. J. (2017). Mediation analysis with time varying exposures and mediators. Journal of the Royal Statistical Society Series B: Statistical Methodology, 79(3), 917938.Google Scholar
VanderWeele, T., & Vansteelandt, S. (2014). Mediation analysis with multiple mediators. Epidemiologic Methods, 2(1), 95115.Google Scholar
VanderWeele, T. J., Vansteelandt, S., & Robins, J. M. (2014). Effect decomposition in the presence of an exposure-induced mediator-outcome confounder. Epidemiology, 25(2), 300306.Google Scholar
Vansteelandt, S., & Dukes, O. (2022). Assumption-lean inference for generalised linear model parameters. Journal of the Royal Statistical Society Series B: Statistical Methodology, 84(3), 657685.Google Scholar
Vansteelandt, S., Bekaert, M., & Lange, T. (2012). Imputation strategies for the estimation of natural direct and indirect effects. Epidemiologic Methods, 1(1), 131158.Google Scholar
Vo, T.-T., Williams, N., Liu, R., Rudolph, K. E., & Díaz, I. (2024). Recanting twins: Addressing intermediate confounding in mediation analysis. arXiv Preprint arXiv:2401.04450.Google Scholar
Westreich, D., & Cole, S. R. (2010). Invited commentary: positivity in practice. American Journal of Epidemiology, 171(6). https://doi.org/10.1093/aje/kwp436Google Scholar
Westreich, D., & Greenland, S. (2013). The table 2 fallacy: Presenting and interpreting confounder and modifier coefficients. American Journal of Epidemiology, 177(4), 292298.Google Scholar
Westreich, D., Edwards, J. K., Lesko, C. R., Stuart, E., & Cole, S. R. (2017). Transportability of trial results using inverse odds of sampling weights. American Journal of Epidemiology, 186(8), 10101014.Google Scholar
Wheatley, P. (1971). The pivot of the four quarters: A preliminary enquiry into the origins and character of the ancient chinese city (p. 602). Aldine.Google Scholar
Williams, N. T., & Díaz, I. (2021). lmtp: Non-parametric causal effects of feasible interventions based on modified treatment policies. https://doi.org/10.5281/zenodo.3874931Google Scholar
Wilson, D. S. (2008). Evolution and religion: The transformation of the obvious. In Bulbulia, J., Sosis, R., Harris, E., Genet, R., Genet, C., & Wyman, K. (Eds.), The evolution of religion: Studies, theories, & critiques (pp. 2329). Collins Foundation Press.Google Scholar
Young, J. G., Hernán, M. A., & Robins, J. M. (2014). Identification, estimation and approximation of risk under interventions that depend on the natural value of treatment using observational data. Epidemiologic Methods, 3(1), 119.Google Scholar
Figure 0

Table 1. Terminology

Figure 1

Table 2. Elements of Causal Graphs

Figure 2

Table 3. Five elementary causal structures in a causal directed acyclic graph

Figure 3

Table 4. Graphical conventions we use for representing effect modification

Figure 4

Table 5. Effect Modification

Figure 5

Table 6. Single World Interventions Graphs $( {{\cal G}_{3-4}} ) $ present separate causal diagrams for each treatment to be contrasted. A Single World Intervention Template $( {{\cal G}_2} ) $ is a ‘graph value function’ that produces the individual counterfactual graphs (Richardson & Robins, 2013a). On the other hand, causal directed acyclic graphs, such as $( {{\cal G}_1} ) $, require positing interventional distributions. The formalism underpinning these interventional distributions is mathematically equivalent to formalism underpinning the potential outcomes framework, assuming the errors of the underlying structural causal models that define the nodes on which interventions occur are independent (Richardson & Robins, 2013a). Single World Intervention Graphs (SWIGs), however, permit the comparison of distinct interventions in our causal diagram without requiring that the non-parametric structural equations that correspond to nodes on a causal graph have independent error structures. This is useful when attempting to identify the causal effects of sequential treatments, refer to supplementary materials S2

Figure 6

Table 7.

Figure 7

Table 8. In causal mediation, the quantities that we require to obtain natural direct and indirect effects, namely $\mathbb{E}[Y(1,M(0))]$, cannot be experimentally observed because we cannot treat someone and observe the level of their mediator if they were not treated

Figure 8

Table 9. Assumptions of Causal Mediation Analysis

Figure 9

Table 10. Four fixed treatment regimens in time-series data where exposure varies

Figure 10

Table 11. Six causal contrasts in time-series data where exposure varies

Figure 11

Table 12. Single World Intervention Graph for Sequential Treatments

Supplementary material: File

Bulbulia supplementary material

Bulbulia supplementary material
Download Bulbulia supplementary material(File)
File 212 KB