2,427
Views
1
CrossRef citations to date
0
Altmetric
Articles

Confounding adjustment methods for multi-level treatment comparisons under lack of positivity and unknown model specification

, , , & ORCID Icon
Pages 2570-2592 | Received 20 Aug 2020, Accepted 29 Mar 2021, Published online: 07 Apr 2021

Abstract

Imbalances in covariates between treatment groups are frequent in observational studies and can lead to biased comparisons. Various adjustment methods can be employed to correct these biases in the context of multi-level treatments (> 2). Analytical challenges, such as positivity violations and incorrect model specification due to unknown functional relationships between covariates and treatment or outcome, may affect their ability to yield unbiased results. Such challenges were expected in a comparison of fire-suppression interventions for preventing fire growth. We identified the overlap weights, augmented overlap weights, bias-corrected matching and targeted maximum likelihood as methods with the best potential to address those challenges. A simple variance estimator for the overlap weight estimators that can naturally be combined with machine learning is proposed. In a simulation study, we investigated the performance of these methods as well as those of simpler alternatives. Adjustment methods that included an outcome modeling component performed better than those that focused on the treatment mechanism in our simulations. Additionally, machine learning implementation was observed to efficiently compensate for the unknown model specification for the former methods, but not the latter. Based on these results, we compared the effectiveness of fire-suppression interventions using the augmented overlap weight estimator.

1. Introduction

Many empirical studies seek to evaluate the effect of a treatment or intervention. This evaluation can be attempted using randomized or observational experiments. In the former, pre-randomization characteristics are expected to be similar across treatment groups. As such, outcome differences can be causally attributed to the treatment. However, randomized trials are often difficult to realize because they may be unethical, impractical, or untimely [Citation12]. Thus, relying on observational experiments is often necessary. Unfortunately, imbalances in covariates between treatment groups are frequent in observational studies and can lead to biased treatment comparisons. This major challenge is known as confounding, wherein differences in outcomes between treatment groups are at least partly attributable to systematic differences in baseline covariates.

The motivation for the current study was the comparison of the effect of multiple fire-suppression interventions on fire growth in Alberta, Canada. In a previous attempt at tackling this question using observational data, a reversed association was observed wherein more aggressive interventions were associated with greater fire growth [Citation37]. We believed two main reasons could explain this unexpected association. First, important confounding by indication appeared plausible, since more aggressive interventions would be preferred for wildfires that are expected to grow quickly. Adjusting for further potential confounders may help reducing bias, but how best to adjust for these confounders was unclear. We also hypothesized that the positivity assumption may be violated in these data. Positivity refers to the fact that all units should have a positive probability of receiving each possible treatment and is required for many adjustment methods to yield unbiased estimators. In our context, it seemed possible, for example, that some fires may be too large when initially discovered to consider less aggressive interventions, thus violating the positivity assumption.

Various methods can be used to correct confounding bias. The performance of these methods have been compared in multiple simulation studies in the case of a binary treatment [Citation4–6,Citation15,Citation20,Citation21,Citation26,Citation28] and for performing pairwise comparisons between multi-level treatments [Citation8,Citation22,Citation25,Citation35,Citation42,Citation43]. In summary, methods that rely on binary propensity score for multi-level treatments comparisons, focusing on comparing treatment levels two at a time, can yield biased and nontransitive comparisons and are not recommended [Citation25,Citation42]. Stratification was susceptible to produce results with important residual bias [Citation4–6,Citation26]. Standardization is generally the best performing method in ideal situations, in terms of bias and mean-squared error (MSE) [Citation4,Citation15,Citation20,Citation21,Citation26]. However, when the model used for standardization is incorrectly specified, the estimator is biased [Citation8,Citation20,Citation21,Citation26]. Weighting methods were susceptible to be biased and have a large variance if the model used for constructing the weights was misspecified [Citation15,Citation20,Citation42] or under positivity violation [Citation22,Citation25,Citation28,Citation35]. Trimming, which removes from the analysis observations in areas where there is a lack of positivity, has been observed to improve the performance of multiple methods in presence of positivity violations [Citation15,Citation22]. The augmented inverse probability weighting estimator has been observed to have mixed performances. This estimator performed competitively to standardization in ideal situations and retained similar performances under model misspecifications [Citation26], but it performed poorly under positivity violations [Citation15,Citation21]. The targeted maximum likelihood estimator (TMLE) was observed to have good performances under ideal situations and under model misspecifications, in addition to being more robust to positivity violations [Citation20]. Similar results were observed concerning bias-corrected matching [Citation15,Citation20,Citation21]. The use of machine learning algorithms was observed to be beneficial for preventing the bias attributable to model misspecifications for TMLE, BCM and standardization, but not necessarily for weighting [Citation20].

From a theoretical perspective, methods that combine treatment and outcome modeling are generally expected to be n-consistent as long as each component is consistently estimated at a rate n1/4 (see for example [Citation17] and references therein). This allows using a greater array of machine learning methods without sacrificing desirable asymptotic properties. This is unlike methods that rely on modeling only the treatment or the outcome which typically require n-consistency of their model to be n-consistent.

Weighting estimators that are robust to positivity violations have also been developed in recent years [Citation21–23,Citation43]. In particular, [Citation23] consider a general class of weighting estimators for binary treatments, that notably include as specific cases standard weighting and weighting with trimmed weights. They theoretically derived, under certain conditions, the lowest variance estimator in this class, which they called the overlap weights. An extension of these overlap weights to multi-level treatments has also been developed [Citation22]. The overlap weights were observed to outperform the usual weighting estimator, with or without trimming, as well as matching estimators, both in terms of bias and MSE [Citation22,Citation28]. An outcome regression augmented overlap weighting estimator can be constructed in order to improve the robustness to model misspecification [Citation22,Citation28].

Overall, a few methods thus stand out according to their performance in terms of bias, MSE, and their robustness to model misspecifications or positivity violations: bias-corrected matching, TMLE and overlap weights. While bias-corrected matching and TMLE have been compared in simulation studies with binary treatments, they have never been compared together in scenarios with multi-level treatment or against the overlap weights. Which of these methods yield the best estimator of pairwise average treatment effects, including in scenarios with model misspecifications or positivity violations, is thus unknown. To inform our analysis of the wildfire data, we investigated the performance of TMLE, bias-corrected matching and overlap weights for performing pairwise comparisons between multi-level treatments, notably under lack of positivity and unknown model specification.

The remainder of this article is structured as follows. In the next section, the notation used throughout the article is first introduced. Then, the methods being compared are formally presented, as well as the causal assumptions on which they rely. We also propose some methodological developments concerning the overlap weight estimators, including a convenient asymptotic variance estimator. In Section 3, we present the simulation study we have conducted to compare bias-corrected matching, TMLE, overlap weights, and augmented overlap weights. Standardization, inverse probability weighting and matching were also included as benchmark comparators. We further examine the potential of machine learning implementations of these methods to deal with the unknown model specification. Both a fully synthetic Monte Carlo simulation, and a plasmode simulation are presented. Plasmode simulations combine real and synthetic data to produce more realistic simulation scenarios [Citation10]. Section 4 presents the analysis of the wildfire data, which is informed by the results of our simulation study. We conclude with a discussion in Section 5.

2. Methodology

Let Y be the outcome of interest, TT={1,2,,k} be the multi-level treatment, and X={X1,,Xp}X be a set of covariates, where T and X are the support of variables T and X, respectively. We assume that a sample of independent observations {Xi,Ti,Yi}, with i=1,,n, is drawn from a given population. We also denote by Yi(t) the potential outcome of observation i under treatment level t, that is, the outcome that would have been observed for observation i had the treatment taken level t, potentially contrary to the fact. Using this notation, the population average pairwise treatment effect comparing treatment levels t and t is τt,t=E[Y(t)Y(t)].

The fundamental problem of causal inference is that it is only possible to observe one potential outcome for each observation, the one corresponding to the factual treatment level. Since the other potential outcomes are missing, the causal effect τt,t cannot be estimated from the observed data without making some assumptions. Henceforth, we make the following usual assumptions (see, for example, [Citation12, Chapter 3]): (1) Y(t)T|X for all tT, (2) 0<P(T=t|X)<1 for all tT and XX and (3) Ti=tYi=Yi(t). The first assumption is exchangeability and indicates that treatment allocation is independent of the potential outcomes conditional on measured covariates. Informally, this means there are no factors that would simultaneously explain T and Y beyond what may be explained by X. Assumption (2) is the positivity assumption that entails that each observation had a strictly positive probability of being assigned to any treatment level. The third is the consistency assumption and indicates that the observed outcome for observation i is equal to its potential outcome under its factual treatment level.

2.1. Standardization

A first possible estimator of τt,t consists of computing a standardized mean difference of the outcome between treatment groups t and t (see, for example, [Citation12, Chapter 13]): τ^t,tstan=1ni=1n[E^(Y|T=t,Xi)E^(Y|T=t,Xi)], where E^(Y|T=t,Xi) and E^(Y|T=t,Xi) are the estimated expectations of the outcome under treatment level t and t, respectively, and covariates Xi. These estimated expectations can be obtained using a parametric regression model of Y conditional on T and X, such as a linear regression if Y is continuous or a logistic regression if Y is binary. More sophisticated methods, including machine-learning algorithms, could also be employed. Although a variance estimator for τ^t,tstan that accounts for the estimation of treatment model parameters exists [Citation44], it can be computationally intensive since it involves a double sum over the observations' index. Alternatively, inferences can be produced using the nonparametric bootstrap, which also allows accounting for parameters estimation.

2.2. Inverse probability weighting

Inverse probability weighting estimators of τt,t are inspired by the method proposed by Hovitz and Thompson [Citation14]. Robins [Citation32] introduced inverse probability weighting estimators of the parameters of longitudinal marginal structural models, which include as a special case the following estimator of τt,t: τ^t,tIPW=1ni=1n[I(Ti=t)YiI(Ti=t)YiP^(T=ti|Xi)], where I() denotes the usual indicator function and P^(T=ti|Xi) is the estimated probability that the treatment takes level ti conditional on Xi. P^(T=ti|Xi) can, for example, be obtained from a multinomial logistic regression. A conservative asymptotic variance estimator for τ^t,tIPW is conveniently obtained using a sandwich estimator, treating P^(T=ti|Xi) as known [Citation26,Citation33].

2.3. Matching and bias-corrected matching

Scotina et al. [Citation35] present a matching estimator as well as a bias-corrected matching estimator based on the work of Abadie and Imbens [Citation1,Citation2]. This matching estimator first imputes the missing potential outcomes of each observation by the observed outcomes of observations from the other treatment group. More precisely, the potential outcome of observation i under treatment t, Yi(t), for tti, is imputed with the average of the observed outcome of the m observations from treatment group t whose covariates' values are closest to Xi according to some distance metric. The potential outcome Yi(ti) is deemed observed (Yi(ti)=Yi). This matching procedure is performed with replacement, allowing each observation to be used multiple times for imputing potential outcomes of different observations. Comparisons between treatment groups are then performed directly on the (observed/imputed) potential outcomes.

More formally, let A be some definite positive matrix and ||x||A=(xAx)1/2 define the distance metric. Let Mit denote the set of the m observations ‘closest’ to unit i in treatment group tti, that is, such that jMit||XjXi|| is minimal and jI(jMit)=m. Then Y^i(t)={Yi,ift=ti,1mjMitYj,iftti, and the matching estimator is τ^t,tmatch=1ni=1n[Y^i(t)Y^i(t)]. Similar matching estimators for multi-level treatments had previously been introduced by [Citation9] and [Citation42].

The bias-corrected matching estimator proposed by [Citation35] uses a regression method for imputing the missing potential outcomes: Y^ibc(t)={Yi,ift=ti,1mjMit[Yj+E^(Y|T=t,Xi)E^(Y|T=t,Xj)],iftti. The bias-corrected matching estimator is: τ^t,tBCM=1ni=1n[Y^ibc(t)Y^ibc(t)]. The consistency of the matching estimator τ^t,tmatch and the bias-corrected matching estimator τ^t,tBCM relies on further conditions, in addition to the causal assumptions presented earlier (for more details, see [Citation35]). Asymptotic variance estimators of τ^t,tmatch and of τ^t,tBCM are provided in [Citation35]. These variance estimators were derived assuming the matching is performed on fixed X and bias-correction employs a non-parametric series estimator for the outcome expectation. However, the variance estimators were observed to behave well in other circumstances in simulation studies, notably when matching on the logit of the estimated generalized propensity score vector P^(T=t|X) [Citation35].

2.4. Targeted maximum likelihood

Targeted maximum likelihood estimation is a general methodology introduced by van der Laan and Rubin [Citation40] for constructing doubly-robust semiparametric efficient estimators. Double-robustness entails that the estimator is consistent for τt,t if either the outcome model or the treatment model component of the algorithm described below is consistent, but not necessarily both. Moreover, the TMLE has minimal variance among the class of semiparametric estimators when both models are correctly specified. Luque-Fernandez et al. [Citation27] provide an introductory tutorial on TMLE.

An algorithm for obtaining a TMLE, if Y is binary, is as follows [Citation36].

  • For j{t,t}

    1. Denote Q0(j,Xi)=E^(Y|T=j,Xi)

    2. Define weights wi(j)=I(Ti=j)P^(T=j|Xi)

    3. Run a logistic regression of Y with an intercept, logit[Q0(j,X)] as an offset term and weights wi(j). Denote the estimated intercept term as ϵ^

    4. Let Q1(j,Xi)=expit{logit[Q0(j,Xi)]+ϵ^}

The TMLE estimator is then τ^t,tTMLE=1ni=1n[Q1(t,Xi)Q1(t,Xi)]. If Y is a continuous variable, it is recommended to rescale Y and Q0(j,X) such that values lie between 0 and 1 before performing step 3 [Citation20,Citation39, Chapter 7]. Let a and b be the known limits of Y, then the rescaled outcome and expectations are Y=Yaba and Q0,(j,X)=Q0(j,X)aba. The causal effect estimate τ^t,tTMLE is computed after back-transforming Q1(t,Xi) and Q1(t,Xi) into the original scale.

An asymptotic estimator for the variance of τ^t,tTMLE is based on the sample variance of the efficient influence function (see, for example, [Citation39, Chapter 5]). This variance estimator is consistent when both of these models are estimated with consistent estimators and asymptotically conservative when only the treatment model estimator is consistent [Citation39].

2.5. Overlap weights

When using the overlap weights, each observation is attributed a weight wi=h(Xi)/P^(T=ti|Xi), where h(Xi)=[l=1k1/P^(T=l|Xi)]1. The overlap weight estimator is then τ^t,tOW=i=1nI(Ti=t)Yiwii=1nI(Ti=t)wii=1nI(Ti=t)Yiwii=1nI(Ti=t)wi. This specific choice of h(Xi) yields an estimator with minimal variance, assuming homoscedastic residual variances of the potential outcomes [Citation22].

It is important to note that τ^t,tOW is not generally a consistent estimator for the population average pairwise treatment effect τt,t. In fact, the estimand of τ^t,tOW is the so-called average pairwise treatment effect in the overlap population: τt,t=XX{E[Y(t)|X]E[Y(t)|X]}f(X)h(X)μ(dX)XXf(X)h(X)μ(dX), where f(X) is the marginal density of X with respect to an appropriate measure μ. This corresponds to a weighted average treatment effect, where h(X) is the weight function. The estimand of the overlap weights focuses on the treatment effect in the part of the population where the positivity assumption holds. Indeed, parts of the populations for which at least one of the treatment is impossible have a weight of 0 in the estimand.

A variance estimator for τ^t,tOW that accounts for the estimation of the weights is provided by [Citation22]. Because this variance estimator requires computing derivatives of the weights according to parameters of the treatment model, it combines poorly with machine learning methods, which may lack a simple parametric expression. We thus propose an alternative variance estimator similar to the variance estimator commonly employed for inverse probability weighting estimators. This simpler conservative variance estimator for τ^t,tOW is obtained by considering the weights as known. It corresponds to the empirical variance of the nonparametric influence function of τt,t (given by [Citation28]), scaled by 1/n: Var^(τ^t,tOW)=1n2i=1n[I(Ti=t)(Yiτ^t)wi1ni=1nh(Xi)I(Ti=t)(Yiτ^t)wi1ni=1nh(Xi)]2, where τ^t=i=1nI(Ti=t)Yiwii=1nI(Ti=t)wi is the treatment t weighted mean. This variance estimator is motivated by the fact the behavior of a semiparametric estimator is asymptotically the same as the one from its influence function [Citation40, Chapter 5]. An important advantage of our proposed variance estimator is that effect estimates and inferences based on this variance estimator correspond to those produced by standard software when running a weighted generalized estimating equation regression with the treatment as the sole predictor and employing the robust variance estimator (for example, using proc genmod in SAS with a repeated statement, or using geeglm in the geepack package in R [Citation11]).

As indicated by [Citation22] it is possible to construct an outcome-regression augmented version of the overlap weight estimator: τ^t,tAOW=τ^t,tOW+i=1nh(Xi)[E^(Y|T=t,Xi)E^(Y|T=t,Xi)]i=1nh(Xi)i=1nI(Ti=t)E^(Y|T=t,Xi)wii=1nI(Ti=t)wi+i=1nI(Ti=t)E^(Y|T=t,Xi)wii=1nI(Ti=t)wi. This estimator is semiparametric efficient for estimating τt,t when both the treatment model P^(T=t|X) and the outcome model E^(Y|T,Xi) are correctly specified [Citation22].

Following [Citation28], we can show that τ^t,tAOW has a property analogue to double-robustness.

Theorem 2.1

The estimator τ^t,tAOW is consistent for τt,t when the treatment model is correctly specified, whether the outcome model is correctly specified or not. When the outcome model is correctly specified, but the treatment model is misspecified, τ^t,tAOW is consistent for τ~t,t=XX{E[Y(t)|X]E[Y(t)|X]}f(X)h~(X)μ(dX)XXf(X)h~(X)μ(dX), where h~(X) is the estimand of [l=1k1/P^(T=l|Xi)]1 under the misspecified P^(T=l|Xi).

The proof of Theorem 2.1 is provided in Web Appendix A.

Based on the efficient influence function derived by [Citation28], we propose an asymptotic variance estimator for τ^t,tAOW obtained by computing the empirical variance of the nonparametric efficient influence function, scaled by a factor 1/n Var^(τ^t,tAOW)=1n2i=1n{h(Xi)1ni=1nh(Xi)[E^(Y|T=t,Xi)E^(Y|T=t,Xi)+I(Ti=t){YiE^(Y|T=t,Xi)}P^(T=t|Xi)I(Ti=t){YiE^(Y|T=t,Xi)}P^(T=t|Xi)]+I(Ti=t){YiE^(Y|T=t,Xi)}P^(T=t|Xi)I(Ti=t){YiE^(Y|T=t,Xi)}P^(T=t|Xi)]τ^t,tAOW}2. This variance estimator is consistent if both the treatment and outcome models are estimated with consistent estimators. Like the variance estimator for τ^t,tOW we proposed, this variance estimator can be conveniently computed when machine learning methods are employed for fitting the treatment or outcome model. Indeed, the variance formulas are relatively simple functions of observed variables and predicted values, thus not directly depending on how the model was fitted.

3. A simulation study

We now describe the simulation study we have conducted to compare the empirical performance of the methods presented in the previous section.

3.1. Monte Carlo simulation

3.1.1. Simulation design and scenarios

For our data-generating process, we considered three covariates X1, X2, X3 arbitrarily generated as follow: X1N(0,1), X2N(2X1,1) and X3Bernoulli(0.4). The treatment variable T was simulated according to a multinomial logistic regression with three levels (1, 2 and 3), where the first was the reference category. The probability of membership in each group was given by: π1=[1+exp(β21X1+β22X2+β23X3+β24X1X3+β25X12)+exp(β31X1+β32X2+β33X3+β34X1X3+β35X12)]1,π2=exp(β21X1+β22X2+β23X3+β24X1X3+β25X12)π1,π3=exp(β31X1+β32X2+β33X3+β34X1X3+β35X12)π1. The outcome Y was generated from a normal distribution YN(μ3,1), where μ3=γ0+γ1X1+γ2X2+γ3X3+γ4X2X3+γ5X22+λ1I(T=2)+λ2I(T=3).

We defined four simulation scenarios where the association between the covariates and the treatment was either weak (T) or strong (T+), and the association between the covariates and the outcome was either weak (Y) or strong (Y+). Parameters λ1 and λ2 represent the true treatment effects of level 2 and 3 as compared to level 1 (τ21 and τ31), respectively, and were set to λ1=1.0 and λ2=1.5. These Monte Carlo scenarios were devised such that the estimand of the overlap weights is the same as the one from the other methods. Such a choice facilitates the comparison between estimators. However, in the plasmode simulation presented in Section 3.2, the estimand of the overlap weight estimators does not correspond with that of the other methods.

The specific choice of the values for β and γ was made to attain the following desiderata: (1) in Scenario TY, the confounding bias for both λ parameters is between 10% and 30%, (2) the bias for Scenarios TY+ and T+Y is between 40% and 60%, (3) the bias for Scenario T+Y+ is above 100%, (4) there is a good overlap of the treatment probabilities' distribution between treatment groups in Scenarios T, and (5) there is a poor overlap in Scenarios T+ (near positivity violation).

In Scenarios T, the β values were β21=0.2, β22=0.2, β23=0.2, β24=0.1, β25=0.1, β31=0.2, β32=0.1, β33=0.2, β34=0.1, β35=0.1. In Scenarios T+, they were β21=0.8, β22=0.8, β23=0.8, β24=0.2, β25=0.2, β31=0.5, β32=0.5, β33=0.8, β34=0.2, β35=0.2. In Scenarios Y, the γ values were γ0=0, γ1=0.2, γ2=0.2, γ3=0.2, γ4=0.1 and γ5=0.1. In Scenarios Y+, they were γ0=0, γ1=0.5, γ2=0.5, γ3=0.5, γ4=0.2 and γ5=0.2.

For each scenario, we simulated 1000 independent data sets each of sample size 500, 1000 and 2000 under the above conditions. We then estimated τ21 and τ31 utilizing the standardization (stan), inverse probability weighting (IPW), matching (match), bias-corrected matching (BCM), targeted maximum likelihood (TMLE), overlap weights (OW) and augmented overlap weights (AOW) estimators presented previously. For match and BCM, the matching was performed on the logit of the estimated generalized propensity score (the vector logit[P^(T=t|X)]). First, the estimators were implemented using correctly specified parametric models, that is, a linear regression of Y on T, X1, X2, X3, X2X3 and X22, and a multinomial logistic regression of T on X1, X2, X3, X1X3 and X12. This allows investigating the performance of the estimators under ideal circumstances. Next, the estimators were implemented using parametric models that include only main terms, excluding interaction or quadratic terms. This corresponds to a common implementation since the correct model is unknown in practice. Finally, the estimators were implemented using machine learning approaches. For the outcome, we used a Super Learner. The Super Learner is an ensemble method that yields an estimate corresponding to a weighted average of the prediction of multiple procedures, where the weights are determined using cross-validation [Citation38]. The prediction procedures we used were a main terms only linear regression, a main, interactions and quadratic terms linear regression, and a generalized additive model with cubic splines. This was performed using the SuperLearner package in R [Citation31]. Because this package does not accommodate multinomial dependent variables, we used a polychotomous regression and multiple classification with the R package polspline [Citation18,Citation19] to model the exposure. This approach uses linear splines and their tensor products to produce predictions. The machine learning methods we chose were selected such that their rate of convergence was n.

We computed the following measures to assess and compare the performance of the estimators: bias, standard deviation (Std), ratio of the mean estimated standard error to the standard deviation of the estimates, square-root of the mean-squared error (RMSE) and the proportion of the replications in which 95% confidence intervals included the true value of the treatment effect (Coverage CI). For stan, inferences were produced using a nonparametric bootstrap estimate of the variance with 200 samples. For IPW, TMLE, OW, and AOW, the influence function variance estimator was used. For match and BCM, we used the variance estimator proposed by [Citation35]. To assess bias levels, we calculated an unadjusted estimate of the treatment effect.

3.1.2. Simulation results

Tables  and compare the bias, standard deviation, RMSE and coverage of confidence intervals across methods for n = 1000. Similar results were obtained for n = 500 and n = 2000. A detailed presentation of these results is available in the Supplemental Online Material. Figure  depicts the ratio of the mean estimated standard error to the standard deviation of the estimates of τ31. Similar results were observed for τ21 and are available in the Supplemental Online Material. A figure depicting the overlap in the treatment probabilities distribution according to treatment group and scenarios is also available in the Supplemental Online Material.

Figure 1. Ratio of the mean estimated standard error to the standard deviation of the estimates of τ31 for the correct parametric implementation according to sample size (dark gray = 500, gray = 1000, light gray = 2000), the strength of the association between covariates and treatment (columns; weak=T or strong=T+) and between the covariates and outcome (weak=Y or strong=Y+), and implementation (rows; Cor = Correct parametric, Inc = Incorrect parametric, ML = machine learning).

Figure 1. Ratio of the mean estimated standard error to the standard deviation of the estimates of τ31 for the correct parametric implementation according to sample size (dark gray = 500, gray = 1000, light gray = 2000), the strength of the association between covariates and treatment (columns; weak=T− or strong=T+) and between the covariates and outcome (weak=Y− or strong=Y+), and implementation (rows; Cor = Correct parametric, Inc = Incorrect parametric, ML = machine learning).

Table 1. Estimate of the treatment effect in the Scenario with a weak association between covariates and treatment, a weak association between covariates and outcome (TY), and a sample size of 1000.

Table 2. Estimate of the treatment effect in the Scenario with a strong association between covariates and treatment, a weak association between covariates and outcome (T+Y), and a sample size of 1000.

Table 3. Estimate of the treatment effect in the Scenario with a weak association between covariates and treatment, a strong association between covariates and outcome (TY+), and a sample size of 1000.

Table 4. Estimate of the treatment effect in the Scenario with a strong association between covariates and treatment, a strong association between covariates and outcome (T+Y+), and a sample size of 1000.

Under a correct parametric specification, all methods managed to almost completely eliminate the bias in Scenarios TY and TY+. In Scenarios with positivity violations (T+Y and T+Y+), IPW and match yielded results with substantial residual bias, while the other methods still achieved close to unbiased estimation. In all scenarios, stan produced the estimates with lowest standard deviation and RMSE, followed closely by AOW, and by TMLE in Scenarios TY and TY+. In Scenarios T+Y and T+Y+, the standard deviation and RMSE of IPW and match were drastically larger than those of the other methods. The coverage of confidence intervals of all methods was close to 95% or slightly conservative, except for IPW and match in Scenarios T+Y and T+Y+ where confidence intervals included the true effect much less often than expected.

When using an incorrect parametric specification, important bias reduction was still achieved, but residual confounding bias remained for all adjustment methods. In Scenarios TY and TY+, BCM and match were the methods that reduced the bias the most and were the only methods that yielded confidence intervals with appropriate coverage. In Scenarios T+Y and T+Y+, AOW and BCM were the methods that produced the most important bias reduction and the coverage of their confidence intervals were the closest, yet inferior, to 95%.

When a machine learning implementation was used, stan, BCM, TMLE and AOW produced results very similar to those obtained under the correct parametric implementations, with regards to bias, standard deviation, RMSE and coverage of confidence intervals. However, IPW, match and OW produced biased results with increased standard deviations and RMSE. The coverage of their confidence intervals was also generally below the expected level, except for match that retained appropriate coverage in Scenarios TY and TY+.

In all scenarios, we noticed that the bias of match and BCM tended to decrease with increased sample size, even under an incorrect parametric implementation.

Figure  indicates that the variance estimators we proposed for OW and AOW performed well. As expected, the variance estimator for OW tends to overestimate the true variance. The estimated standard error of AOW was very similar to its Monte Carlo standard deviation. Figure  also reveals that the usual sandwich estimator of the variance of IPW, which is expected to be conservative, sometimes drastically underestimated the true variance in Scenarios TY+ and T+Y+. The variance estimators of stan, match and BCM were conservative in Scenarios T+Y and T+Y+ when using a correct parametric implementation. In all other cases, the estimated and empirical variances were similar.

3.2. Plasmode simulation

3.2.1. Plasmode datasets

We used the data on forest fires in Alberta, Canada, as a basis for our plasmode simulation. These data are produced and published online by the Government of Alberta [Citation41]. The purpose of the analysis was to compare various interventions for fighting wildfires in Alberta on their probability of preventing the fire to grow after its initial assessment. We considered only fires caused by lightning from 1996 to 2014. Observations for which size at ‘being held’ was smaller than the size at initial attack were also removed. ‘Being held’ is defined as in [Citation37] as a state when no further increase in size is expected.

Potential confounders considered in the plasmode simulation are the ecological region in which the fire occurred (Clear Hills Upland, Mid-Boreal Uplands, Other), the number of fires active at the time of initial assessment of each fire, fuel type (Boreal Spruce, Boreal Mixedwood – Green, other), month of the year the fire was first assessed (‘May or June’, July, ‘August, September or October’), how the fire was discovered (air patrol, lookout, unplanned), response time in hours between the moment the fire was reported to first assessment by forestry personnel, and the natural logarithm (ln) of the size of the fire at the time of the initial attack. Response time was truncated at the 95th percentile of the distribution to limit the influence of extreme observations, which caused convergence problems of models in some replications of the simulation. Additional potential confounders were available but were not considered in the plasmode simulation. As will be explained shortly, the plasmode simulation was run on a reduced sample of the total data available. As such, including all potential confounders in the simulations caused convergence issues.

Some levels of categorical variables having few instances were dropped, and observations in those categories were removed. The final database contained 8591 observations, the 7 covariates presented above, and the treatment variable indicating the method of intervention used by the firefighters to suppress the wildfire and the outcome variable. The categories for the treatment were heli-attack crew with helicopter but no rappel capability (HAC1H; 53.8%), heli-attack crew with helicopter and rappel capability (HAC1R; 15.8%), fire-attack crew with or without a helicopter and no rappel capability (HAC1F; 6.3%), Air tanker (15.3%) and Ground-based action (8.8%). The binary response variable was 1 or 0 depending on whether the fire did or did not increase in size between initial attack and ‘being held’ (n =1982 and 6645, respectively).

For generating plasmode data, we first fitted a random forest regression of the outcome variable conditional on treatment and all seven covariates using the package randomForest in R with the default settings [Citation24]. Similarly, a random forest regression of the treatment according to the covariates was fitted. These models were then used to generate simulated treatment and outcome variables. This approach naturally allows for treatment-covariate interactions in generating the outcome and thus allows the estimand of the overlap weight estimators to differ from that of the other estimators. The true causal effects were estimated using a Monte Carlo simulation. Briefly, for each observation in the complete population, we generated counterfactual outcomes that would have been observed under each possible treatment according to the true outcome model. The difference in risk of fire progression were then computed for each treatment level, as compared to HAC1H, which was used as the reference level. For the overlap weights, risk differences weighted according to h(X) as defined in Section 2.5 were computed. The true effects comparing Air tanker, Ground-based action, HAC1F and HAC1R to HAC1H in the entire population were, respectively, 0.099, −0.011, 0.012 and −0.003. In the overlap population, the true risk differences were 0.101, −0.024, 0.003 and −0.002. To simplify the presentation, both sets of parameters will be denoted as τ21, τ31, τ41 and τ51 in the results below.

A plasmode simulated dataset was obtained as follows. First, we drew with replacement a random sample of size 2000 from the original data. Then, for each observation, a new simulated treatment and a new simulated outcome were generated, with P(T=t|X) and P(Y=1|T,X) determined by the fitted random forest models. Based on this process, we simulated 1000 independent data sets and compared adjustments methods using the same performance metrics as in the Monte Carlo simulation (see Section 3.1.1). For each method, we considered either a parametric models implementation that included only main terms, or a machine learning implementation. For the treatment, these implementations were the same as described in Section 3.1.1. The implementations for the outcome were also similar to those described in Section 3.1.1, but replacing linear regressions by logistic regressions. Since the outcome and treatment data were generated nonparametrically, the correct models are unknown.

3.2.2. Simulation results

A figure displaying the distribution of the treatment probabilities according to treatment groups is available in the Supplemental Online Material. While a good level of overlap between treatment groups was present, multiple observations had treatment probabilities close to 0, thus suggesting possible practical positivity violations.

The results presented in Table  show that a large amount of bias is present when no adjustment is performed, especially for τ21. When using a parametric implementation, most methods achieved a bias reduction for estimating all parameters, except stan and match that increased the bias for one parameter. Additionally, a substantial bias remained for estimating some parameters for stan, match, and AOW. The RMSE of stan, TMLE, and OW were smaller than those of the crude estimates for all parameters, whereas IPW, match, BCM, and AOW had an increased RMSE for at least one parameter. The lowest RMSE for all parameters was produced by stan. Most methods yielded 95% confidence intervals with coverage rate close to the expected level for τ21, τ41 and τ51, but not for τ31. Only IPW and OW had close to adequate coverage for τ31. match and BCM yielded inadequate confidence intervals for all parameters, because of an underestimation of the true variance (see Figure 3 in the Supplemental Online Material). When using a machine learning implementation, the performance of most adjustment methods was overall marginally improved. Indeed, the bias and RMSE were generally smaller, and coverage closer to 95%. For IPW, however, the results were somewhat worse.

Table 5. Estimate of treatment effect in plasmode simulation using 2000 observations. True effects are τ21=0.099, τ31=0.011, τ41=0.012 and τ51=0.003 for all methods except OW and AOW and τ21=0.101, τ31=0.024, τ41=0.003 and τ51=0.002 for the these two methods. HAC1H is the reference category . τ21, τ31, τ41 and τ51 refer to treatment effect associated to Air tanker, Ground-based action, HAC1F and HAC1R, respectively.

Figure 3 in the Supplemental Online Material represents the ratio of the mean estimated standard error to the standard deviation of the estimates. The variances estimated using our proposed estimators for OW and AOW were similar to the Monte Carlo variances. In fact, all variance estimators except those of match and BCM performed well in the plasmode simulation.

4. Application

4.1. Context

Wildfires have great economic, environmental and societal impacts. A key tool of fire management is fire suppression by initial attack teams who seek to limit the growth, and hence the final size, of the fire [Citation7]. These interventions by initial attack differ in terms of the size and training of the fire-fighting crews, and the mechanical resources provided to them. However, few studies have attempted estimating the causal effect of initial attack on wildfire growth.

Recently, [Citation37] compared initial attack interventions using public data on wildfire in Alberta, Canada. More aggressive interventions were associated with greater fire growth, opposite to the expected direction of the causal effect. These results suggest that important confounding by indication may be present in fire management agency records.

4.2. Data

The data we considered are similar to those used for performing the plasmode simulation and those employed by [Citation37]. More precisely, we considered all fires on provincial land or in other public lands caused by lighting between 2003 and 2014 recorded in the historical wildfire database of Alberta's Agriculture and Forestry ministry [Citation41].

The final data consisted of 5439 observations. The fire-attack interventions being compared are the same as those in the plasmode simulation: heli-attack crew with helicopter but no rappel capability (HAC1H), heli-attack crew with helicopter and rappel capability (HAC1R), fire-attack crew with or without a helicopter and no rappel capability (HAC1F), Air tanker, and Ground-based action. The outcome of interest was whether the fire grew in size between initial assessment and ‘being held '.

Based on experts' knowledge and the scientific literature, the following 11 potential confounders were considered: (1) Initial Spread Index, (2) Fire Weather Index, (5) ecological region in which the fire occurred (Clear Hills Upland, Mid-Boreal Uplands, Other), (6) fuel type at initial assessment (Boreal Spruce, Boreal Mixedwood – Green, Other), (7) period of day (AM or PM), (8) month of the year the fire was first assessed (‘May or June’, July, ‘August, September or October’), (9) response time in hours between the moment the fire was reported to first assessment by fire fighters, (10) the number of fires active at the time of initial assessment of each fire, and (11) the natural logarithm (ln) of the size of the fire at the initial attack. The Initial Spread Index and the Fire Weather Index are used to predict the rate of progression of fires and fire danger, respectively, by Alberta's Wildfire Management Branch. These indexes are notably based on daily weather variables. To account for the fact that initial attack decisions would be based not only on current and recent weather, but also on future forecast, we used the values of these variables for the day of fire assessment, the two days prior to assessment and the two days following assessment (for a total of five variables for each index). More information on these variables is available in [Citation37] and references therein.

4.3. Analysis

Not all interventions would be reasonable choices to suppress some fires given their characteristics at the moment of assessment. As such, it seems more relevant to target the average effect of interventions only for those fires for which multiple options are reasonable. Based on this and on the results from our simulation study, we decided to use the augmented overlap-weights estimator with a machine learning implementation. The treatment probabilities were estimated using a polychotomous regression and the outcome expectations were estimated with the Super Learner using a main term generalized linear model, a generalized additive model with spline terms, and random forests as prediction procedures. Note that we no longer limit the machine learning methods we consider to those having a convergence rate of n, since the augmented overlap weight estimator we use combines outcome and treatment modeling, thus allowing the use of more flexible machine learning methods. The generalized linear model with interaction and quadratic terms was not considered as a prediction procedure because of the small sample size in some levels of categorical variables. All potential confounders were included as independent variables in both models. Year was considered as a categorical variable (12 levels). Inferences were produced using the variance estimator we introduced in Section 2.5. P-values were adjusted for multiple comparisons with Holm'smethod [Citation13].

4.4. Results

Characteristics of the considered fires according to the initial intervention that was chosen for suppressing the fire are presented in Web Appendix C. Important imbalances between treatment groups were observed for calendar year, how the fire was discovered, ecological region, fuel type, response time, number of active fires and initial size of the fire, and some imbalance was observed for the fire weather index on the day of initial assessment. To the best of our knowledge, no method has been developed to assess post-adjustment balance for the augmented overlap weight estimator.

Adjusted associations between initial interventions for suppressing fires and the probability of fire growth are reported in Table . Air tanker, which is the most aggressive intervention, is associated with the largest probability of fire growth. All interventions are associated with probabilities of fire growth similar to or greater than that of ground-based action, which we consider to be the least aggressive intervention. The various heli-attack crew interventions are all associated with similar probabilities of fire growth.

Table 6. Association between initial intervention used to suppress the fire and probability of fire growth between initial assessment and being held.

5. Conclusion and discussion

The initial motivation for this work was to compare the effect of five initial attack interventions on wildfire growth. Multiple analytical challenges were expected in attempting to estimate this effect, including important confounding by indication, possible non-positivity and unknown model specifications. We reviewed studies evaluating the empirical performance of adjustment methods and identified those that appeared best suited to address these challenges: the overlap weights, the bias-corrected matching and the targeted maximum likelihood estimation.

Regarding the overlap weight estimator, we demonstrated that when it is augmented with an outcome regression, it benefits from a property analogue to double-robustness in the multi-level treatment case, following the proof of [Citation28] for the binary case. Additionally, we have proposed a simple asymptotic variance estimator for the overlap weights and the augmented overlap weights based on the semiparametric theory. These variance estimators offer multiple benefits. First, they are expressed as the sample variance of a relatively simple quantity that do not require advanced coding or mathematical skills to compute. This is particularly true for the overlap weights, since estimation and inferences correspond to those produced by common generalized estimating equations routines with a robust variance estimator. A further advantage of our proposed variance estimator is that it is agnostic to the model used for computing the point estimates. Hence, it can be computed as easily whether the point estimates are obtained with parametric regression models or machine learning algorithms.

We have also conducted a simulation study comparing the overlap weights, the augmented overlap weights, the bias-corrected matching and the TMLE estimator, as well as standardization, regular inverse probability weighting and matching as benchmark comparators. Most of these methods had never been compared to one another in a multi-level treatment setting. Our proposed variance estimator for the overlap weight estimators performed well, even with small sample sizes. We also observed that methods that combine outcome and treatment modeling (bias-corrected matching, TMLE and augmented overlap weights) performed better than those solely based on the treatment model (inverse probability weighting, matching and overlap weights) in terms of bias reduction, standard deviation of estimates and RMSE. Under positivity violations, inverse probability weighting and matching were biased and had increased variability, while all other methods were relatively unaffected. We also observed that the variance estimator for the matching and bias-corrected matching estimators of [Citation35] performed poorly in the plasmode simulation. We hypothesize that this is because of the binary nature of the outcome.

Unsurprisingly, when an incorrect parametric implementation was used, all adjustment methods were biased. However, the bias for matching and bias-corrected matching tended to get smaller as sample size increased, whereas the bias remained constant for the other methods. A plausible explanation for this phenomenon is that these matching methods are less model dependent, since they involve imputing the missing counterfactual outcomes by the observed outcomes from subjects in the other treatment groups. As sample size increases, the pool of control subjects get larger, thus allowing observations to be matched with others that are more similar to them. We did not observe that methods that combine outcome and exposure modeling performed better than the others under models misspecification. While it is expected that double-robustness theoretically helps protecting against the bias attributable to model misspecifications, this property requires that at least one of the two models involved is correct for unbiased estimation. When both models are incorrect to some extent, as should arguably be expected in practice, double-robust methods do not necessarily produce estimates with less bias than others [Citation16,Citation20]. However, a theoretical advantage of combining outcome and treatment modeling is the possibility to use machine learning methods that have a relatively slow convergence rate without sacrificing the convergence rate of the treatment effect estimator [Citation17].

The results of our simulation study finally indicate that machine learning methods can be helpful for preventing misspecification bias for some, but not all, adjustment methods. Indeed, for all methods that include an outcome modeling component (standardization, bias-corrected matching, TMLE and augmented overlap weight), the results under a machine learning implementation were very similar to those obtained under the ideal case of a correct parametric specification for all considered performance metrics. As such, the routine use of machine learning algorithms for performing confounder adjustment may have benefits to help preventing bias without paying a price in terms of power. On the other hand, methods that focus only on modeling the treatment (inverse probability weighting, matching and overlap weight) had considerable bias when employing a machine learning implementation. A possible explanation for these results is that machine learning algorithms may tend to more often predict treatment probabilities close to 0 or 1 because they are better able to fit the observed data, thus exacerbating practical positivity violations.

Standardization was the method that overall performed best, closely followed by the augmented overlap weight in most scenarios and sometimes also by TMLE. In practice, we warn that the choice of an approach should not only be guided by its performance, but also by the question of interest. For example, if the goal is to inform about a decision that would apply at the whole population level, the average effect in the entire population would be the one that is of most interest, whereas if the question is to help the decision in cases where several treatment options are possible, the methods that estimate an effect in the subgroups where there is overlap would likely be more appropriate.

Based on the context of the wildfire growth problem and the results of the simulation study, we decided to estimate the effect of initial attack interventions using the augmented overlap weight estimator, implemented with machine learning algorithms. The adjusted associations we observed were counter-intuitive: the more aggressive the intervention was, the larger was the probability that fires grew. These associations are very unlikely to represent the true causal effect. Based on our simulation study's results, we do not believe these results can be explained by residual confounding due to measured confounders or to non-positivity. We thus hypothesize that residual confounding due to important non-measured confounders is present, despite the fact we have made important efforts to include multiple potential confounders in our analysis. Having accounted for the factors previously shown to affect the probabilities of fire growth after initial attack in this system [Citation3], we are unable to advance a more precise explanation. Another possible explanation for our counterintuitive results may be the presence of large effect modification, that is, the treatment effect may vary widely according to the initial characteristics of the fire. It may thus be the case that the less aggressive interventions are indeed more effective than more aggressive interventions in most situations, but less effective in some specific situations. Dynamic treatment regime methods may thus be interesting to explore in future studies to garner a better understanding of the effect of fire suppression methods[Citation29,Citation30,Citation34].

Ultimately, the findings from this study originate from simulation studies and are thus limited to the contexts that were investigated. However, the plasmode simulation based on real data helped understanding the performance of these approaches in a realistic setting. Bias elimination relies on the fact that all confounders are measured, which arguably never occurs in practice. As such, at least some amount of residual confounding due to unmeasured confounders is inevitably present in real data analyses. This is one plausible explanation for the counter-intuitive associations we have observed in the wildfire analysis. Comparing the effect of fire-fighting interventions using observational data is fraught with multiple challenges. Our work contribute to facing these challenges by helping ruling out residual confounding due to known confounders and positivity violations as possible explanations to the unexpected associations that are observed. We hope our study will stimulate others to attempt facing the remaining challenges to be able to better inform fire managers regarding the best course of action when wildfires occur.

Supplemental material

Supplementary data

Download PDF (414.6 KB)

Acknowledgments

This research was funded by the Natural Sciences and Engineering Research Council of Canada grant # 2016-06295. S. A. Diop was also supported by a scholarship of Excellence from the Faculté de sciences et de génie of Université Laval. D Talbot is a Fonds de recherche du Québec - Santé Chercheur-Boursier.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

This work was supported by Fonds de Recherche du Québec – Santé [grant number 265385] and Natural Sciences and Engineering Research Council of Canada [grant number 2016-06295].

References

  • A. Abadie and G.W. Imbens, Large sample properties of matching estimators for average treatment effects, Econometrica 74 (2006), pp. 235–267.
  • A. Abadie and G.W. Imbens, Bias-corrected matching estimators for average treatment effects, J. Bus. Econ. Stat. 29 (2011), pp. 1–11.
  • M.C. Arienti, S.G. Cumming and S. Boutin, Empirical models of forest fire initial attack success probabilities: the effects of fuels, anthropogenic linear features, fire weather, and management, Canadian J. Forest Res. 36 (2006), pp. 3155–3166.
  • P.C. Austin, The performance of different propensity-score methods for estimating relative risks, J. Clin. Epidemiol. 61 (2008), pp. 537–545.
  • P.C. Austin, The performance of different propensity-score methods for estimating differences in proportions (risk differences or absolute risk reductions) in observational studies, Stat. Med. 29 (2010), pp. 2137–2148.
  • P.C. Austin, P. Grootendorst and G.M. Anderson, A comparison of the ability of different propensity score models to balance measured variables between treated and untreated subjects: a Monte Carlo study, Stat. Med. 26 (2007), pp. 734–753.
  • S. Cumming, Effective fire suppression in boreal forests, Canadian J. Forest Res. 35 (2005), pp. 772–786.
  • P. Feng, X.H. Zhou, Q.M. Zou, M.Y. Fan and X.S. Li, Generalized propensity score for estimating the average treatment effect of multiple treatments, Stat. Med. 31 (2012), pp. 681–697.
  • M. Frölich, Programme evaluation with multiple treatments, J. Econ. Surv. 18 (2004), pp. 181–224.
  • G. L. Gadbury, Q. Xiang, L. Yang, S. Barnes, G. P. Page, D. B. Allison and G. Gibson, Evaluating statistical methods using plasmode data sets in the age of massive public databases: an illustration using false discovery rates, PLoS Genet. 4 (2008), p. e1000098.
  • U. Halekoh, S. Højsgaard and J. Yan. The r package geepack for generalized estimating equations, J. Stat. Softw. 15 (2006), pp. 1–11.
  • M. Hernán and J. Robins, Causal Inference: What If, Chapman & Hall/CRC, Boca Raton, FL, 2020.
  • S. Holm, A simple sequentially rejective multiple test procedure, Scandinavian J. Stat. 6 (1979), pp. 65–70.
  • D.G. Horvitz and D.J. Thompson, A generalization of sampling without replacement from a finite universe, J. Am. Stat. Assoc. 47 (1952), pp. 663–685.
  • M. Huber, M. Lechner and C. Wunsch, The performance of estimators based on the propensity score, J. Econom. 175 (2013), pp. 1–21.
  • J.D. Kang and J.L. Schafer, Demystifying double robustness: a comparison of alternative strategies for estimating a population mean from incomplete data, Stat. Sci. 22 (2007), pp. 523–539.
  • E.H. Kennedy, Semiparametric theory and empirical processes in causal inference, in Statistical Causal Inferences and Their Applications in Public Health Research, H. He, P. Wu, D.-G. Chen, eds., Springer, Cham, 2016, pp. 141–167.
  • C. Kooperberg, polspline: Polynomial Spline Routines, R package version 1.1.14, 2019. Available at https://CRAN.R-project.org/package=polspline.
  • C. Kooperberg, S. Bose and C.J. Stone, Polychotomous regression, J. Am. Stat. Assoc. 92 (1997), pp. 117–127.
  • N. Kreif, S. Gruber, R. Radice, R. Grieve and J.S. Sekhon, Evaluating treatment effectiveness under model misspecification: a comparison of targeted maximum likelihood estimation with bias-corrected matching, Stat. Methods. Med. Res. 25 (2016), pp. 2315–2336.
  • L. Li and T. Greene, A weighting analogue to pair matching in propensity score analysis, Int. J. Biostat. 9 (2013), pp. 215–234.
  • F. Li and F. Li, Propensity score weighting for causal inference with multiple treatments, Ann. Appl. Stat. 13 (2019), pp. 2389–2415.
  • F. Li, K.L. Morgan and A.M. Zaslavsky, Balancing covariates via propensity score weighting, J. Am. Stat. Assoc. 113 (2018), pp. 390–400.
  • A. Liaw and M. Wiener, Classification and regression by randomforest, R News 2 (2002), pp. 18–22. Available at https://CRAN.R-project.org/doc/Rnews/.
  • M.J. Lopez and R. Gutman, Estimation of causal effects with multiple treatments: a review and new ideas, Stat. Sci. 32 (2017), pp. 432–454.
  • J.K. Lunceford and M. Davidian, Stratification and weighting via the propensity score in estimation of causal treatment effects: a comparative study, Stat. Med. 23 (2004), pp. 2937–2960.
  • M.A. Luque-Fernandez, M. Schomaker, B. Rachet and M.E. Schnitzer, Targeted maximum likelihood estimation for a binary treatment: A tutorial, Stat. Med. 37 (2018), pp. 2530–2546.
  • H. Mao, L. Li and T. Greene, Propensity score weighting analysis and treatment effect discovery, Stat. Methods. Med. Res. 28 (2019), pp. 2439–2454.
  • E.E. Moodie, T.S. Richardson and D.A. Stephens, Demystifying optimal dynamic treatment regimes, Biometrics 63 (2007), pp. 447–455.
  • S.A. Murphy, Optimal dynamic treatment regimes, J. R. Stat. Soc.: Ser. B (Stat. Methodol.) 65 (2003), pp. 331–355.
  • E. Polley, E. LeDell, C. Kennedy and M. van der Laan, SuperLearner: Super Learner Prediction, R package version 2.0-24, 2018. Available at https://CRAN.R-project.org/package=SuperLearner.
  • J.M. Robins, Marginal structural models, in Proceedings of the Section on Bayesian Statistical Science, Vol. 1997, American Statistical Association, Alexandria, VA, 1998, pp. 1–10.
  • J.M. Robins, Marginal structural models versus structural nested models as tools for causal inference, in Statistical Models in Epidemiology, the Environment, and Clinical Trials, M. Halloran and B. Donald, eds., Springer, New York, 2000, pp. 95–133.
  • J.M. Robins, Optimal structural nested models for optimal sequential decisions, Proceedings of the Second Seattle Symposium in Biostatistics, D. Lin, P.J. Heagerty, eds., Springer, New York, 2004, pp. 189–326.
  • A.D. Scotina, F.L. Beaudoin and R. Gutman, Matching estimators for causal effects of multiple treatments, Stat. Methods. Med. Res. 29 (2019), pp. 1051–1066.
  • A. A. Siddique, M. E Schnitzer, A. Bahamyirou, G. Wang, T. H Holtz, G. B Migliori, G. Sotgiu, N. R Gandhi, M. H Vargas, D. Menzies and A. Benedetti, Causal inference with multiple concurrent medications: a comparison of methods and an application in multidrug-resistant tuberculosis, Stat. Methods. Med. Res. 28 (2019), pp. 3534–3549.
  • P-O. Tremblay, T. Duchesne, S. G. Cumming and J. A. Jones, Survival analysis and classification methods for forest fire size, PLoS One 13 (2018), p. e0189860.
  • M.J. Van der Laan, E.C. Polley and A.E. Hubbard, Super learner, Stat. Appl. Genet. Mol. Biol. 6 (2007), Article 25.
  • M.J. Van der Laan and S. Rose, Targeted Learning: Causal Inference for Observational and Experimental Data, Springer, New York, 2011.
  • M.J. Van Der Laan and D. Rubin, Targeted maximum likelihood learning, Int. J. Biostat. 2 (2006), p. 28.
  • Wildfire Management Branch – Alberta Agriculture and Forestry, Historical Wildfire Database, 2015. Available at http://wildfire.alberta.ca/resources/historical-data/historical-wildfire-database.aspx.
  • S. Yang, G.W. Imbens, Z. Cui, D.E. Faries and Z. Kadziola, Propensity score matching and subclassification in observational studies with multi-level treatments, Biometrics 72 (2016), pp. 1055–1065.
  • K. Yoshida, S. Hernández-Díaz, D.H. Solomon, J.W. Jackson, J.J. Gagne, R.J. Glynn and J.M. Franklin, Matching weights to simultaneously compare three treatment groups: comparison to three-way matching, Epidemiology 28 (2017), pp. 387–395.
  • G. Zou, Assessment of risks by predicting counterfactuals, Stat. Med. 28 (2009), pp. 3761–3781.