Skip to content
Publicly Available Published by De Gruyter July 2, 2015

The Bayesian Causal Effect Estimation Algorithm

  • Denis Talbot EMAIL logo , Geneviève Lefebvre and Juli Atherton

Abstract

Estimating causal exposure effects in observational studies ideally requires the analyst to have a vast knowledge of the domain of application. Investigators often bypass difficulties related to the identification and selection of confounders through the use of fully adjusted outcome regression models. However, since such models likely contain more covariates than required, the variance of the regression coefficient for exposure may be unnecessarily large. Instead of using a fully adjusted model, model selection can be attempted. Most classical statistical model selection approaches, such as Bayesian model averaging, do not readily address causal effect estimation. We present a new model averaged approach to causal inference, Bayesian causal effect estimation (BCEE), which is motivated by the graphical framework for causal inference. BCEE aims to unbiasedly estimate the causal effect of a continuous exposure on a continuous outcome while being more efficient than a fully adjusted approach.

1 Introduction

Estimating causal exposure effects in observational studies demands a vast knowledge of the domain of application. For instance, to estimate the causal effect of an exposure on an outcome, the graphical framework to causality usually involves postulating a causal graph to identify an appropriate set of confounding variables [1]. Specifying such a graph can be difficult, especially in subject areas where prior knowledge is scarce or limited.

Investigators often bypass difficulties related to the identification and selection of confounders through the use of fully adjusted outcome regression models. Such models express the outcome variable as a function of the exposure variable and all available potential confounding variables. A fully adjusted outcome regression model is commonly assumed to yield an unbiased estimator of the true effect of the exposure. However, since such models likely contain more covariates than required, the variance of the regression coefficient for exposure may be unnecessarily large. Instead of using a fully adjusted model, model selection can be attempted.

Most classical statistical model selection approaches do not readily address causal effect estimation. One such approach is Bayesian model averaging (BMA) [2, 3]. BMA averages quantities of interest (e.g. a regression coefficient or the value of a future observation) over all possible models under consideration: in the average, each estimate is weighted by the posterior probability attributed to the corresponding model. When the goal is prediction, BMA accounts for the uncertainty associated with model choice and produces confidence intervals that have adequate coverage probabilities [4]. Unfortunately, BMA can perform poorly when used to estimate a causal effect of exposure [5, 6].

Wang et al. [6] suggested two novel approaches that modify BMA to specifically target causal effect estimation: Bayesian adjustment for confounding (BAC) and two-stage Bayesian adjustment for confounding (TBAC). Graph-based simulations presented in Wang et al. [6] show that the causal effect estimators of BAC and TBAC are unbiased in a variety of scenarios, hence supporting their adequacy for causal inference. A theoretical justification for the use of BAC for causal inference purposes is further discussed in Lefebvre, Atherton, and Talbot [7]. However, some simulations comparing BAC and TBAC to fully adjusted models show little difference in the variance of the causal effect estimators of each method [6, 8]. Moreover, the choice of BAC’s hyperparameter ω has been recognized as challenging [9]. The value ω= has been recommended if one seeks an unbiased causal exposure effect estimator [7]. Lefebvre et al. [7] proposed using cross-validation and bootstrap for selecting an ω value that aims to minimize the mean-square-error (MSE) of the BAC’s causal effect of exposure estimator. These results suggest that the optimal ω value not only depends on the data-generating scenario, but also on sample size, thus making it very hard in practice to select an appropriate ω value.

In this paper we propose a new model averaging approach to causal inference: Bayesian causal effect estimation (BCEE). BCEE aims to unbiasedly estimate the causal effect of a continuous exposure on a continuous outcome, while being more efficient than a fully adjusted approach. With a sample of finite size, however, this is an ambitious objective. Hence, through a user-selected hyperparameter, BCEE enables an analyst to consider various degrees of trade-off between bias and variance for the estimator. While BCEE shares some similarities with TBAC, one distinctive feature of our approach is that its motivation lies in the graphical framework for causal inference (e.g. Pearl [1]).

The paper is structured as follows. In Section 2, we present the BCEE algorithm and discuss, in Section 3, a number of aspects of its practical implementation. We compare BCEE to some existing approaches for causal effect estimation in Section 4. In Section 5, we apply BCEE to a real dataset where we estimate the causal effect of mathematical perceived competence on the self-reported average in mathematics for highschool students in the province of Quebec. We conclude in Section 6 with a discussion of our results and provide suggestions for further research.

2 Bayesian causal effect estimation (BCEE)

Before presenting BCEE in Section 2.3, we first describe the modeling framework in Section 2.1 and provide a proposition and corollary concerning directed acyclic graphs (DAGs) in Section 2.2. The description of how the proposition and the corollary are used to develop BCEE is presented in Section 2.4. We conclude, in Section 2.5, with a toy example that sheds light on BCEE’s properties. Note that although we refer to BCEE as a Bayesian algorithm, strictly speaking, it is approximately Bayesian since it requires specifying prior distributions only for a subset of the parameters. To simplify the discussion, we motivate BCEE from a frequentist perspective.

2.1 Modeling framework

We consider estimating the causal effect of a continuous exposure on a continuous outcome. Let X be the random exposure variable, Y be the random outcome variable and U={U1,U2,...,UM} be a set of M available, pre-exposure, potentially confounding random covariates. Let i index the units of observations, i=1,...,n. Our goal is to estimate the causal effect of exposure using a linear regression model for the outcome with normal, independent and identically distributed errors. Assuming the set U is sufficient to identify the average causal effect and the model is correctly specified, a fully adjusted linear regression model can be used to estimate the causal effect. Under such assumptions, parameter β encodes the average causal effect of a unit increase in X on Y in the linear model

(1)EYi|Xi,sUi=δ0+βXi+m=1MδmUim,

where δ0 is the intercept and δm is the regression coefficient associated with covariate Um. A disadvantage to using a fully adjusted outcome model is that the variance of the exposure effect estimator βˆ can be large. Therefore, one might want to include a reduced number of covariates in the outcome model (1), that is, to adjust for a strict subset of U also sufficient to estimate the causal effect of X on Y.

Consider G an assumed causal directed acyclic graph (DAG) compatible with the distribution of the observed covariates in G, {Y,X,U}. Let D={D1,D2,...,DJ}U be the set of parents (direct causes) of X in G. Then using Pearl’s back-door criterion [1], it is straightforward to show that adjusting for the set D is sufficient to avoid confounding. In other words, the parameter β in the linear model

(2)EYi|Xi,Di=δ0+βXi+j=1JδjDij

can also be interpreted as the average causal effect of X on Y. It can also be shown that outcome models adjusting for sets of pre-exposure covariates that at least include the direct causes of exposure are unbiased; BAC may be seen to be exploiting this feature [7]. Adjusting for the set of direct causes of X in the outcome model thus seems appealing since D is generally smaller than the full set U. However, this approach can also yield an estimator of β, βˆ, whose variance is large unless those direct causes of X are also strong predictors of Y (e.g. Lefebvre et al. [7]).

BAC, TBAC and BCEE all rely on the fact that the set of direct causes of X is sufficient for estimating the causal effect and that this set of covariates can be identified from the data. A differentiating feature of BCEE is that it aims to disfavor outcome models that include one or more direct causes of X that are unnecessary to eliminate confounding. This is viewed as desirable since these variables generally increase the variance of βˆ. By doing so, BCEE targets sufficient models

(3)EYi|Xi,Zi=δ0+βXi+k=1KδkZik

for which the variance of βˆ is smaller than the variance of βˆ in model (1) and the variance of βˆ obtained using BAC or TBAC. In Section 2.2 we present a proposition and a corollary that underlie the functioning of BCEE.

2.2 A motivation based on directed acyclic graphs

The results presented in this section are based on Pearl’s back-door criterion and are thus obtained from a graphical perspective to causality using directed acyclic graphs (DAGs). For a brief review of this framework, we refer the reader to the appendix of VanderWeele and Shpitser [10].

Proposition 2.1 presented below gives a sufficient condition to identify a set Z that yields an unbiased estimator βˆ of the causal effect of X in eq. (3). Corollary 2.1 starts with such a sufficient set Z and provides conditions under which a direct cause of X included in Z can be excluded so that the resulting set Z is also sufficient. Remark that this corollary is akin to Proposition 1 from VanderWeele and Shpitser [10]. In the sequel, the concept of d-separation is used to entail notions of conditional independence between variables. Moreover, the distribution-free adjustment defined in Pearl [1] relates to the adjustment in the linear model setting introduced in Section 2.1. For instance, see Chapter 5 from Pearl [1] and section 5.3.2 in particular.

Proposition 2.1

Consider data compatible with a causal DAG G. LetD={D1,D2,...,DJ}be the set of direct causes of X and letZbe a set of covariates which we consider adjusting for. Adjusting forZis sufficient to identify the average causal effect of X on Y if

  1. no descendants of X are inZand

  2. if for eachDjD, either

    1. DjZor

    2. ifDjZthen Y andDjare d-separated by{XZ}.

Proof: see Appendix A.1.

Corollary 2.1

Consider aDjZand letZ=ZDj.

  1. IfDjand Y are d-separated by{XZ}then all back-door pathsXDjYare blocked byZ.

  2. If in addition to 1., Zis sufficient to identify the average causal effect according to Proposition 2.1, thenZis also sufficient to identify the average causal effect of X on Y.

Proof: see Appendix A.2.

We now address how the proposition and the corollary are used in the linear regression setting presented in Section 2.1. First, Theorem 1.2.4 from Pearl [1] states the quasi-equivalence between d-separation and conditional independence. That is, unless a very precise tuning of parameters occurs, d-separation of Y and Dj by {XZ} is equivalent to conditional independence between Y and Dj given {XZ}. Hence, we can replace d-separation by conditional independence in Proposition 2.1 and in Corollary 2.1. Under the assumption that all variables in the graph G are multivariate normal, we have that conditional independence is equivalent to zero partial correlation and thus to zero regression parameter in the linear model [11]. More specifically, if Y and Dj are conditionally independent given {XZ}, then the regression parameter associated to Dj in the linear regression of Y on Dj, X and Z is 0; and this parameter is 0 only if Y and Dj are conditionally independent given {XZ}. The assumption of multivariate normality is quite stringent; a weaker assumption is that model (1) is correctly specified (see Appendix B).

2.3 The BCEE algorithm

BCEE is viewed as a BMA procedure where the prior distribution of the outcome model is informative and constructed by using estimates from earlier steps of the algorithm, including the exposure model. In this section, we introduce BCEE and define the aforementioned prior distribution. The connections between Proposition 2.1, Corollary 2.1 and BCEE’s prior distribution are discussed in Section 2.4.

We now define the outcome model using the same model averaging notation as in BAC and TBAC. Let αY=(α1Y,...,αMY) be an M-dimensional vector for the inclusion of the covariates U in the outcome model, where component αmY equals 1 if covariate Um is included in the model and αmY equals 0 if covariate Um is not included, m=1,...,M. Letting i index the units of observation, i=1,...,n, the outcome model is the following normal linear model

(4)Yi=δ0αY+βαYXi+m=1MαmYδmαYUim+ϵiαY,

where δmαY and βαY denote respectively the unknown regression coefficients associated with Um and X in the outcome model specified by αY. The parameter δ0αY denotes the unknown intercept in model αY and the distribution of the error terms is given by ϵiαYiidN(0,σαY2).

Given model (4) and a prior distribution P(αY), the use of BMA for the estimation of the exposure effect requires first obtaining the posterior distribution of the outcome model P(αY|Y)P(Y|αY)P(αY). Standard implementation of BMA often involves selecting a uniform prior distribution P(αY)=1/2MαY, in which case P(αY|Y)P(Y|αY). The model-averaged exposure effect is then given by

(5)E[β|Y]=αYβαYPβαY|αY,YdβαYPαY|Y.

In BCEE, we utilize an informative prior distribution rather than the usual non-informative one. This distribution aims to give the bulk of the prior probability to outcome models in which βαY has a causal interpretation according to Proposition 2.1, and that cannot be reduced according to Corollary 2.1. As will be seen, this prior distribution is constructed by borrowing information from the data.

The first step in the construction of BCEE’s prior distribution PB(αY) is to compute the posterior distribution of the exposure model. This step is also present in TBAC and is performed in BCEE to identify possible causal exposure models and thus likely direct causes of the exposure. Recall that direct causes of exposure play a pivotal role in both Proposition 2.1 and Corollary 2.1. We now introduce the notation for the exposure model. Let αX=(α1X,...,αMX) be an M-dimensional vector for the inclusion of the covariates U in the exposure model. The exposure model is the following normal linear model

(6)Xi=δ0αX+m=1MαmXδmαXUim+ϵiαX,

where δmαX denotes the unknown regression coefficient of Um, m=1,...,M, in the exposure model specified by αX. The parameter δ0αX denotes the unknown intercept in αX and ϵiαXiidN(0,σαX2). In this step, each model αX is attributed a weight corresponding to its posterior probability, P(αX|X)P(X|αX)P(αX). For simplification, P(αX) is taken to be uniform (that is, P(αX)=1/2MαX), although other prior distributions could be considered.

We are now ready to define PB(αY), which depends not only on P(αX|X), but also on the regression coefficients δmαY. Remember that Proposition 2.1 and Corollary 2.1 both require verifying conditional independences. This can be achieved through the examination of the outcome model regression coefficients (see the final remarks of Section 2.2). To simplify the presentation, we assume for now that the true values of the regression coefficients are provided by an oracle. The BCEE prior distribution is as follows:

PB(αY)=αXPBαY|αXP(αX|X),where
PBαY|αXm=1MQαYαmY|αmX.

For vectors αY and αX, QαYαmY|αmX is given by one of the following:

(7)QαYαmY=1|αmX=1=ωmαYωmαY+1,QαYαmY=0|αmX=1=1ωmαY+1,QαYαmY=1|αmX=0=12,QαYαmY=0|αmX=0=12,

where ωmαY is defined in (8). To properly define ωmαY we must first define the notion of an m-nearest neighbor outcome model. For a given model αY where αmY=0, the m-nearest neighbor model to αY, αY(m), has exactly the same covariates as αY except with αmY=1 instead of αmY=0. In the case where αmY=1, there is no need to define an m-nearest neighbor model. We now define a new set of regression parameters:

δ˜mαY={δmαYifαmY=1δmαY(m)ifαmY=0.

For example, if U={U1,U2} and αY=(1,0) then δ˜1αY=δ1αY can be directly taken from model αY, whereas δ˜2αY=δ2αY(2) needs to be taken from model αY(2)=(1,1).

With this additional notation, we define the hyperparameter ωmαY as:

(8)ωmαY=ω×δ˜mαYσUmσY2,

where 0ω is a user-defined hyperparameter, σUm and σY are respectively, the (true) standard deviations of Um and Y. Note that δ˜mαYσUm/σY is a standardization of δ˜mαY which makes it insensitive to the measurement units of both Y and Um. In practice, we cannot rely on an oracle to provide δ˜mαY; in the sequel, we use the maximum likelihood estimator of δ˜mαY instead. Also, the true values of σUm and σY are not known and are estimated by sUm and sY. The prior distribution PB(αY) thus has an empirical Bayes flavor. Once PB(αY) is obtained, the posterior distribution of the outcome model P(αY|Y) is computed and the posterior exposure effect calculated according to eq. (5). In Section 3.2, we discuss how one can account for using the data for the specification of PB(αY) to obtain appropriate inferences.

2.4 The rationale behind BCEE

In this section, we explain in detail how BCEE’s prior distribution PB(αY) is motivated by causal graphs through Proposition 2.1 and Corollary 2.1.

To begin, recall that the first step of BCEE serves to identify likely exposure models. Classical properties of Bayesian model selection ensure that the true (structural) exposure model, the one including only and all direct causes of X (D={D1,...,DJ}), is asymptotically attributed all the posterior probability by the first step of BCEE (e.g. Haughton [12], Wasserman [13]). This result follows from assuming that the set of potential confounding covariates U includes all direct causes of X and no descendants of X and that the specification of the model is correct: that is, the true exposure model is indeed a normal linear model of the form Xi=δ0+j=1JδjDij+ϵiX, with ϵiXiidN(0,σX2).

The algorithm BCEE aims to give the bulk of the posterior weight to outcome models in which βαY has a causal interpretation according to Proposition 2.1 and that cannot be reduced according to Corollary 2.1. In such outcome models, αY includes any given direct cause (identified in the first step) only if the inclusion of this direct cause of exposure is necessary for βαY to have a causal interpretation in αY. To do so, PB(αY) places small prior weight on outcome models which do not respect condition 2 of Proposition 2.1. In such models, some direct causes of X are excluded (condition point 2a from Proposition 2.1) and Y is dependent on those excluded direct causes of X given X and the potential confounding covariates already included (condition point 2b from Proposition 2.1). Moreover, PB(αY) seeks to limit the prior weight attributed to outcome models that could be reduced according to Corollary 2.1. In such models, some direct causes of X are included, but these are not associated with Y conditionally on X and the other covariates included.

To illustrate how Proposition 2.1 and Corollary 2.1 motivate the formulation of PB(αY) we provide the following thought experiment. To simplify our presentation, we assume that the direct causes of exposure are known and that the outcome model (1) is correctly specified. Moreover, we order the elements of U so that the first J elements are D, that is {U1,...,UM}={D1,...,DJ,UJ+1,...,UM}. For ease of interpretation, we also assume that the covariates U are standardized, although, due to the way ωmαY is defined, this is not necessary in practice. We consider four different situations to illustrate how BCEE functions. In each situation, a direct cause of exposure Dj=Uj is either included or excluded from the outcome model αY and the maximum likelihood estimate |δ˜^jαY| is either close to 0 or large. The anticipated magnitudes of QαY(αjY|αjX) and of PB(αY|αX) for each situation are presented in Table 1. Considering jointly those four situations, we see that only outcome models that both correctly identify the average causal effect of exposure and that solely include necessary direct causes of exposure receive non-negligible prior probabilities. In the next paragraph, we describe in detail the first situation, which supposes that direct cause Dj is omitted from αY and its associated estimated parameter |δ˜^jαY| is large.

Suppose αY does not include Dj. Note that QαY(αjY=0|αjX=1) depends on δ˜^jαY through ωjαY. Therefore, PB(αY|αX) also depends on δ˜^jαY. If |δ˜^jαY| is large, then Y is likely not independent of Dj conditionally on X and the potential confounding covariates included in αY. It follows that αY does not respect condition 2b from Proposition 2.1. Since the value of ωjαY is large, QαY(αjY=0|αjX=1) is small and so is PB(αY|αX). In this situation, PB(αY) is well behaved: the model αY is not sufficient to identify the average causal effect of exposure and hence it receives little prior probability. A similar reasoning can be applied for situation 4 of Table 1. The reasoning for situations 2 and 3 is also quite similar, but requires invoking Corollary 2.1 to determine whether the inclusion of Dj is necessary or not.

Table 1:

Magnitudes of QαY(αjY|αjX) and PB(αY|αX) for four situations defined by the inclusion of a direct cause of exposure Dj and the magnitude of |δ˜^jαY|.

SituationDjδ˜^jαYYDjX,ZωjαYQαY(αjYαjX)PB(αYαX)
(1)Excl.LargeNot likelyLargeClose to 0Close to 0
(2)Incl.Close to 0LikelyClose to 0Close to 0Close to 0
(3)Incl.LargeNot likelyLargeClose to 1Depends
(4)Excl.Close to 0LikelyClose to 0Close to 1Depends

Remark in Table 1 that in situations 3 and 4, where QαY(αjY|αjX) is close to 1, PB(αY|αX) depends in a large part on the QαY associated with the other direct causes of exposure. If none of the QαY are close to 0, then PB(αY|αX) is non-negligible and hence favors models that identify the causal effect according to Proposition 2.1 and Corollary 2.1. However, if any of the QαY is close to 0, then PB(αY|αX) is close to 0.

2.5 A toy example

We consider a toy example to gain preliminary insights on the finite sample properties of BCEE. We generated a sample of size n=500 satisfying the following relationships:

X=U1+U2+ϵX
Y=X+0.1U1+ϵY,

with U1, U2N(0,1) and ϵX, ϵYN(0,1), all independent.

The first step of BCEE is to calculate the posterior distribution of the exposure model P(αX|X). The four possible exposure models in this example are:

α1X:X|(α1X=0,α2X=0),
α2X:X|U1(α1X=1,α2X=0),
α3X:X|U2(α1X=0,α2X=1),
α4X:X|U1,U2(α1X=1,α2X=1).

We approximate P(X|αX) using exp[0.5BIC(αX)] [14], where BIC(αX) is the Bayesian information criterion for exposure model αX. In our example, model α4X receives all posterior weight, that is P(αX=(1,1)|X)=1.

Next, we compute the posterior distribution of the outcome model using PB(αY). We take ω=100n, a choice that is subsequently discussed in Section 3.1. The four possible outcome models are:

α1Y:Y|X(α1Y=0,α2Y=0),
α2Y:Y|X,U1(α1Y=1,α2Y=0),
α3Y:Y|X,U2(α1Y=0,α2Y=1),
α4Y:Y|X,U1,U2(α1Y=1,α2Y=1).

Note that only models α2Y and α4Y correctly identify the causal effect of exposure. We present the calculation of PB(αY|αX) for model α2Y. Since we obtained P(αX=(1,1)|X)=1, we only need to calculate PB(αY=(1,0)|αX=(1,1))Qα2Y(α1Y=1|α1X=1)Qα2Y(α2Y=0|α2X=1). We have:

Qα2Y(α1Y=1|α1X=1)=ω1α2Yω1α2Y+1,
Qα2Y(α2Y=0|α2X=1)=1ω2α2Y+1.

We get ω1α2Y=ω(δ˜^1α2Y×sU1/sY)2=100500(0.14×1.00/2.01)2=9.75. Note that because U1 is included in α2Y, δ˜^1α2Y=δ^1α2Y. Also, we have ω2α2Y=ω(δ˜^2α2Y×sU2/sY)2=100500(0.01×1.04/2.01)2=0.05. Because U2 is not in α2Y, we get the regression parameter estimate for U2 from its 2-nearest neighbor model, that is δ˜^2α2Y=δ^2α4Y. Finally, the value of the (unnormalized) prior probability of model α2Y is 0.8658.

Following the same process for the three other outcome models, we calculate the prior probabilities. From there, we calculate the posterior distribution of the outcome model using the relationship P(αY|Y)P(Y|αY)PB(αY). Again, we use exp[0.5BIC(αY)] to approximate P(Y|αY). Table 2 provides the results with the details of the intermediate steps.

Table 2:

Calculation of the BCEE outcome model posterior distribution with intermediate steps.

ModelU.PB(αY)PB(αY)BICBMA P(αY|Y)P(αY|Y)
α1Y0.02300.02291,435.820.46020.0254
α2Y0.86580.86181,435.810.46290.9625
α3Y0.07490.07461,440.040.05600.0101
α4Y0.04090.04071,442.000.02090.0021

We see from these results how BCEE, as compared to BMA, shifts the posterior weight toward models that identify the causal effect of exposure. In fact, in this toy example, BCEE puts almost all the posterior weight on the true outcome model. BCEE accomplishes this by using an informative prior distribution for the outcome model that borrows information both from the exposure selection step and from neighboring regression coefficient estimates in the outcome models.

3 Practical considerations regarding BCEE

In this section we discuss practical considerations regarding the usage of the BCEE algorithm. First, we discuss the choice of the hyperparameter ω value in eq. (8), then we suggest two alternative ways of implementing BCEE.

3.1 Choice of ω

Recall that BCEE’s prior distribution PB(αY) depends on a user-selected hyperparameter ω. In what follows, we suggest making ω proportional to n on the basis of asymptotic results related to the quantities QαY in eq. (7). Without loss of generality, we only discuss the case QαY(αmY=1|αmX=1). Indeed, the cases QαY(αmY=1|αmX=0) and QαY(αmY=0|αmX=0) are trivial because the two quantities are both equal to 1/2. Moreover, the case QαY(αmY=0|αmX=1) is essentially equivalent to the case QαY(αmY=1|αmX=1) since these quantities are closely (and negatively) associated. Remark that because we consider the case where αmY=1, δ˜mαY=δmαY. However, we present the reasoning in terms of δ˜mαY to allow a direct generalization to the case where αmY=0.

Assume that the true outcome model is a normal linear model of the form (1) and first consider the case δ˜mαY=0 for a given model αY. Then covariate Um is conditionally independent of Y given the (other) covariates included in model αY. Hence Um should be left out of αY on the basis of Corollary 2.1. It is thus desirable that QαY(αmY=1|αmX=1)0 as n, which happens if ω^mαY=ω×(δ˜^mαYsUm/sY)20 as n.

Consider the case δ˜mαY0. According to Proposition 2.1, it is now desirable that QαY(αmY=1|αmX=1)1 as n, since this would allow for covariates causing less confounding to be forced in the outcome model as n grows. Thus, we need ωˆmαY as n if δ˜mαY0.

If δ˜mαY=0 then δ˜^mαYP0 and thus, for any finite constant value of ω, ωˆmαYP0, where P means convergence in probability. However, if δ˜mαY0, we need to choose ω as a function of sample size n in order to ensure that ωˆmαY as n. We consider rates of convergence to find an appropriate function of n.

Recall that ωˆmαY is a function of the MLE δ˜^mαY (Section 2.3). Under mild regularity conditions, it follows from the results in Yuan and Chan [15] that δ˜^mαYsUm/sYPδ˜mαYσUm/σY at rate Op(1/n), where Op is the usual big-Op notation (Agresti [16], p. 588). Thus (δ˜^mαYsUm/sY)2P(δ˜mαYσUm/σY)2 at rate Op(1/n).

By taking ω=cnb, with 0<b<1, where c is a user-fixed constant that does not depend on sample size, we obtain ωˆmαY (at rate nb) if δ˜mαY0 and ωˆmαYP0 (with convergence rate Op(1/n1b)) if δ˜mαY=0, as desired. The value b=1/2 appears to make a good compromise between the two desired convergence behaviors. The simulation study presented in Section 4 shows that BCEE performs well for ω=cn with 100c1000. We also see that larger values of c yield less bias and more variance in the estimator of the causal effect, and conversely for smaller values of c. Appendix C illustrates how QαY(αmY=1|αmX=1) behaves for different values of c in some simple settings.

3.2 Implementing BCEE

In this section, we first consider a naive implementation of BCEE that closely follows our presentation of the algorithm in Section 2.3. Then we describe a modified implementation that accounts for using the MLE δ˜^mαY in PB(αY).

We perform three steps to sample one draw from the posterior distribution of the average causal exposure effect P(β|Y). Several such draws are taken to obtain approximations to quantities of interest, such as the posterior mean and variance of β. The steps of the sampling procedure are:

  1. Draw αX from the posterior distribution of the exposure model P(αX|X)P(X|αX), using exp[0.5BIC(αX)] to approximate P(X|αX);

  2. Draw αY from the conditional posterior distribution P(αY|αX,Y)PB(αY|αX)P(Y|αY), where the regression coefficients δ˜mαY are estimated by their MLEs and P(Y|αY) is approximated by exp[0.5BIC(αY)];

  3. Draw β from the conditional posterior distribution P(βαY|αY,Y), which we approximate by its limit normal distribution NβˆαY,SEˆ(βˆαY) [17, 18], where βˆαY is the maximum likelihood estimator of βαY and SEˆ(βˆαY) is its estimated standard error.

The sampling for the first two steps is done using Markov chain Monte Carlo model composition (MC3) [19]. We refer to this naive implementation of BCEE as N-BCEE.

Because N-BCEE does not take into account the uncertainty related to the estimation of the regression coefficients δ˜mαY in PB(αY), we anticipate that the confidence (credible) interval for β will be too narrow. Our insight relies on the Empirical Bayes literature, where it has been extensively shown that data-dependent prior distributions lead to confidence intervals that tend to be “too short, inappropriately centered, or both” [20]. Also, narrow confidence intervals for β are observed in simulations presented in Section 4. Although many solutions to this problem have been proposed (see Carlin and Louis [21] for a short discussion), most cannot be realistically applied to BCEE due to the complexity of the algorithm. Therefore, we propose the following simple ad hoc solution, which happens to be notably faster than N-BCEE. We refer to this modified implementation of BCEE as A-BCEE.

A-BCEE is the same as N-BCEE except for step S2. Recall that this step is directed at sampling from the conditional posterior distribution P(αY|αX,Y) using MC3. This MC3 scheme requires calculating a Metropolis-Hasting ratio (RP) which involves the ratio of the (conditional) prior probabilities of the proposed outcome model, α1Y, to the current outcome model, α2Y:

(9)RP=PBα1Y|αXPBα2Y|αX=m=1MQα1YαmY|αmX/Cm=1MQα2YαmY|αmX/C=m=1MQα1YαmY|αmXQα2YαmY|αmX,

where C is a normalizing constant such that PB(αY|αX)=m=1MQαYαmY|αmX/C. In RP, α1Y and α2Y are two neighbor outcome models that differ only by their inclusion of a single covariate Um. A-BCEE utilizes the following simplification for RP:

(10)RPQα1YαmY|αmXQα2YαmY|αmX.

The heuristic for suggesting this approximation is that the individual ratio that is the most likely to significantly differ from 1 in eq. (9) is the one associated to covariate Um, that is Qα1YαmY|αmX/Qα2YαmY|αmX. In fact, unless the covariates U are very strongly correlated with each other, we expect the δ˜^mαYs (mm) to be of the same magnitude between two neighboring models. Note that we also expect many terms in the RP product to be exactly equal to 1 since an individual ratio equals 1 when its corresponding covariate is not included in the exposure model Qα1YαmY|αmX=0/Qα2YαmY|αmX=0=1. Simulations were performed to verify the validity of approximation (10) (results not presented).

Using simplified RP (10), it becomes an easy task to incorporate the variability associated with the estimation of the δ˜αYs. We assume that δ˜mαY~N(δ˜^mαY,SE^(δ˜^mαY)), where SE^(δ˜^mαY) is the estimated standard error of δ˜^mαY. In summary, in step S2 of the sampling procedure of A-BCEE we simply draw δ˜mαY from N(δ˜^mαY,SE^(δ˜^mαY)) and use it in approximation (10). We remark that this strategy is akin to specifying an empirical Bayes type of hyperprior for δ˜αY.

The finite sample properties of N-BCEE and A-BCEE are studied and compared in some simulation scenarios presented in the next section. We also consider nonparametric bootstrap [22] in a few simple and small scale simulations as an alternative to A-BCEE to correct confidence intervals. Note that, due to computing time, this bootstrapped BCEE (B-BCEE) approach is considerably less practical than A-BCEE to evaluate in simulations and to apply to real data sets of moderate to large sizes.

4 Simulation studies

In this section, we study the finite sample properties of BCEE in various simulation scenarios. The first primary objective of the simulations is to compare BCEE to standard or related methods that are used to estimate total average causal effects of exposure. The second primary objective is to study the sensitivity of BCEE to the choice of its user-selected hyperparameter ω. In Appendix D, we study two other secondary objectives relating to the large, whilst finite, properties of BCEE and to the performance of B-BCEE.

To achieve the two main objectives, we examine 24 different simulation scenarios obtained by considering three factors: data-generating process (DGP1, DGP2, DGP3 and DGP4), sample size (200, 600 and 1,000) and true causal effect of exposure (β=0.1 or β=0). The four data-generating processes are described below.

The first data-generating process (DGP1) satisfies the following relationships:

U3=U2+ϵ3
U5=U4+ϵ5
X=U1+U2+U4+ϵX
Y=U3+0.1U4+U5+βX+ϵY,

with U1,U2,U4,ϵ3,ϵ5,ϵX,ϵYN(0,1) all independent. The set of available covariates is U={U1,U2,,U5}.

The second data-generating process (DGP2) involves a larger number of covariates than DGP1 and features an indirect effect of X on Y:

U1=U4+ϵ1
U2=U4+ϵ2
U3=U4+ϵ3
U5=U1+ϵ5
X=U1+U2+U3+ϵX
U6=0.5X+U3+ϵ6
Y=0.1U4+0.1U5+βU6+0.5βX+ϵY,

where U4,ϵ1,ϵ2,ϵ3,ϵ5,ϵX,ϵ6,ϵYN(0,1) all independent. The set of available covariates is U={U1,U2,,U5,U7,,U15}, where U7,,U15 are all independent N(0,1). We exclude U6 from the set of potential confounding covariates since one must not adjust for descendants of the exposure X to identify the total average causal effect. Here the total effect of X on Y (direct effect plus indirect effect through U6) is 0.5β+0.5β=β. For simulation purposes, we consider the model αY=(0,0,1,1,1,0,...,0) as the “true” outcome model.

The third data-generating process (DGP3) is similar to the first simulation example in Wang et al. [6] but includes only 18 additional (noise) covariates (instead of 49):

X=0.7U1+(10.72)ϵX
Y=0.1U1+0.1U2+βX+ϵY,

where U1,U2,ϵX,ϵYN(0,1) all independent. The set of available covariates is U={U1,U2,,U20}, where U3,,U20 are also independent N(0,1).

The fourth data-generating process (DGP4) is inspired by a DAG presented in Morgan and Winship [23], Figure 1.1, page 25:

X=0.1U1+0.1U2+0.1U3+ϵX
U6=U3+ϵ6
Y=0.1U4+0.5U5+0.5U6+βX+ϵY,

where ϵX,ϵ6,ϵYN(0,1) all independent. Covariates U1,U2,U3,U4,U5 are also N(0,1) and are all independent except U1,U2 and U1,U4 for which we have Cov(U1,U2)=0.7 and Cov(U1,U4)=0.7. Notice that U1 is a collider between U2 and U4 and thus Cov(U2,U4)=0.

For each of the 24 simulation scenarios, we randomly generated 500 datasets. We estimated the average causal effect of exposure using 8 different procedures: (1) the true outcome model, (2) the fully adjusted model, (3) Bayesian model averaging (BMA) with a uniform prior distribution on the outcome model, (4) Bayesian adjustment for confounding (BAC) with ω chosen with cross-validation criterion CVm(ω) proposed in Lefebvre et al. [7], (5) BAC with ω=, (6) Two-stage Bayesian adjustment for confounding (TBAC) with ω=, (7) N-BCEE, and (8) A-BCEE. For both N-BCEE and A-BCEE, we used ω=cn and considered c=100, c=500 and c=1000. For each scenario and each method of estimation, we computed the average causal effect estimate (Mean), the average standard error estimate (SEE), the standard deviation of the estimates (SDE), the root mean squared error (MSE) and the coverage probability of 95% confidence intervals (CP). All 95% confidence intervals were computed using the normal approximation βˆ±1.96SEE. Tables 3, 4, 5 and 6 summarize the results for β=0.1. The marginal posterior probability of inclusion of each potential confounding covariate can be found in Tables 11 to 14 in Appendix E. The results for β=0 are similar (not presented).

Table 3:

Comparison of estimates of β obtained from the true outcome model, the fully adjusted model, BMA, BAC, TBAC, N-BCEE, and A-BCEE for 500 Monte Carlo replicates of the first data-generating process (DGP1).

nMethodMeanSEESDEMSECP
200True model0.1000.0450.0470.04794
200Fully adjusted model0.0980.0720.0740.07494
200BMA0.1130.0470.0470.04895
200BAC (CVm(ω))0.1040.0550.0640.06492
200BAC (ω=)0.0980.0720.0740.07494
200TBAC (ω=)0.0980.0720.0740.07494
200N-BCEE (c=100)0.1080.0510.0550.05693
200N-BCEE (c=500)0.1040.0550.0620.06292
200N-BCEE (c=1000)0.1020.0570.0650.06593
200A-BCEE (c=100)0.1070.0550.0540.05495
200A-BCEE (c=500)0.1040.0610.0600.06096
200A-BCEE (c=1000)0.1030.0630.0630.06396
600True model0.1000.0260.0250.02596
600Fully adjusted model0.1000.0410.0390.03997
600BMA0.1110.0270.0270.02994
600BAC (CVm(ω))0.1050.0310.0350.03595
600BAC (ω=)0.1000.0410.0390.03997
600TBAC (ω=)0.1000.0410.0390.03996
600N-BCEE (c=100)0.1080.0290.0310.03193
600N-BCEE (c=500)0.1060.0300.0330.03493
600N-BCEE (c=1000)0.1050.0310.0340.03493
600A-BCEE (c=100)0.1080.0300.0300.03195
600A-BCEE (c=500)0.1050.0330.0320.03297
600A-BCEE (c=1000)0.1050.0350.0330.03397
1,000True model0.1010.0200.0200.02095
1,000Fully adjusted model0.1000.0320.0330.03394
1,000BMA0.1110.0210.0220.02592
1,000BAC (CVm(ω))0.1020.0260.0300.03093
1,000BAC (ω=)0.1000.0320.0330.03394
1,000TBAC (ω=)0.1000.0320.0330.03394
1,000N-BCEE (c=100)0.1070.0220.0240.02594
1,000N-BCEE (c=500)0.1050.0230.0260.02694
1,000N-BCEE (c=1000)0.1040.0240.0260.02794
1,000A-BCEE (c=100)0.1070.0230.0240.02595
1,000A-BCEE (c=500)0.1050.0260.0260.02696
1,000A-BCEE (c=1000)0.1040.0270.0270.02796
Table 4:

Comparison of estimates of β obtained from the true outcome model, the fully adjusted model, BMA, BAC, TBAC, N-BCEE, and A-BCEE for 500 Monte Carlo replicates of the second data-generating process (DGP2).

nMethodMeanSEESDEMSECP
200True model0.1020.0460.0450.04596
200Fully adjusted model0.1040.0750.0780.07894
200BMA0.1480.0440.0460.06768
200BAC (CVm(ω))0.1180.0520.0750.07776
200BAC (ω=)0.1030.0730.0770.07795
200TBAC (ω=)0.1030.0730.0760.07695
200N-BCEE (c=100)0.1200.0530.0680.07183
200N-BCEE (c=500)0.1100.0580.0730.07486
200N-BCEE (c=1000)0.1070.0600.0740.07488
200A-BCEE (c=100)0.1200.0620.0660.06992
200A-BCEE (c=500)0.1120.0670.0710.07295
200A-BCEE (c=1000)0.1100.0680.0720.07395
600True model0.1000.0260.0260.02696
600Fully adjusted model0.1020.0420.0410.04195
600BMA0.1330.0300.0320.04670
600BAC (CVm(ω))0.1060.0360.0420.04285
600BAC (ω=)0.1020.0410.0410.04196
600TBAC (ω=)0.1020.0410.0410.04196
600N-BCEE (c=100)0.1140.0320.0370.04086
600N-BCEE (c=500)0.1080.0340.0390.03991
600N-BCEE (c=1000)0.1060.0350.0390.04091
600A-BCEE (c=100)0.1140.0360.0370.03992
600A-BCEE (c=500)0.1090.0380.0380.03994
600A-BCEE (c=1000)0.1070.0390.0390.03994
1,000True model0.1000.0200.0210.02195
1,000Fully adjusted model0.1000.0320.0320.03295
1,000BMA0.1210.0240.0270.03480
1,000BAC (CVm(ω))0.1000.0290.0310.03192
1,000BAC (ω=)0.0990.0320.0320.03195
1,000TBAC (ω=)0.0990.0320.0320.03195
1,000N-BCEE (c=100)0.1070.0250.0290.02990
1,000N-BCEE (c=500)0.1030.0260.0290.02990
1,000N-BCEE (c=1000)0.1020.0260.0300.03091
1,000A-BCEE (c=100)0.1080.0280.0280.02993
1,000A-BCEE (c=500)0.1040.0290.0290.02994
1,000A-BCEE (c=1000)0.1030.0300.0290.03095
Table 5:

Comparison of estimates of β obtained from the true outcome model, the fully adjusted model, BMA, BAC, TBAC, N-BCEE, and A-BCEE for 500 Monte Carlo replicates of the third data-generating process (DGP3).

nMethodMeanSEESDEMSECP
200True model0.1030.1000.1000.10095
200Fully adjusted model0.1010.1050.1040.10496
200BMA0.1490.0850.0860.09989
200BAC (CVm(ω))0.1160.0870.1030.10490
200BAC (ω=)0.1010.1000.1010.10195
200TBAC (ω=)0.1020.1010.1010.10195
200N-BCEE (c=100)0.1130.0930.1000.10193
200N-BCEE (c=500)0.1060.0960.1010.10194
200N-BCEE (c=1000)0.1040.0970.1010.10194
200A-BCEE (c=100)0.1160.0960.0980.09995
200A-BCEE (c=500)0.1090.0980.0990.10095
200A-BCEE (c=1000)0.1080.0990.1000.10095
600True model0.0980.0570.0600.06096
600Fully adjusted model0.0980.0580.0610.06196
600BMA0.1380.0540.0610.07280
600BAC (CVm(ω))0.1040.0540.0650.06587
600BAC (ω=)0.0970.0580.0600.06096
600TBAC (ω=)0.0970.0570.0600.06095
600N-BCEE (c=100)0.1080.0560.0640.06488
600N-BCEE (c=500)0.1010.0560.0620.06292
600N-BCEE (c=1000)0.1000.0560.0610.06192
600A-BCEE (c=100)0.1110.0570.0630.06490
600A-BCEE (c=500)0.1040.0570.0620.06292
600A-BCEE (c=1000)0.1030.0570.0620.06294
1,000True model0.0980.0440.0430.04396
1,000Fully adjusted model0.0980.0450.0430.04395
1,000BMA0.1300.0450.0500.05879
1,000BAC (CVm(ω))0.1020.0430.0460.04691
1,000BAC (ω=)0.0980.0450.0430.04396
1,000TBAC (ω=)0.0980.0440.0430.04396
1,000N-BCEE (c=100)0.1060.0440.0480.04891
1,000N-BCEE (c=500)0.1010.0440.0450.04593
1,000N-BCEE (c=1000)0.1000.0440.0440.04494
1,000A-BCEE (c=100)0.1080.0450.0480.04892
1,000A-BCEE (c=500)0.1030.0450.0460.04694
1,000A-BCEE (c=1000)0.1020.0450.0450.04594
Table 6:

Comparison of estimates of β obtained from the true outcome model, the fully adjusted model, BMA, BAC, TBAC, N-BCEE, and A-BCEE for 500 Monte Carlo replicates of the fourth data-generating process (DGP4).

nMethodMeanSEESDEMSECP
200True model0.1030.0540.0520.05296
200Fully adjusted model0.1050.0720.0680.06895
200BMA0.1190.0600.0540.05796
200BAC (CVm(ω))0.1100.0610.0620.06395
200BAC (ω=)0.1030.0720.0680.06895
200TBAC (ω=)0.1050.0710.0670.06796
200N-BCEE (c=100)0.1080.0610.0630.06493
200N-BCEE (c=500)0.1060.0640.0660.06694
200N-BCEE (c=1000)0.1050.0650.0660.06695
200A-BCEE (c=100)0.1100.0660.0620.06296
200A-BCEE (c=500)0.1070.0680.0640.06496
200A-BCEE (c=1000)0.1070.0680.0650.06596
600True model0.0990.0310.0310.03195
600Fully adjusted model0.0970.0410.0430.04395
600BMA0.1100.0360.0360.03892
600BAC (CVm(ω))0.1000.0370.0420.04292
600BAC (ω=)0.0960.0410.0430.04395
600TBAC (ω=)0.0960.0410.0420.04394
600N-BCEE (c=100)0.1020.0360.0400.04092
600N-BCEE (c=500)0.0990.0370.0410.04192
600N-BCEE (c=1000)0.0980.0370.0420.04292
600A-BCEE (c=100)0.1020.0380.0400.04094
600A-BCEE (c=500)0.1000.0390.0410.04194
600A-BCEE (c=1000)0.0990.0400.0410.04194
1,000True model0.0990.0240.0240.02496
1,000Fully adjusted model0.0990.0320.0320.03295
1,000BMA0.1070.0280.0290.03092
1,000BAC (CVm(ω))0.1000.0290.0320.03291
1,000BAC (ω=)0.0980.0320.0320.03294
1,000TBAC (ω=)0.0980.0320.0320.03293
1,000N-BCEE (c=100)0.1020.0280.0300.03092
1,000N-BCEE (c=500)0.1000.0280.0310.03192
1,000N-BCEE (c=1000)0.1000.0290.0320.03192
1,000A-BCEE (c=100)0.1020.0300.0310.03194
1,000A-BCEE (c=500)0.1010.0300.0310.03194
1,000A-BCEE (c=1000)0.1000.0310.0320.03294

We start by discussing the results pertaining to non-BCEE methods for estimating the average causal effect of exposure. Then, we discuss the results for BCEE and contrast them to the former results.

As expected, Bayesian model averaging (BMA) can perform very poorly to estimate the average causal effect. More precisely, the simulation results show that the bias can be substantial when the most important confounding covariates are only slightly associated with the outcome (DGP2 and DGP3). For instance, in DGP2, U3 and U4 are important confounding covariates often excluded by BMA (see Table 12 in Appendix E). Similarly, in DGP3, U1 is often excluded by BMA (see Table 13). This situation also yields confidence intervals with poor coverage probabilities. Although increasing sample size seems to reduce the bias, the coverage probability remains mostly unchanged. In situations where the most important confounding covariates are strongly associated with the outcome (DGP1 and DGP4), BMA performs very well both in terms of mean squared error (MSE) and coverage probability.

The simulation results also support the claim that BAC and TBAC with ω= do not yield a notable reduction in the variance of the estimated causal effect as compared to the fully adjusted model. This is partly due to the fact that BAC and TBAC tend to include more covariates than needed to achieve unbiasedness (see Appendix E). Moreover, using BAC with cross-validation criterion CVm(ω) gives relatively poor results. Even though this method sometimes gives smaller MSE than BAC with ω=, the estimated standard error remarkably underestimate the true standard error (the standard deviation of the estimates of β). One possible explanation for this underestimation is that BAC with CVm(ω) neglects the uncertainty associated with the choice of the hyperparameter ω.

The simulation results show that the choice of using ω=cn, c[100,1000], for A-BCEE and N-BCEE is reasonable. The results do not appear too sensitive to the choice of c in this interval. The simulation results also confirm that N-BCEE can yield lower than expected coverage probabilities. This seems to be particularly true in complex scenarios that contain many covariates, such as DGP2, DGP3 and DGP4.

Despite sometimes producing slightly biased estimates, A-BCEE performs at least as well as BAC and TBAC with ω= in terms of MSE. The bias is small enough that in all simulation scenarios we considered, A-BCEE (with any c) yields appropriate coverage probability. In general, A-BCEE gives less weight to variables only associated with the exposure than BAC and TBAC (see Appendix E). In DGP1, A-BCEE outperforms BAC and TBAC with ω= in terms of MSE. In DGP2 and DGP4, A-BCEE has smaller MSE than BAC and TBAC although comparatively to a lesser extent. Results are quite similar between BAC, TBAC and A-BCEE in DGP3. Note that in DGP3, the true model and the fully adjusted model have the same MSE. There is thus no possible gain in using another model than the fully adjusted model. Figure 1 illustrates the distribution of βˆ obtained by using A-BCEE and BAC with ω= for all four data-generating processes with n=200 (analogous figures are displayed in Appendix F with n=600 and n=1000). This figure shows how estimates obtained with A-BCEE, despite being slightly biased, are more concentrated around the true value β than estimates obtained with BAC. Moreover, Figure 1 illustrates the bias-variance tradeoff associated with the choice of c in A-BCEE: smaller values of c, as compared to larger values of c, favor a reduced variance in the estimator of the causal effect at the cost of an increase in bias.

Figure 1: Comparison of the distribution of βˆ$$\hat \beta $$ obtained from BAC (ω=∞$$\omega = \infty $$) and A-BCEE (c=100, 500, and 1,000) for all four data-generating processes and a sample size n=200$$n = 200$$. The red line corresponds to the true value β$$\beta $$=0.1.
Figure 1:

Comparison of the distribution of βˆ obtained from BAC (ω=) and A-BCEE (c=100, 500, and 1,000) for all four data-generating processes and a sample size n=200. The red line corresponds to the true value β=0.1.

On the basis of these results, we hypothesized that BCEE would perform best when (1) there are some direct causes of the exposure that are strongly associated with the exposure, and (2) there exists variables that can d-separate those direct causes from the outcome. In such situations, we expect BCEE to favor models excluding those direct causes and including the d-separating variables. To verify this, we simulated data according to a fifth data-generating scenario (DGP5) which meets these two conditions. The equations for DGP5 are:

U5=U1+U2+U3+U4+ϵ5
X=U1+U2+U3+U4+ϵX
Y=U5+βX+ϵY,

where ϵ5,ϵX,ϵY N(0,1), all independent. In this example, BCEE’s prior distribution, PB(αY), is devised to give non negligible prior weight to the two following sufficient outcome models: (i) the one including {U1,U2,U3,U4}, and (ii) the one including only {U5}. However, because the marginal likelihood of the model (ii) should dominate the one of model (i) for large n, we expect the second outcome model to receive increased posterior weight as n grows. To reduce computational burden, we only considered β=0.1 and did not estimate β with N-BCEE. The results are presented in Table 7. Those results show how under such ideal conditions, the MSE obtained by using A-BCEE is much smaller than the one obtained using the fully adjusted outcome model, BAC or TBAC. In fact, A-BCEE’s MSE is similar to the MSE of the true outcome model. Moreover, Table 15 in Appendix E reveals that models including U5, but excluding U1, U2, U3 and U4 are favored by A-BCEE, particularly for the larger sample sizes. Indeed, the marginal posterior probabilities of covariates U1 to U4 decrease with sample size while the posterior probability of U5 remains at 1 for all sample sizes considered. This is as opposed to BAC and TBAC where the full model (including U1 to U5) receives a posterior probability of 1 at all sample sizes.

Table 7:

Comparison of estimates of β obtained from the true outcome model, the fully adjusted model, BMA, BAC, TBAC and A-BCEE for 500 Monte Carlo replicates of the fifth data-generating process (DGP5).

nMethodEstimateSEESDEMSECP
200True model0.1030.0530.0540.05492
200Fully adjusted model0.1020.0720.0760.07694
200BMA0.1030.0540.0550.05593
200BAC (CVm(ω))0.1020.0590.0660.06692
200BAC (ω=)0.1020.0720.0760.07694
200TBAC (ω=)0.1020.0720.0760.07695
200A-BCEE (c=100)0.1030.0550.0560.05693
200A-BCEE (c=500)0.1030.0590.0590.05994
200A-BCEE (c=1000)0.1020.0610.0610.06195
600True model0.0990.0310.0290.02996
600Fully adjusted model0.0970.0410.0400.04096
600BMA0.0990.0310.0290.02996
600BAC (CVm(ω))0.0970.0340.0360.03695
600BAC (ω=)0.0970.0410.0400.04096
600TBAC (ω=)0.0970.0410.0400.04096
600A-BCEE (c=100)0.0980.0310.0300.03096
600A-BCEE (c=500)0.0980.0330.0300.03097
600A-BCEE (c=1000)0.0980.0340.0310.03197
1,000True model0.1000.0240.0230.02395
1,000Fully adjusted model0.1000.0320.0310.03194
1,000BMA0.1000.0240.0230.02396
1,000BAC (CVm(ω))0.1010.0270.0270.02795
1,000BAC (ω=)0.1000.0320.0310.03194
1,000TBAC (ω=)0.1000.0320.0310.03195
1,000A-BCEE (c=100)0.1000.0240.0230.02396
1,000A-BCEE (c=500)0.1000.0250.0230.02396
1,000A-BCEE (c=1000)0.1000.0250.0230.02396

5 Application: estimation of the causal effect of perceived mathematical competence on grades in mathematics

In this section we use A-BCEE to estimate the causal effect of perceived competence in mathematics (measured on a scale from 1 to 7) on self-reported grades (in %) in mathematics. We consider longitudinal data obtained from 1,430 students during their first three years of highschool. Participants lived in various regions throughout Quebec, Canada. The data were collected by postal questionnaires every year for a period of three years (time 1, time 2 and time 3). Further details can be found in Guay et al. [24].

We used measures of perceived competence in mathematics at time 2 as the exposure and grades in mathematics at time 3 as the outcome to estimate the causal effect of interest. Recall that A-BCEE requires specifying a set of potential confounding covariates that includes all direct causes of the exposure and none of its descendants. Moreover, it is beneficial that this set also includes strong predictors of the outcome. We took advantage of the longitudinal feature of the data to build the set of potential confounding covariates. Because a cause always precedes its effect in time, we constructed the set of potential confounding covariates by including variables at time 1 that were potential direct causes of perceived-competence at time 2. We also included variables at time 2 that were thought to be strong predictors of grades in mathematics at time 3.

We selected the following 26 covariates: gender, highest level of education reached by the mother, highest level of education reached by the father, perceived competence in mathematics (at time 1), perceived autonomy support from the mother, perceived autonomy support from the father, perceived autonomy support from the mathematics teacher, perceived autonomy support from friends at school, self-reported mathematics’ grades, intrinsic motivation in mathematics, identified motivation in mathematics, introjected motivation in mathematics, externally regulated motivation in mathematics, victimization and sense of belonging to school. All variables except the first four were considered both at times 1 and 2.

Before applying A-BCEE on these data, we obtained some descriptive statistics. We drew scatter plots of the outcome versus the exposure and versus each potential confounding covariate to roughly verify the linearity assumption and to check for outliers. For the same reasons, we drew scatter plots of the exposure versus each potential confounding covariate. We also noticed that only 46.5% of the participants have complete information for all the selected covariates. The variables measured at time 1 have generally few missing cases (between 1.8% and 8.3%), but the variables measured at times 2 and 3 have a larger degree of missingness (between 26.4% and 36.4%). We performed multiple imputation [25] to account for the missing data, using 50 imputed datasets to ensure the power falloff is negligible [26].

We estimated the causal effect of perceived competence on grades in mathematics using the fully adjusted outcome model, A-BCEE with ω=cn (c=100, 500, 1,000), BAC, and TBAC (with ω=). Results are summarized in Table 8. The computational burden of BCEE on these data is manageable and comparable to the one of TBAC, although quite heavier than the one of BAC when using the BACprior package [27]. The approximate running times of A-BCEE, BAC, and TBAC on one imputed dataset are respectively, 22.5, 1.2, and 21.2 min on a PC with 2.4 GHz and 8 Gb RAM.

Because Step S1 of A-BCEE aims to find the direct causes of the exposure, it is reasonable to only allow covariates measured before the exposure to be selected in this step. Hence, we ran the A-BCEE algorithm a second time, but this time excluding the possibility that covariates measured at time 2 enter the exposure model. We denote this implementation of A-BCEE as A-BCEE* in Table 8.

Table 8:

Comparison of the estimated causal effect of perceived mathematical competence in mathematics on self-reported mathematics’ grades.

MethodEstimateSEECI
Fully adjusted model0.6930.460(–0.208, 1.594)
BAC (ω=)0.7290.462(–0.178, 1.635)
TBAC (ω=)0.7780.465(–0.133, 1.690)
A-BCEE (c=100)0.8070.451(–0.076, 1.691)
A-BCEE (c=500)0.7900.456(–0.105, 1.685)
A-BCEE (c=1000)0.7860.459(–0.113, 1.685)
A-BCEE* (c=100)0.8230.445(–0.049, 1.696)
A-BCEE* (c=500)0.8080.444(–0.062, 1.679)
A-BCEE* (c=1000)0.8030.444(–0.066, 1.673)

Table 8 shows that the results from A-BCEE and A-BCEE* are very similar. This is not surprising since the marginal posterior probability of inclusion of covariates do not differ much between A-BCEE and A-BCEE* (not shown). Using A-BCEE instead of the fully adjusted model slightly decreases the standard error of estimate, between 0.3% and 3.5%, which translates in a small decrease of the 95% confidence intervals’ width. Moreover the standard errors of estimate for BAC and TBAC are slightly larger than the one for the fully adjusted model in this illustration. Although the point estimates appear to vary substantially between methods, the differences are small relative to the magnitude of the estimated standard errors. We conclude that perceived competence in mathematics at one point in time likely has little or no causal effect on self-reported grades in mathematics a year later.

6 Discussion

We have introduced the Bayesian causal effect estimation (BCEE) algorithm to estimate causal exposure effects in observational studies. This novel data-driven approach avoids the need to rely on the specification of a causal graph and aims to control the variability of the estimator of the exposure effect. BCEE employs a prior distribution that is motivated by a theoretical proposition embedded in the graphical framework to causal inference. We also proposed a practical implementation of BCEE, A-BCEE, that accounts for the fact that this prior distribution uses information from the data. Using simulation studies, we found that A-BCEE generally achieves at least some reduction of the MSE of the causal effect estimator as compared to the MSE generated by a fully-adjusted model approach or by other data-driven approaches to causal inference, such as BAC and TBAC, thus resulting in estimates that are overall closer to the true value. In some circumstances, the reduction of the MSE can be substantial. Moreover, confidence intervals with appropriate coverage probabilities were obtained. Hence, we believe that BCEE is a promising algorithm to perform causal inference.

Some current limitations of BCEE could be addressed in future research. The generalization to non continuous exposure variable (e.g. binary) is straightforward. Recall that the first step of BCEE aims at identifying the direct causes of the exposure. As in the normal case we have considered, classical Bayesian procedures asymptotically select the true exposure model with probability 1 when assuming X belongs to an exponential family (e.g. Bernoulli) and that an adequate parametric model is considered [12]. The generalization of BCEE to other types of outcome variables is less straightforward. One could specify a generalized linear model for the outcome of the form g(EYi|Xi,Ui)=δ0+βXi+m=1MδmUim. However, unless g is the identity or the log link, such models are generally not collapsible for β over covariate Um [28]. In other words, the true value of β, and thus its interpretation, depends on whether Um is included or not in the outcome model, even when Um is a not confounding covariate. In such circumstances, averaging the estimated value of β over different outcome models would not be advisable.

We think that BCEE can be particularly helpful to those working in fields where current subject-matter knowledge is sparse. To facilitate usage of the BCEE algorithm, we provide an R package named BCEE (available at http://cran.r-project.org).

Funding statement: Funding: Government of Canada – Natural Sciences and Engineering Research Council of Canada, Fonds de Recherche du Québec – Santé, Fonds de Recherche du Québec – Nature et Technologies.

Acknowledgements

Dr. Talbot has been supported by doctoral scholarships from the Natural Sciences and Engineering Research Council of Canada (NSERC) and the Fonds de recherche du Québec: Nature et Technologies. Dr. Lefebvre’s is a Chercheur-Boursier from the Fonds de recherche du Québec – Santé and is also supported by NSERC. Dr. Atherton’s research is supported by NSERC.

Appendix

A Proofs

A.1 Proof of Proposition 2.1

Proof. First, we know from Pearl [1] section 3.3.1 that a set Z is sufficient for identifying the causal effect of an exposure X on an outcome Y if (i) no descendants of X are in Z and (ii) Z blocks all back-door paths between X and Y. According to condition 1 we assume that there are no descendants of X in Z. Suppose that G admits some back-door paths. All back-door paths are such that the second variable appearing in the path is a direct cause of X; the back-door paths thus have the form XDjY.

Suppose that a direct cause Dj is included in Z. Then Dj (and therefore the set Z) blocks all back-door paths of the form XDjY. Indeed no variable in ZDj can reopen a path XDjY once closed by Dj. Therefore, all back-door paths admitting a direct cause in Z are blocked by Z.

It remains to show that all back-door paths for which the second variable in the path is not a direct cause included in Z are closed when condition 2b in the proposition holds. Consider DjZ. Now assume that Y and Dj are d-separated by {XZ}. By the definition of d-separation, this means that every path connecting Dj to Y is blocked by {XZ}. Recall that all back-door paths associated with this Dj are of the form XDjY. Because by (2b) Dj and Y are d-separated by {XZ} and since each subpath DjY in these back-door paths does not contain the variable X, these subpaths are blocked by Z. This reasoning is applied to each DjZ separately.

The proof is complete by the back-door criterion as we realize that all back-door paths, whether their Dj is contained in Z or not, are blocked by Z. □

A.2 Proof of Corollary 2.1

Proof. 1. Suppose that G admits some back-door paths of the form XDjY. If Dj and Y are d-separated by {XZ}, then by definition of d-separation all paths between Dj and Y are blocked by {XZ}. Using the same argument as the one used in the third paragraph of the proof of Proposition 2.1, it follows that all back-door paths XDjY are blocked by Z.

2. To prove that Z is sufficient for estimating the causal effect of X on Y, we show that all back-door paths between X and Y are blocked by Z.

First, we consider the back-door paths that admit Dj as second variable. From point 1) of the corollary, we already know that these back-door paths are blocked.

Next, we divide the back-door paths that do not admit Dj as second variable into two categories: (1) the paths whose second variable is a DjZ, jj, and (2) the paths whose second variable is a DjZ. For 1), following the same argument as in the second paragraph of the proof of Proposition 2.1, we know that all back-door paths whose second variable is a DjZ are blocked.

The case where the second variable is a DjZ is more involved. Here, note that Dj is not in Z either since Z=ZDj. The fact that Z is sufficient to identify the average causal effect according to Proposition 2.1 implies that Dj and Y are d-separated by {XZ}. Therefore, every path between Dj and Y is blocked by {XZ}. For those paths that do not include Dj, it is easy to see that they are also blocked by {XZ}. For those paths that include Dj, that is, paths of the form DjDjY, we know from point 1. that they are blocked in the subpaths DjY by {XZ}. Thus, every path between Dj and Y is blocked by {XZ}, whether or not it includes Dj. Using the same arguments as the ones used in the third paragraph of the proof of Proposition 2.1, it follows that all back-door paths XDjY are blocked by {XZ}. The whole reasoning is applied for each possible Dj, according to their inclusion or exclusion in Z.

Hence, all back-door paths between X and Y in G are blocked by {XZ}. Also, because Z is sufficient to identify the average causal effect according to Proposition 2.1, Z does not include any descendants of X and therefore Z does not either. According to the back-door criterion, Z is thus sufficient to identify the average causal effect and the proof is complete.□

B General conditions for the equivalence of zero regression coefficient and conditional independence

We show that the independence of Y and Uk conditional on X and U1,...,Uk1,Uk+1,...,UM is equivalent to having regression parameter δk associated to Uk in the linear regression of Y on X and U equal to zero under less stringent assumptions than multivariate normality for the covariates X and U.

Consider the same normal linear model as in eq. (1)

Yi=δ0+βXi+m=1MδmUim+ϵi,

where ϵiiidN(0,σ2). We assume that this model is correctly specified, that is, the data for Y is generated according to eq. (1) with possibly some regression coefficients set to 0. However, we make no assumptions about the distribution of variables X and U. To simplify the notation, we denote {U1,...,Uk1,Uk+1,...UM} by UUk. We consider the case where Uk is a continuous variable. Similar arguments can be used when Uk is discrete or has a mixture distribution. Using a conditional normal distribution for Y, we have

f(y|x,u)=12πσ2exp12σ2yδ0+βx+m=1Mδmum2

and the conditional distribution of Y|X,UUk can be calculated as

(11)f(y|x,uuk)=f(uk|x,uuk)f(y|x,u)duk.

If δk=0

f(y|x,u)=12πσ2exp12σ2yδ0+βx+mkδmum2,

and the expression (11) for f(y|x,uuk) becomes

f(uk|x,uuk)12πσ2exp12σ2yδ0+βx+mkδmum2duk
=f(y|x,u)f(uk|x,uuk)duk,

which equals f(y|x,u).

Thus if δk=0 in eq. (1) then YUk|X,UUk. Also, it is obvious that if YUk|X,UUk, then δk=0. Therefore, assuming model (1) is correctly specified we have that YUk|X,UUk if and only if δk=0. Recall that no assumptions were made concerning the distribution of X and UUk.

C The behavior of ωOT8019aa4fBI"QαY

In Figure 2 we examine how the term QαY(αmY=1|αmX=1) in the definition of PB(αY) behaves as a function of the constant c, the sample size n and the standardized parameter δ˜mαYsUm/sY. Specifically, we take ω=cn, as suggested in Section 3.2, and plot the QαY(αmY=1|αmX=1) values as a function of c[0,1000] for fixed values of n (n=200,600,1000) and δ˜mαYsUm/sY (δ˜mαYsUm/sY=0.1,0.05,0.01).

Figure 2: QαYαmX=1|αmX=1$${Q_{{{\bf{\alpha}}^{\bf{Y}}}}}\left({\alpha _m^X = 1|\alpha _m^X = 1} \right)$$ with ω=cn$$\omega = c\sqrt n $$ as a function of c∈[0,1000]$$c \in [0,1000]$$ for n=200,600,1000$$n = 200,600,1000$$ and δ˜mαYsUm/sY=0.1$$\tilde \delta _m^{{\alpha ^Y}}{s_{{U_m}}}/{s_Y} = 0.1$$ (a), δ˜mαYsUm/sY=0.05$$\tilde \delta _m^{{\alpha ^Y}}{s_{{U_m}}}/{s_Y} = 0.05$$ (b) and δ˜mαYsUm/sY=0.01$$\tilde \delta _m^{{\alpha ^Y}}{s_{{U_m}}}/{s_Y} = 0.01$$ (c).
Figure 2:

QαYαmX=1|αmX=1 with ω=cn as a function of c[0,1000] for n=200,600,1000 and δ˜mαYsUm/sY=0.1 (a), δ˜mαYsUm/sY=0.05 (b) and δ˜mαYsUm/sY=0.01 (c).

In Figure 2(A), we see that, for all sample sizes considered, QαY(αY=1|αX=1) rapidly increase from 0 to the limit 1 as c goes from 0 to 1,000. This behavior is desirable since a standardized regression parameter of 0.1 is non negligible. A similar pattern is seen in Figure 2(B), although the progression of QαY(αY=1|αX=1) from 0 to 1 is slightly less rapid. In Figure 2(C), the progression of QαY(αY=1|αX=1) is much slower, especially for the smaller sample size. This behavior is desirable as well since an effect size of 0.01 would usually be considered as negligible.

D Additional simulations

We now address the secondary objectives with additional simulations. The first secondary goal is to study the large, whilst finite, sample properties of BCEE. To do this, we examine four different simulation scenarios obtained by considering the four data-generating processes (DGP1, DGP2, DGP3 and DGP4) with a sample size of 10,000. Once again, for each scenario, we randomly generated 500 datasets. We estimated the average causal effect of exposure using A-BCEE and N-BCEE with ω=cn. Because the sample size is large and the computational burden is heavy, we considered only one value of c (c=500). The results are shown in Table 9. These simulations suggest that A-BCEE and N-BCEE with ω=cn unbiasedly estimate the causal effect of exposure when n is large and when BCEE’s working assumptions hold (i.e., U includes all direct causes of X and the normal linear model is a correct specification for both X and Y).

Table 9:

Estimates of β for N-BCEE and A-BCEE with a sample size of n=10,000 for 500 Monte Carlo replicates of each data-generating process.

DGPMethodMeanSEESDEMSECP
1N-BCEE (c=500)0.1000.00680.00670.006797
1A-BCEE (c=500)0.1000.00690.00660.006697
2N-BCEE (c=500)0.1000.00740.00750.007596
2A-BCEE (c=500)0.1000.00790.00750.007598
3N-BCEE (c=500)0.1000.01400.01410.014195
3A-BCEE (c=500)0.1000.01400.01410.014195
4N-BCEE (c=500)0.0990.00830.00840.008496
4A-BCEE (c=500)0.0990.00890.00860.008697

The second secondary objective is to study the performance of B-BCEE to correct the confidence intervals of N-BCEE. Since this bootstrapped implementation is very computationally intensive, we only considered two simulation scenarios: DGP1 and DGP4 with a sample size of 200. In this case, only 100 datasets were generated for each scenario. We estimated the causal effect of exposure using the fully adjusted model, the true outcome model, BMA, BAC, TBAC, A-BCEE, N-BCEE and B-BCEE. For A-BCEE, N-BCEE and B-BCEE we took ω=cn with c=500. For B-BCEE we performed 200 bootstrap resamplings and considered an estimate with and without a bias correction. The results are presented in Table 10. We find that the non-parametric bootstrap implementation of BCEE yields correct estimates of the standard error of estimate and correct coverage probabilities. However, B-BCEE does not seem to be as efficient nor as practical as A-BCEE.

Table 10:

Comparison of estimates of β obtained from the true model, the fully adjusted model, BMA, BAC, TBAC, N-BCEE, A-BCEE and B-BCEE for the first and fourth data-generating processes (DGP1 and DGP4). Sample size is n=200, 100 datasets were generated for each data-generating process. For B-BCEE, 200 bootstrap resamplings were performed.

DGPMethodMeanSEESDEMSECP
1True model0.1050.0450.0530.05393
1Fully adjusted model0.1040.0720.0750.07594
1BMA0.1210.0480.0500.05493
1BAC (ω=)0.1040.0720.0750.07594
1TBAC (ω=)0.1040.0720.0750.07493
1N-BCEE (c=500)0.1120.0560.0630.06492
1A-BCEE (c=500)0.1110.0620.0610.06296
1B-BCEE (c=500, no bias corr.)0.1120.0670.0630.06496
1B-BCEE (c=500, w/bias corr.)0.1070.0670.0660.06696
4True model0.1110.0630.0630.06396
4Fully adjusted model0.1060.0720.0640.06496
4BMA0.1200.0600.0510.05596
4BAC (ω=)0.1050.0720.0640.06496
4TBAC (ω=)0.1060.0710.0640.06396
4N-BCEE (c=500)0.1080.0640.0610.06197
4A-BCEE (c=500)0.1090.0680.0600.06096
4B-BCEE (c=500, no bias corr.)0.1080.0710.0610.06197
4B-BCEE (c=500, w/bias corr.)0.1070.0710.0620.06298

E Marginal posterior probabilities of inclusion of potential confounding covariates

Table 11:

Marginal posterior probability of inclusion of potential confounding covariate Um, m=1,...,5, for BMA, BAC, TBAC, N-BCEE, and A-BCEE for 500 Monte Carlo replicates of the first data-generating process (DGP1). The covariates included in the true outcome model are {U3,U4,U5}.

nMethodU1U2U3U4U5
200BMA0.110.111.000.181.00
200BAC (CVm(ω))0.350.351.000.411.00
200BAC (ω=)1.001.001.001.001.00
200TBAC (ω=)1.001.001.001.001.00
200N-BCEE (c=100)0.190.241.000.371.00
200N-BCEE (c=500)0.360.411.000.541.00
200N-BCEE (c=1000)0.440.491.000.611.00
200A-BCEE (c=100)0.290.351.000.441.00
200A-BCEE (c=500)0.510.561.000.631.00
200A-BCEE (c=1000)0.600.641.000.701.00
600BMA0.060.061.000.241.00
600BAC (CVm(ω))0.320.331.000.471.00
600BAC (ω=)1.001.001.001.001.00
600TBAC (ω=)1.001.001.001.001.00
600N-BCEE (c=100)0.110.151.000.441.00
600N-BCEE (c=500)0.220.301.000.601.00
600N-BCEE (c=1000)0.280.371.000.661.00
600A-BCEE (c=100)0.150.211.000.451.00
600A-BCEE (c=500)0.340.421.000.631.00
600A-BCEE (c=1000)0.440.511.000.701.00
1,000BMA0.050.041.000.331.00
1,000BAC (CVm(ω))0.380.371.000.611.00
1,000BAC (ω=)1.001.001.001.001.00
1,000TBAC (ω=)1.001.001.001.001.00
1,000N-BCEE (c=100)0.090.111.000.551.00
1,000N-BCEE (c=500)0.190.221.000.691.00
1,000N-BCEE (c=1000)0.250.281.000.741.00
1,000A-BCEE (c=100)0.120.151.000.541.00
1,000A-BCEE (c=500)0.300.341.000.701.00
1,000A-BCEE (c=1000)0.390.441.000.751.00
Table 12:

Marginal posterior probability of inclusion of potential confounding covariate Um, m=1,...,5,7,8, for BMA, BAC, TBAC, N-BCEE, and A-BCEE for 500 Monte Carlo replicates of the second data-generating process (DGP2). The covariates included in the true outcome model are {U3,U4,U5}.

nMethodU1U2U3U4U5U7U8
200BMA0.140.120.200.180.310.100.10
200BAC (CVm(ω))0.440.410.480.180.320.120.11
200BAC (ω=)1.001.001.000.220.400.150.15
200TBAC (ω=)1.001.001.000.180.290.140.14
200N-BCEE (c=100)0.450.390.550.260.340.140.14
200N-BCEE (c=500)0.640.580.720.280.350.170.17
200N-BCEE (c=1000)0.710.660.780.290.360.180.18
200A-BCEE (c=100)0.640.560.660.190.300.130.13
200A-BCEE (c=500)0.790.730.810.190.300.140.14
200A-BCEE (c=1000)0.840.790.850.190.300.140.14
600BMA0.120.090.390.250.640.080.07
600BAC (CVm(ω))0.630.590.750.220.620.090.07
600BAC (ω=)1.001.001.000.210.570.070.06
600TBAC (ω=)1.001.001.000.190.530.090.08
600N-BCEE (c=100)0.370.260.730.280.650.090.08
600N-BCEE (c=500)0.560.450.840.290.630.110.09
600N-BCEE (c=1000)0.640.540.880.290.610.120.10
600A-BCEE (c=100)0.560.430.760.220.590.080.07
600A-BCEE (c=500)0.740.630.860.210.560.090.08
600A-BCEE (c=1000)0.800.700.900.200.560.090.08
1,000BMA0.120.080.550.330.820.060.06
1,000BAC (CVm(ω))0.690.660.860.280.750.060.06
1,000BAC (ω=)1.001.001.000.250.710.050.05
1,000TBAC (ω=)1.001.001.000.250.690.070.07
1,000N-BCEE (c=100)0.340.230.830.340.780.070.07
1,000N-BCEE (c=500)0.500.390.900.340.760.080.08
1,000N-BCEE (c=1000)0.580.470.920.340.750.080.08
1,000A-BCEE (c=100)0.500.370.840.280.750.060.07
1,000A-BCEE (c=500)0.690.570.910.270.720.070.07
1,000A-BCEE (c=1000)0.750.650.930.260.710.070.07
Table 13:

Marginal posterior probability of inclusion of potential confounding covariate Um, m=1,...,4, for BMA, BAC, TBAC, N-BCEE, and A-BCEE for 500 Monte Carlo replicates of the third data-generating process (DGP3). The covariates included in the true outcome model are {U1,U2}.

nMethodU1U2U3U4
200BMA0.170.260.080.09
200BAC (CVm(ω))0.480.290.090.10
200BAC (ω=)1.000.280.080.10
200TBAC (ω=)1.000.300.140.15
200N-BCEE (c=100)0.570.320.140.16
200N-BCEE (c=500)0.750.340.170.19
200N-BCEE (c=1000)0.800.350.180.20
200A-BCEE (c=100)0.620.300.130.14
200A-BCEE (c=500)0.770.300.130.14
200A-BCEE (c=1000)0.830.300.130.15
600BMA0.290.500.070.05
600BAC (CVm(ω))0.750.530.070.06
600BAC (ω=)1.000.520.070.05
600TBAC (ω=)1.000.510.090.08
600N-BCEE (c=100)0.680.520.100.08
600N-BCEE (c=500)0.820.530.110.10
600N-BCEE (c=1000)0.870.530.120.10
600A-BCEE (c=100)0.680.510.090.07
600A-BCEE (c=500)0.810.510.090.08
600A-BCEE (c=1000)0.850.510.090.08
1,000BMA0.410.680.050.05
1,000BAC (CVm(ω))0.860.700.060.05
1,000BAC (ω=)0.990.690.060.05
1,000TBAC (ω=)1.000.680.070.07
1,000N-BCEE (c=100)0.760.690.070.07
1,000N-BCEE (c=500)0.880.690.080.08
1,000N-BCEE (c=1000)0.910.690.090.09
1,000A-BCEE (c=100)0.750.680.070.07
1,000A-BCEE (c=500)0.850.680.070.07
1,000A-BCEE (c=1000)0.890.680.070.07
Table 14:

Marginal posterior probability of inclusion of potential confounding covariate Um, m=1,...,6, for BMA, BAC, TBAC, N-BCEE, and A-BCEE for 500 Monte Carlo replicates of the fourth data-generating process (DGP4). The covariates included in the true outcome model are {U4,U5,U6}.

nMethodU1U2U3U4U5U6
200BMA0.150.140.130.221.001.00
200BAC (CVm(ω))0.320.150.310.221.001.00
200BAC (ω=)0.970.230.980.241.001.00
200TBAC (ω=)0.870.230.990.251.001.00
200N-BCEE (c=100)0.360.170.360.291.001.00
200N-BCEE (c=500)0.530.220.570.321.001.00
200N-BCEE (c=1000)0.600.250.660.331.001.00
200A-BCEE (c=100)0.500.190.480.231.001.00
200A-BCEE (c=500)0.650.210.660.241.001.00
200A-BCEE (c=1000)0.700.210.730.241.001.00
600BMA0.140.100.080.351.001.00
600BAC (CVm(ω))0.440.160.410.291.001.00
600BAC (ω=)0.990.241.000.271.001.00
600TBAC (ω=)0.960.231.000.261.001.00
600N-BCEE (c=100)0.330.120.220.411.001.00
600N-BCEE (c=500)0.500.180.410.411.001.00
600N-BCEE (c=1000)0.580.210.500.411.001.00
600A-BCEE (c=100)0.480.160.320.311.001.00
600A-BCEE (c=500)0.670.190.530.291.001.00
600A-BCEE (c=1000)0.730.200.610.281.001.00
1,000BMA0.130.090.070.521.001.00
1,000BAC (CVm(ω))0.480.180.420.421.001.00
1,000BAC (ω=)0.990.301.000.351.001.00
1,000TBAC (ω=)0.990.301.000.341.001.00
1,000N-BCEE (c=100)0.300.120.190.571.001.00
1,000N-BCEE (c=500)0.450.190.360.561.001.00
1,000N-BCEE (c=1000)0.530.220.440.541.001.00
1,000A-BCEE (c=100)0.470.180.260.441.001.00
1,000A-BCEE (c=500)0.670.230.470.401.001.00
1,000A-BCEE (c=1000)0.730.250.560.381.001.00
Table 15:

Marginal posterior probability of inclusion of potential confounding covariate Um, m=1,...,5, for BMA, BAC, TBAC, and A-BCEE for 500 Monte Carlo replicates of the fifth data-generating process (DGP5). The true outcome model includes only U5.

nMethodU1U2U3U4U5
200BMA0.110.120.120.111.00
200BAC (CVm(ω))0.420.420.420.421.00
200BAC (ω=)1.001.001.001.001.00
200TBAC (ω=)1.001.001.001.001.00
200A-BCEE (c=100)0.210.220.220.211.00
200A-BCEE (c=500)0.450.450.460.451.00
200A-BCEE (c=1000)0.550.550.560.551.00
600BMA0.080.070.080.081.00
600BAC (CVm(ω))0.420.410.420.421.00
600BAC (ω=)1.001.001.001.001.00
600TBAC (ω=)1.001.001.001.001.00
600A-BCEE (c=100)0.120.110.120.121.00
600A-BCEE (c=500)0.300.290.310.311.00
600A-BCEE (c=1000)0.400.400.410.411.00
1,000BMA0.060.070.060.061.00
1,000BAC (CVm(ω))0.410.410.410.401.00
1,000BAC (ω=)1.001.001.001.001.00
1,000TBAC (ω=)1.001.001.001.001.00
1,000A-BCEE (c=100)0.080.090.090.081.00
1,000A-BCEE (c=500)0.220.230.230.221.00
1,000A-BCEE (c=1000)0.320.330.330.321.00

F Comparison of the distribution of β^ obtained from A-BCEE and BAC

Figure 3: Comparison of the distribution of βˆ$$\hat \beta $$ obtained from BAC (ω=∞$$\omega = \infty $$) and A-BCEE (c=100, 500, and 1,000) for all four data-generating processes and a sample size n=600$$n = 600$$. The red line corresponds to the true value β$$\beta $$=0.1.
Figure 3:

Comparison of the distribution of βˆ obtained from BAC (ω=) and A-BCEE (c=100, 500, and 1,000) for all four data-generating processes and a sample size n=600. The red line corresponds to the true value β=0.1.

Figure 4: Comparison of the distribution of βˆ$$\hat \beta $$ obtained from BAC (ω=∞$$\omega = \infty $$) and A-BCEE (c=100, 500, and 1,000) for all four data-generating processes and a sample size n=1000$$n = 1000$$. The red line corresponds to the true value β$$\beta $$=0.1.
Figure 4:

Comparison of the distribution of βˆ obtained from BAC (ω=) and A-BCEE (c=100, 500, and 1,000) for all four data-generating processes and a sample size n=1000. The red line corresponds to the true value β=0.1.

References

1. PearlJ. Causality: models, reasoning, and inference, 2nd ed. New York: Cambridge University Press, 2009.10.1017/CBO9780511803161Search in Google Scholar

2. HoetingJA, MadiganD, RafteryAE, VolinskyCT. Bayesian model averaging: a tutorial. Stat Sci1999;14:382401.Search in Google Scholar

3. RafteryAE, MadiganD, HoetingJA. Bayesian model averaging for linear regression models. J Am Stat Assoc1997;92:17991.10.1080/01621459.1997.10473615Search in Google Scholar

4. RafteryAE, ZhengY. Discussion: performance of Bayesian model averaging. J Am Stat Assoc2003;98:9318.10.1198/016214503000000891Search in Google Scholar

5. CrainiceanuCM, DominiciF, ParmigianiG. Adjustment uncertainty in effect estimation. Biometrika2008;95:63551.10.1093/biomet/asn015Search in Google Scholar

6. WangC, ParmigianiG, DominiciF. Bayesian effect estimation accounting for adjustment uncertainty. Biometrics2012;68:66171.10.1111/j.1541-0420.2011.01731.xSearch in Google Scholar PubMed

7. LefebvreG, AthertonJ, TalbotD. The effect of the prior distribution in the Bayesian adjustment for confounding algorithm. Comput Stat Data Anal2014;70:22740.10.1016/j.csda.2013.09.011Search in Google Scholar

8. VansteelandtS. Discussions. Biometrics2012;68:6758.10.1111/j.1541-0420.2011.01734.xSearch in Google Scholar PubMed

9. WangC, ParmigianiG, DominiciF. Rejoinder: Bayesian effect estimation accounting for adjustment uncertainty. Biometrics2012;68:6806.10.1111/j.1541-0420.2011.01735.xSearch in Google Scholar

10. VanderWeeleTJ, ShpitserI. A new criterion for confounder selection. Biometrics2011;67:140613.10.1111/j.1541-0420.2011.01619.xSearch in Google Scholar PubMed PubMed Central

11. BabaK, ShibataR, SibuyaM. Partial correlation and conditional correlation as measures of conditional independence. Aust NZ J Stat2004;46:65764.10.1111/j.1467-842X.2004.00360.xSearch in Google Scholar

12. HaughtonDMA. On the choice of a model to fit data from an exponential family. Ann Stat1988;16:34255.10.1214/aos/1176350709Search in Google Scholar

13. WassermanL. Bayesian model selection and model averaging. J Math Psychol2000;44:92107.10.1006/jmps.1999.1278Search in Google Scholar PubMed

14. ClydeM. Subjective and objective Bayesian statistics, 2nd ed. S. James Press, New Jersey: Wiley-Interscience, 2003Search in Google Scholar

15. YuanK-H, ChanW. Biases and standard errors of standardized regression coefficients. Psychometrika2011;76:67090.10.1007/s11336-011-9224-6Search in Google Scholar PubMed

16. AgrestiA. Categorical data analysis, 3rd ed. New Jersey: John Wiley & Sons, 2014.Search in Google Scholar

17. DawidAP. On the limiting normality of posterior distributions. Proc Cambridge Philos Soc1970;67:62533.10.1017/S0305004100045953Search in Google Scholar

18. WalkerAM. On the asymptotic behavior of posterior distributions. J R Stat Soc Ser B1969;31:808.10.1111/j.2517-6161.1969.tb00767.xSearch in Google Scholar

19. MadiganD, YorkJ, AllardD. Bayesian graphical models for discrete data. Int Stat Rev1995;63:21532.10.2307/1403615Search in Google Scholar

20. CarlinBP, GelfandAE. Approaches for empirical Bayes confidence intervals. J Am Stat Assoc1990;85:10514.10.1080/01621459.1990.10475312Search in Google Scholar

21. CarlinBP, LouisTA. Bayes and empirical Bayes methods for data analysis, 2nd ed. Boca Raton: Chapman & Hall/CRC, 2000.10.1201/9781420057669Search in Google Scholar

22. LairdNM, LouisTA. Empirical Bayes confidence intervals based on bootstrap samples. J Am Stat Assoc1987;82:73950.10.1080/01621459.1987.10478490Search in Google Scholar

23. MorganSL, WinshipC. Counterfactuals and causal inference: methods and principles for social research. New York: Cambridge University Press, 200710.1017/CBO9780511804564Search in Google Scholar

24. GuayF, LaroseS, RatelleC, SénécalC, VallerandRJ, VitaroF (2011): Mes amis, mes parents et mes professeurs: Une analyses comparée de leurs effets respectifs sur la motivation, la réussite, l‘orientation et la persévérance scolaires. Technical Report 2007-PE-118485, Fonds de recherche du Québec - Société et culture, Québec.Search in Google Scholar

25. LittleRJ, RubinDB. Statistical analysis with missing data, 2nd ed. New Jersey: John Wiley & Sons, 2002.Search in Google Scholar

26. GrahamJW, OlchowskiAE, GilreathTD. How many imputations are really needed? Some practical clarifications of multiple imputation theory. Prev Sci2007;8:20613.10.1007/s11121-007-0070-9Search in Google Scholar PubMed

27. TalbotD, LefebvreG, AthertonJ (2014): BACprior: Choice of the Hyperparameter Omega in the Bayesian Adjustment for Confounding (BAC) Algorithm. Available at: http://CRAN.R-project.org/package=BACprior, R package version 2.0.Search in Google Scholar

28. GreenlandS, RobinsJM, PearlJ. Confounding and collapsibility in causal inference. Stat Sci1999;14:2946.10.1214/ss/1009211805Search in Google Scholar

Published Online: 2015-7-2
Published in Print: 2015-9-1

©2015 by De Gruyter

Downloaded on 27.5.2024 from https://www.degruyter.com/document/doi/10.1515/jci-2014-0035/html
Scroll to top button