Home Estimating Average Treatment Effects Utilizing Fractional Imputation when Confounders are Subject to Missingness
Article Open Access

Estimating Average Treatment Effects Utilizing Fractional Imputation when Confounders are Subject to Missingness

  • Nathan Corder EMAIL logo and Shu Yang
Published/Copyright: December 31, 2020
Become an author with De Gruyter Brill

Abstract

The problem of missingness in observational data is ubiquitous. When the confounders are missing at random, multiple imputation is commonly used; however, the method requires congeniality conditions for valid inferences, which may not be satisfied when estimating average causal treatment effects. Alternatively, fractional imputation, proposed by Kim 2011, has been implemented to handling missing values in regression context. In this article, we develop fractional imputation methods for estimating the average treatment effects with confounders missing at random. We show that the fractional imputation estimator of the average treatment effect is asymptotically normal, which permits a consistent variance estimate. Via simulation study, we compare fractional imputation’s accuracy and precision with that of multiple imputation.

MSC 2010: 62F12; 62F30

1 Introduction

It is commonplace in scientific research for investigators to rely on observational data to address questions of interest. While randomized experiments are the gold standard for drawing causal inferences about the effect of a treatment (also known as exposure, intervention, regime, or policy), in many cases, randomized experiments are difficult or infeasible to implement for logistical, financial or ethical reasons. For example, it would be unethical to force people to smoke to study the causal effect of smoking on health outcomes. Instead, researchers must utilize observational data and make careful corrections to address various biases. Undeniably, it is considerably more difficult to draw correct causal conclusions from observational data than from a randomized experiment. The main reason is due to confounding induced by non-randomization of treatments. Usually, researchers make unverifiable assumptions to draw causal conclusions of treatment effects, such as unconfoundedness of the treatment-outcome relationship, after adjusting for a set of confounders. Current causal inference methods, including propensity score methods [1], outcome regression methods, and doubly robust methods [2, 3, 4, 5], have been developed to remove confounding bias, mainly in the settings where confounders are fully observed. However, observational data is also highly prone to missingness. Thus it is important, and at many times critical, to handle missing data properly to avoid introducing additional bias to the data analysis.

1.1 Missing Data

Despite the best intentions of researchers, missing data is nearly impossible to avoid in real-world settings. Fortunately, as prevalent as missingness is, so too are methods with which to address missingness; however, the type of missingness matters when selecting a method.

Missing data occurs by way of one of three mechanisms in observational data. The first and simplest type is Missingness Completely at Random (MCAR) [6]. In this setting, whether or not an observation is missing is independent of both the observed and missing data. The second is Missingness at Random (MAR). Here, whether or not an observation is missing depends only on the observable data but not the missing data. Lastly, missing data can be Missing Not at Random (MNAR). In this setting, even after conditioning on all observed data, missingness will still depend on the missing portion of the data.

Confirming which missingness type is present for a given data set can not be validated from only the observed data. Little [7] proposed a chi-square based test capable of checking if the MCAR assumption was violated, and more recently Mohan and Pearl have been able to use m-graphs to refute instances where the MAR and MNAR assumptions were proposed but shown not to hold [8, 9]. Despite not being able to validate if a particular missingness model is true, it is still common for arguments towards one of these assumptions to be made, relying on the knowledge of subject matter experts for the data at hand. If the observed data is sufficient to explain the missingness, the MAR assumption is plausible, and it is this assumption we carry forward for the majority of our discussion. In Section 6 we will discuss the extension to MNAR.

1.2 Approaches to Address Missing Data

Many methods have been proposed to address missingness when assuming a MAR pattern. Two of the most common approaches in practice are complete case estimation (CC) and multiple imputation (MI). In particular, MI was favored by the National Research Council in 2010 as one of its preferred means of addressing missing data in clinical trials [10]. Under CC, all records with missing data are excluded, and treatment effects are estimated only on fully observed cases. CC can be biased under MAR; more importantly, if multiple variables have missing values, there will only be a small portion of complete cases in the data. By throwing out a large portion of the data, the effective sample size shrinks, inflating variances. Therefore, CC suffers from loss of efficiency by utilizing less of the observed data in the final analysis [11].

MI, on the other hand, is traditionally recommended for MAR in part due to its improved efficiency over CC and due to it being applicable to a wider range of missingness mechanisms. With MI, the full joint distribution of the data is estimated (either empirically or modeled based on distributional assumptions), and from this, a series of M new imputed data sets are drawn. In each imputed data set, MI fills in each missing value with an imputed value by sampling from the posterior predictive distribution of the missing value given the observed values. Then, full sample analyses can be applied to each imputed data sets, and these multiple results are summarized by an easy-to-implement combining rule for inference [12].

MI has produced valid frequentist inference in a wide range of applications [13]. At the same time, Rubin’s variance estimator for MI has been shown to not always be consistent [14, 15, 16, 17, 18, 19, 20, 21, 22]. For inference using MI consistent, imputation must be proper (see [12] for a precise definition). Practically speaking, for any proper imputations, under sufficiently regular models, Rubin’s combining rule provides a consistent estimator of the parameter of interest and a weekly unbiased estimator of its asymptotic variance. When the imputation model is correctly specified and the MAR assumption holds, Meng [23] showed that a sufficient condition for imputation to be proper is that the imputation model and the analysis model are congenial. For example, the imputation model is correctly specified and the analysis is efficient under the same model. Congeniality is sometimes more elusive than it would appear. Even when the imputation model is correctly specified, Yang and Kim [22] showed that MI is not necessarily congenial for method of moments estimation. Therefore, some common statistical procedures may be incompatible with MI. From a causal inference perspective, this poses a problem as the validity of Rubin’s variance estimator has not been fully explored for many full sample estimation methods used widely in causal inference. Certain otherwise unbiased and consistent full sample causal inference methods (outcome regression, weighting, matching, etc.) may lose these properties when applied in conjunction with MI and MI-produced data sets. Many of the most common estimates for average treatment effects are based on method of moments estimators and are thusly susceptible to inaccurate variance estimates when using Rubin’s variance estimator for MI. For researchers desiring to make causal claims when utilizing MI, it is imperative for the variance properties of their estimators to, therefore, be either validated or an alternative method must be proposed.

1.3 Fractional Imputation as an Alternative

As an alternative to CC and MI, there are likelihood-based methods that can be applied. When using these methods, the key insight is that under fully observed confounders, the full sample estimators are obtained by solving estimating equations. In the presence of partially observed confounders, the corresponding estimators can be obtained by solving conditional estimating equations which integrate out the missing confounders given the observed data. There are two difficulties in this approach. First, it requires consistent estimators in the conditional distribution of the missing confounders given the observed data, such as MLE. In the presence of missing values, an EM algorithm is typically used. Second, numerical integration is needed. Integration approximated by imputation was considered by many authors, such as Monte Carlo EM methods [24]. For Monte Carlo EM algorithms, in each E-step, the imputed values are regenerated, and thus the computation can be quite heavy. Also, the convergence of Monte Carlo sequence of the estimators is not guaranteed for fixed Monte Carlo sample size [25].

In practice, EM algorithms may not be feasible when the conditional expectation in the E-step is not available in a closed form. Instead, fractional imputation (FI) has been proposed to serve as a computational tool for implementing the expectation step (E-step) in the EM algorithm [24, 26, 27], which simplifies computation by drawing on importance sampling to obtain the fractional weights and reducing the iterative computation burden over other simulation methods such as Markov Chain Monte Carlo. See [28, 29, 30, 31] for applications of FI outside of the causal inference context. The main idea in FI is to produce a complete data set by imputation where each imputed value is associated with a fractional weight, by which the observed likelihood can be approximated by the weighted average of the imputed data likelihood. The resulting estimator approximates the maximum likelihood estimator.

For illustrative purposes, we can compare the format of the imputed data via FI to the more commonly encountered imputed data via MI. Suppose we have a data set where some of the records are subject to missingness. MI will attempt to address this missingness by creating M data sets where each imputed value is imputed exactly once in each of the M data sets as depicted in Figure 1. In this image X1 is a fully observed covariate, and A and Y are fully observed treatment indicator and response variables respectively. X2 is a covariate subject to missingness. R1 and R2 are missingness indicators for X1 and X2 where Rij = 0 indicates Xj is missing for record i. Records with missingness are indicated in red on the left. The completed data sets (with X2 imputed) are on the right with the records now including imputed values for X2 highlighted in green.

Figure 1 Imputed Data via MI
Figure 1

Imputed Data via MI

FI instead seeks to generate a single imputed data set, but one where each imputed record also includes a fractional weight. That fractional weight is indicative of how likely the imputed data is to occur under the distribution of the completed data set. Figure 2 shares all the same features as Figure 1 except now includes an additional column in the imputed data for the fractional weight (ω) that will be incorporated into analyses.

Figure 2 Imputed Data via FI
Figure 2

Imputed Data via FI

At its most basic level, the difference in imputation approach is analogous to a debate between a wide vs a tall data set. It could even be argued that when recombining data sets imputed via MI an implicit "weight" of 1/M is assigned to every record. This implicit 1/M weight acts similarly to the fractional weights ω generated via FI. The important distinction between MI and FI lies in how the final "weights" assigned to each imputed record are produced. The weight generation process for FI will be discussed in more depth in Section 3. For the time being, we will assume these weights are generated appropriately and can be incorporated easily into further analyses.

FI is sufficiently versatile that, like MI, it can be deployed along-side both continuous and categorical variables. Under the same common regularity conditions as seen often with MI, we show that the FI estimators are consistent and asymptotically normal. The remainder of this paper focuses on expanding FI into the causal literature by developing an FI-based method for estimating causal treatment effects. Once developed we will validate the method relative to existing causal inference methods. Specifically, we investigate the comparative performance of FI vs MI and CC for estimating causal treatment effects when confounders are subject to missingness.

In summary, the proposed FI framework achieves desirable properties for causal inference:

  1. the same fractionally imputed data set allows for applying general full-sample estimators (which solve certain estimating equations) of the average treatment effect, including regression estimators, inverse probability of weighting estimators, and augmented weighting estimators;

  2. the FI estimators are asymptotically linear and therefore allow resampling methods for variance estimation and inference, which is simple to implement in practice;

  3. the unified FI inference has theoretical guarantees and offers a solution to the uncongeniality issue of MI;

  4. and lastly, FI is not only statistically efficient but also computationally efficient compared to MI, as demonstrated via simulation and real-world application.

The rest of the article is organized as follows. We begin in Section 2 with a description of notation and assumptions. In Section 3 we fully outline the FI process as well as derive the resulting variance estimator for the treatment effect estimator. We implement a simulation study in Section 4 comparing the accuracy and precision of treatment effect estimators when the missingness is addressed by CC, MI, and FI. Section 5 provides a demonstration of FI’s utilization with a real-world health data set. Finally, in Section 6, we end with a discussion of the results and of the implication they have on current and future causal work.

2 Setup and Notation

2.1 Treatment Effect Estimation Notation

Following Neyman [32] and Rubin [33] we use the potential outcomes framework. Treatment is denoted by Ai ∈ {0, 1}, where 0 and 1 are labels for control and treatment respectively. For each subject i, define a pair of potential outcomes {Yi(1), Yi(0)} which represent the outcomes if the subject was treated, Yi(1), and if he or she was not, Yi(0). Implicit in this notation, we make the stable unit treatment value assumption [34]. The observed outcome for subject i is then Yi = Yi(Ai). Let Xi be the vector of confounders for subject i. We assume that {Xi , Ai, Yi(1), Yi(0)}i=1nare independent draws from the distribution {X, A, Y(1), Y(0)}, and therefore, {(Xi,Ai,Yi)}i=1nare independent and identically distributed. The conditional treatment effect is τ(X) = E{Y(1) − Y(0) | X}, and the average treatment effect is τ0 = E{τ(X)}. The average treatment effect cannot be estimated without further assumptions, because for each subject only one potential outcome is observed. The common assumptions for identifying the average treatment effect [1] are as follows:

Assumption 1 (Ignorability). Y(a)⊥A | X for a = 0,1, wheremeans "is conditionally independent of".

Assumption 2 (Sufficient overlap). With probability 1, 0 < c1e(X) ≤ c2 < 1, where e(X) = pr(A = 1 | X) is the propensity score.

Under Assumption 1, adjusting for covariates X creates a randomization-like scenario and removes confounding biases brought on by treatment selection (i.e., for each level of X, the treatment assignment is as good as randomization). However, in practice, there are often many variables in X, some of which are continuous; therefore, directly conditioning on each level of X is difficult. Alternatively, the propensity score has been proposed as a one-dimensional summary of X [1]. The central role of the propensity score lies in the fact that Assumption 1 implies Y(a)⊥A | e(X) for a = 0,1. Therefore, adjusting for the propensity score alone can remove confounding biases.

2.2 Treatment Effect Estimation Under Fully Observed Data

Reliance on the propensity score comes about naturally when we decompose the joint density of X, A, and Y into three particular components. Specifically

(1)f(X,A,Y)=f(X)f(AX)f(YX,A)=f(X)e(X)f(YX,A)

Based on this decomposition, a number of propensity score-based estimators have been proposed for estimating the treatment effect including propensity score matching, subclassification, or weighting. See [35] for a textbook discussion. If we limit the class of propensity estimators to only parametric estimates (of the form e(X | θ) where θ is the vector of parameters used to estimate the propensity score), the two most common estimates of τ from this class are that of Inverse Propensity Weighting (IPW) and Augmented IPW (AIPW). Both are illustrated here as examples. In both examples, let e(X | θ^)be the estimated propensity score where θ has been estimated by some consistent estimator θ^.In practice, e(X | θ^)is typically a logistic regression model.

Example 1

(IPW estimator). The IPW estimator of τ0is

(2)τ^IPW=n1ni=1AiYieXiθ^1AiYi1eXiθ^.

Example 2

(AIPW estimator). Letμ(X,aβ^)be an unbiased estimator of E(Y | X, A = a; β), then the AIPW estimator of τ0is

(3)τ^AIPW=n1i=1nAiYieXiθ^+1AieXiθ^μXi,1β^1AiYi1eXiθ^+11Ai1eXiθ^μXi,0β^.

We adopt the estimating equations convention where we let U(τ; X, A, Y | η) be the estimating function for τ0 under a given set of nuisance parameters η. An unbiased estimate for τ0 can then be derived as the solution to PnU(τ; X, A, Y | η^)= 0 where η^is a consistent estimator of η and Pn is the empirical measure; namely Pnf(X)=n1i=1nf(Xi).As examples, estimating functions for IPW and AIPW are shown below.

Example 3

(IPW estimating function). The estimating function ofτ^IPWis

UIPW(τ;X,A,Yη^)=AYe(Xθ^)(1A)Y1e(Xθ^)τ.

Hereη^=(θ^),the parameter estimates used to calculate the propensity scores.

Example 4

(AIPW estimating function). The estimating function forτ^AIPWis

UAIPW(τ;X,A,Yη^)=[AYe(Xθ^)+{1Ae(Xθ^)}μ(X,1β^)][(1A)Y1e(Xθ^)+{11A1e(Xθ^)}μ(X,0β^)]τ.

Hereη^=(β^,θ^),the parameter estimates used to calculateμ(X,aβ^)and the propensity scores respectively.

Remark 1

(Doubly Robust Estimation). The IPW estimate in Example 3 does not need to model the outcome Y, but it does require a correct model for e(X | θ). On the other hand, the AIPW estimate for τ0obtained from the mean estimating function in Example 4, incorporates a double robust (DR) feature for estimation. That is, UAIPWis an unbiased estimating function for τ0if either e(X | θ) or μ(a, X | β) is correctly specified.

2.3 Missing Data Notation

Whereas all of the above discussion holds under fully observed responses and confounders, in this article we consider the case where X contains missing values. To that end, let R be a collection of indicator variables R = (R1, ..., Rp) corresponding to (X1, ..., Xp) where Rij = 1 indicates that Xj is observed for subject i and Rij = 0 indicates Xj is not observed for subject i. Let R denote the collection of all possible missingness patterns. Let Xobs,i, the observed part of covariates for individual i, consist of Xij with Rij = 1. Similarly let Xmis,i, the missing part of covariates for individual i, consist of Xij with Rij = 0. Note that the dimension of Xobs,i and Xmis,i will vary across individuals. For notational simplicity, we suppress the subscript i for subject.

If we return to equation (1) and incorporate the parameterizations laid out in examples 1 and 2, the decomposition of the joint distribution can be naturally extended to account for missingness. To do so, we rewrite the joint distribution as:

(4)f(X,A,Y,R;α,β,θ,ρ)=f(X)e(Xθ)f(YX,A;β)f(RX,A,Y;ρ)=f(Xobs,Xmis)e(Xθ)f(YX,A;β)f(RX,A,Y;ρ)=f(Xobs)f(XmisXobs;α)e(Xθ)f(YX,A;β)f(RX,A,Y;ρ)

where α is the collection of parameters used to estimate the missing portion of X given the observed portion of X, and ρ is the collection of parameters used in describing the missingness mechanism.

Finally, in this article we are only interested in the case where Xmis follows a MAR pattern, which leads us to our final assumption which completes the basis for Theorem 1 (the proof for which is made available in the appendix).

Assumption 3 (Missing at random). RXmis Xobs ,A,Y

Here we adopt the notion of Rubin’s MAR in the sense that MAR may hold for different variables depending on the missingness pattern. The scientific justification of this assumption may be difficult; however, theoretically it is the weakest condition under which the missingness process can be ignored [6]. Alternatively, one can consider the variable-based taxonomy of MAR [9], where Xmis represents variables that are subject to missingness and Xobs represents variables that are fully observed. Our method development is similar under this notion of MAR but using the new definition of Xmis and Xobs in equation (4).

Theorem 1

Under Assumptions 1–3, τ0is identifiable from the observed data.

A convenient consequence of adopting an MAR framework is that because of Assumption 3, any expectations taken with respect to Xmis will result in f (R|X, A, Y; ρ) falling out of the decomposition; therefore for the remainder of this article, unless otherwise noted, we will be suppressing inclusion of an explicit missingness mechanism term and instead using the decomposition:

(5)f(X,A,Y;α,β,θ)=f(Xobs )f(Xmis Xobs ;α)e(Xθ)f(YX,A;β)

2.4 Treatment Effect Estimation Under MAR

In the presence of missingness, we can let U¯(τ;Xobs,A,Yη^)=E{U(τ;X,A,Yη^)Xobs,A,Y}.We term U¯(τ;Xobs,A,Yη^)the "mean estimating function" of τ0 given the observed data. Note that in defining U¯(τ;Xobs,A,Yη^),η^must now also expand to include any new parameters α^utilized in estimating E(Xmis | Xobs, A, Y; α). From Theorem 1, under Assumptions 1–3, U¯(τ;Xobs,A,Yη^)is an unbiased estimating function of τ0. Therefore, a consistent estimator of τ0 can be obtained by solving PnU¯(τ;Xobs,A,Yη^)=0.

Remark 2

(Doubly Robust Estimation Under MAR). As with IPW estimation under fully observed data, IPW estimation under MAR does not need to model the outcome Y, but it does now require correct models for both e(X | θ) and f (Xmis | Xobs, A, Y; α). On the other hand, the AIPW estimate for τ0will still be unbiased provided f (Xmis | Xobs, A, Y; α) is correctly specified and either e(X | θ) or μ(a, X | β) is correctly specified. The DR feature of AIPW estimation for treatment effects has been shown extensively in full data situations and more recently in the case where exposures are MAR [36].

3 Fractional Imputation

Estimation of treatment effects under fully observed data is straightforward; unfortunately, fully observed data is rarely encountered in practice. Imputation methods are often used to facilitate estimation in the presence of missing values by completing the partially observed portions of the data and coaxing a "full" data set out of a partial one. It is important to note, though, imputation methods only provide a means of addressing the missingness and complete the data set without heed given to effect estimation. Therefore, to examine the consequences of choosing a specific imputation method, we propose a two-stage procedure. In the first stage, referred to as the design stage, we use an imputation technique to fill in the missing covariate values and estimate the propensity scores. In the second stage, referred to as the analysis stage, classical propensity score techniques are applied to estimate the causal parameters. See [37, 38], and [39] for the mention of decomposing causal inference into two different stages.

This framework will be used for both FI and MI methods. The former of these methods we discuss in more detail here. For a more detailed examination of MI see [12].

3.1 Implementing Fractional Imputation

Earlier in this paper, Figure 2 depicted what a completed data set might look like after using FI to impute missing data during the design stage. Besides the shape of the data, the addition of an explicit weight to each record of the imputed data set stands out from more commonly encountered MI data sets. To generate fractional weights, FI deploys a three-step process. First, missing values are imputed from some proposed distribution, then fractional weights are updated, then model parameters for the full joint distribution are re-estimated (now under updated weights). The re-weighting and model updating steps cycle until convergence.

In the first step, sometimes referred to as the imputation step or the I-step [27, 40], every missing value is imputed M times by means of some proposal function h(Xmis | Xobs). This generates M new values Xmis,i(j)for each partially observed record. The choice of h() is arbitrary and left to the imputer, but a convenient choice is f (Xmis | Xobs; α) to align with the decomposition from equation (5). This choice would necessitate being able to provide or estimate a value for α (for instance α = α^0where α^0is the MLE for α calculated only using complete cases). At this first step, every record where a value was imputed will have a fractional weight of ωij=1/M.Note, at the conclusion of every step in FI, fractional weights for each individual, observation, etc. i are held to the conditionj=1Mωij=1.

In the second step, the weighting step or W-step, the fractional weights are updated proportional to the likelihood of the imputed value under the full joint distribution, divided by the likelihood of the imputed value under the proposal distribution h() from the I-step. If choosing the decomposition from equation (5) that would appear as

ωij(t)f(Xobs,i,Xmis,i(j),Ai,Yi;η^(t))h(Xmis,i(j)),

given the current parameter estimates for η^(t)=(α^(t),β^(t),θ^(t)).

In the third step, the maximization step or M-step, the parameter values used to estimate the full joint distribution are updated given the new values of ωij from η^(t) to η^(t+1).The W-step and M-step iterate, setting t=t+1 after each M-step, until the model parameters converge. The resulting data set with the final fractional weights included can then be passed on to the analysis stage as if it were a complete data set (similar to MI passing on M data sets). An example generating a complete data set via FI is included in the appendix.

3.2 Characterizing Fractional Imputation for Estimating Treatment Effects

For illustration, consider the case where X contains only two variables, X = (X1, X2),where X1 is fully observed and X2 is subject to missingness. Let R2 be the response indicator of X2. From Examples 3 and 4, we can obtain an estimator of τ0 by solving the mean estimating equation

(6)i=1nU¯τ;Xobs,i,Ai,Yiη^=i=1nEUτ;Xi,Ai,Yiη^Xobs,i,Ai,Yi=i=1nR2iUτ;Xi,Ai,Yiη^+i=1n1R2iEUτ;Xi,Ai,Yiη^X1i,Ai,Yi=0.

where U(τ;Xi,Ai,Yiη^)denotes either UIPW(τ;Xi,Ai,Yiη^) or UAIPW(τ;Xi,Ai,Yiη^).

In (6), the conditional expectation E{U(τ;Xi,Ai,Yiη^)Xobs,i,Ai,Yi}is often difficult to obtain. The basic idea of FI is to overcome this difficulty by creating a weighted set {(ωij,Xi(j),Ai,Yi):j=1,,M}such that E{U(τ;Xi,Ai,Yiη^)Xobs,i,Ai,Yi}can be approximated byi=1nj=1MωijU(τ;Xi(j),Ai,Yiη^).

Remark 3

Only records where R2i = 0 need imputed, so when utilizing FI, only these observations require a weightωijto be calculated in the weighted set{(ωij,Xi(j),Ai,Yi):j = 1, ..., M}. However, implicit in this representation is the generation of weights ωi = 1 for observations where R2i = 1. While notation for such implicit generation is suppressed in this article, if desired, the weighted set can be rewritten as{(ωij,Xi(j),Ai,Yi):i=1, ..., n; j = 1, ..., mi} where

mi={M  if R2i=01  if R2i=1,andXi(j)={(Xobs,i,xi(j)) if R2i=0(Xobs,i) if R2i=1.

Such notation may be beneficial if equational symmetry is desired, though the FI process and resulting estimates of τ0are unaffected.

Toward that goal of approximating E{U(τ; Xi , Ai , Yi | η^)| Xobs,i , Ai , Yi} as: i=1nj=1MωijU(τ;Xi(j),Ai,Yiη^),notice that the last conditional expectation in (6) can be written as

E{U(τ;Xi,Ai,Yiη^)X1i,Ai,Yi}=U(τ;X1i,x2,Ai,Yiη^)f(X1i,x2,Ai,Yiη^)dx2f(X1i,x2,Ai,Yiη^)dx2,

where f (X1, X2, A, Y | η^)is the joint distribution of (X1, X2, A, Y) with nuisance parameters η set to η^. Furthermore, the joint distribution can be decomposed similar as in (5) to be

(7)f(X1,X2,A,Yη^)=f(X1,X2α^)f(AX1,X2;θ^)f(YX1,X2,A;β^)=f(X1)f(X2X1;α^)e(Xθ^)f(YX1,X2,A;β^),

where we assume f (X2 | X1; α^)and f (Y | X1, X2, A;β^)be correctly specified as f (X2 | X1; α) and f (Y | X1, X2, A; β), respectively.

Under complete response, the maximum likelihood estimator (MLE) of θ can be obtained as a solution to the score equations,

ni=1Sθ;Xi,Ai=0,

where S(θ; X, A) is the score function of θ and can be written as S(θ; X, A) = log f (A | X; θ)/∂θ with f (A | X; θ) = e(X | θ)A{1 − e(X | θ)}1−A [41, 42], which under missingness is rewritten

(8)ni=1R2iSθ;Xi,Ai+ni=11R2iESθ;Xi,AiX1i,Ai,Yi=0.

MLE estimates of α^ and β^can be obtained similarly to (8) for their respective mean score equations. For α^the mean score equation is

(9)ni=1Sα;Xi=ni=1Sα;X1i,X2i=0,

where S(α; X) is the score equation for α written as S(α; X) = S(α; X1, X2) = log f (X2 | X1; α)/∂α. Under MAR, α^can be obtained using only complete cases.

For β^the mean score equation is

ni=1Sβ;Xi,Ai,Yi=0,

where S(β; X, A, Y) is the score function of β and can be written as S(β; X, A, Y) = log f (Y | X, A; β)/∂β, which under missingness is rewritten

(10)ni=1R2iSβ;Xi,Ai,Yi+ni=11R2iESβ;Xi,Ai,YiX1i,Ai,Yi=0.

To obtain the solution to (6), (8), (9) and (10), the EM algorithm can be applied. To do so using FI, the following process can be implemented:

Step 0. Let the initial values for parameters be set to α(0), β(0), and θ(0) which are the MLE of α, β, and θ using only complete cases. For each unit i with R2i = 0, generate M imputed values of X2i, denoted by x2i(j)(j = 1, . . . , M), from a proposal distribution h(x2), e.g. f (x2 | X1; α(0)).

Step 1. At the tth EM iteration, compute the fractional weight

ωij(t)f(x2i(j)X1i;α(t))e{(X1i,x2i(j))θ(t)}f(YiAi,X1i,x2i(j);β(t))h(x2i(j))

subject toj=1Mωij(t)=1.

Step 2. Use ωij(t)and (X1i,x2i(j),Ai,Yi)to update the parameters from (α(t), β(t), θ(t)) to (α(t+1), β(t+1), θ(t+1)) by solving the respective imputed score equations. To update α(t) to α(t+1), solve the imputed score equation

ni=1R2iSα;X1i,X2i+ni=11R2iMj=1ωij(t)Sα;X1i,x2i(j)=0.

To update β(t) to β(t+1), solve the imputed score equation

ni=1R2iSβ;X1i,X2i,Ai,Yi+ni=11R2iMj=1ωij(t)Sβ;X1i,x2i(j),Ai,Yi=0.

To update θ(t) to θ(t+1),solve the imputed score equation

ni=1R2iSθ;Xi,Ai+ni=11R2iMj=1ωij(t)Sθ;X1i,x2i(j),Ai=0.

Step 3. Set t = t + 1 and go to Step 1. Continue until convergence.

Remark 4

Recall under MARα^can be obtained under only complete cases. In such case, α(0)=α^,and there is no need to update the parameter estimate α(t)each iteration. Additionally, ifh(x2)=f(x2X1;α^),the calculation of the weight function simplifies toωij(t)e{(X1i,x2i(j))Tθ(t)}f(YiX1i,x2i(j),Ai;β(t))andj=1Mωij(t)=1.The simplification is not necessary under MAR, but we mention it here for the event when additional computational resource efficiencies are desired for a particular application.

Let η^=(α^,β^,θ^)be the resulting estimates for the nuisance parameters. Note that at each EM iteration, imputed values of X2 are not changed; only fractional weights are updated for each iteration. The weights ωij,obtained at the end of the EM iteration, assigned to imputed values can be called fractional weights. The fractional weight represents a similarity measure between the imputed value and the missing value.

By incorporating these weights, the conditional estimating equation for τO can be approximated by

ni=1R2iUτ;Xi,Ai,Yiη^+ni=11R2iMj=1ωijUτ;X1i,x2i(j),Ai,Yiη^=0,

and τ^can be obtained by solving this imputed estimating equation for τ. Here, U(τ;X,A,Yη^)can be either the IPW or AIPW estimating function.

3.3 Asymptotic Results

Because τ^is obtained through the method of estimating equations, we establish the asymptotic properties of τ^,in a manner similar to [20].

Theorem 2

Let η = (α, β, θ) be a vector of the nuisance parameters, and letη^=(α^,β^,θ^)be the vector of corresponding MLE estimators converging in probability to η0 = (α0, β0, θ0), the true values of the nuisance parameters. Under certain regularity conditions, the solutions to (6), τ^,is consistent for τ0and satisfies

n(τ^τ0)N(0,V),

as n → ∞, where

V=λ1Ωλ1T,λ=E{τU(τ0;X,A,Y)},Ω=Var{U¯(τ0;Xobs,A,Y,η0)+κSobs(η0;Xobs,A,Y)},U¯(τ;Xobs,A,Y,η)=E{U(τ;X,A,Y)Xobs,A,Y,η},κ=E{U(τ0;X,A,Y)SmisT(η0;Xobs,A,Y)}Jobs1,Smis (η;Xobs,A,Y)=E{ηlogf(XmisXobs,A,Y;η)Xobs,A,Y;η},Sobs(η;Xobs,A,Y)=E{S(η;X,A,Y)Xobs,A,Y;η},S(η;X,A,Y)=ηlogf(X,A,Y;η),

and

Jobs=EηTSobsη0;Xobs,A,Y=EηTSη0;X,A,Y+ESmisη0;Xobs,A,Y2

whereB2BBTfor some matrix B.

Proof. LetU¯(τη)n1i=1nE{U(τ;Xi,Ai,Yi)Xobs,i,Ai,Yi,η}.Note thatτ^andη^satisfyU¯(τ^η^)=0as

M → ∞. By use of Taylor expansions we can study the asymptotic properties ofτ^.

First, by a Taylor expansion ofU¯(τη^)aboutη^=η0we obtain

(11)U¯(τη^)=U¯(τη0)+E{ηTU¯(τη0)}(η^η0)+op(n1).

BecauseU¯(τη)=n1U(τ;Xi,Ai,Yi)f(Xmis,iXobs,i,Ai,Yi;η)dXmiswe obtain

(12)EηTU¯(τη)=En1i=1nUτ;Xi,Ai,YifXmis,iXobs,i,Ai,Yi;ηηdXmis=En1i=1nUτ;Xi,Ai,YilogfXmis,iXobs,i,Ai,Yi;ηηfXmis,iXobs,i,Ai,Yi;ηdXmis=EU(τ;X,A,Y)SmisTη;Xobs,A,Y.

To expressη^η0from (11) further, we note that the EM algorithm leads to the MLE of η0and thereforeη^satisfies

(13)n1ni=1ESη^;Xi,Ai,YiXobs,i,Ai,Yi;η^=0,

which depends onη^in two places, namelyS(η^;Xi,Ai,Yi)and the conditional expectation taken with respect tof(Xmis Xobs ,A,Y;η^).Applying another Taylor expansion aboutη^=η0this time in (13) leads to

(14)0=n1i=1nESη0;Xi,Ai,YiXobs,i,Ai,Yi;η0+n1i=1nEηTSη0;Xi,Ai,YiXobs,i,Ai,Yi;η0+ESη0;Xi,Ai,YiSmisTη0,Xi,Ai,YiXobs,i,Ai,Yi;η0+Opn1/2.

Therefore we can express

(15)η^η0Jobs1n1ni=1Sobsη0;Xobs,i,Ai,Yi.

Combining (12) and (15), ignoring the small order terms, (11) can be expressed in a linear form:

Ul¯τη0=n1i=1nU¯τ;Xobs,i,Ai,Yi;η0+EU(τ;X,A,Y)SmisTη;Xobs,i,Ai,YiJobs1Sobsη0;Xobs,i,Ai,Yi]=n1i=1nU¯τXobs,i,Ai,Yi;η0+κSobsη0;Xobs,i,Ai,Yi.

with the l being used to denote the linearization.

Second, note that we now haveU¯(τ^η^)=Ul¯(τ^η0)+op(n1/2).We apply another Taylor expansion, this time onU¯l(τ^η0)aboutτ^=τ0,and we obtain

τ^τ0=E{τTUl¯(τ0η0)}1Ul¯(τ0η0)+op(n1/2).

Because E {Sobs(η0; Xobs, A, Y)} = 0, the first term simplifies as

E{τTUl¯(τ0η0)}=E{τTU¯(τ0;Xobs,A,Y,η0)}=E{τTU¯(τ0;X,A,Y)}=λ.

Lastly, if we combine all the results above, we obtain an asymptotic linearization ofτ^τ0as

(16)τ^τ0=λ1Ul¯τ0η0=λ1n1i=1nU¯τ0;Xobs,i,Ai,Yi,η0+κSobsη0;Xobs,i,Ai,Yi+opn1/2

with κ and λ defined as in the above theorem.

Therefore, the asymptotic variance ofn(τ^τ0)is

λ1Var{U¯(τ0;Xobs,A,Y,η0)+κSobs(η0;Xobs,A,Y)}λ1,

which completes the proof.

As a result of Theorem 2, we can obtain a consistent variance estimator of τ^. Define λ^ and κ^as empirical versions of λ and κ, respectively. For example,λ^=n1i=1nτUτ^;Xobs,i,Ai,Yi,η^.Next define q^i=U¯(τ^;Xobs,i,Ai,Yi,η^)+κ^Sobs(η^;Xobs,i,Ai,Yi).Then we can estimate the variance of τ^using a sandwich formula

Var(τ^)=λ^1n11n1ni=1q^iq^¯2λ^1,

where q^¯=n1i=1nq^i.

3.4 Variance Estimation

Because of how data is imputed under FI, we can obtain further simplification when using IPW or AIPW. The λ terms cancels since τU(τ0;X,A,Y)=1under either estimator, and it can be shown that the κSobs(η0) term falls out of Ω for observations where X2 is observed. However, even after these mild simplifications, it is still apparent that exact variance estimates will be difficult to obtain. Only in rare situations will the derivatives of the score functions be anything other than impractical to calculate. Theorem 2 suggests large sample approximation is appropriate, and a bootstrap or jackknife estimator can serve as a more practical alternative.

4 Simulation Study

In the current causal inference literature, there has not been a side-by-side comparison of FI and MI with respect to how they perform estimating average treatment effects or of their corresponding variance estimates in the same setting. To examine how FI performs compared to MI, we adapt a simulation study framework used by Lunceford and Davidian [5]. We modify this setting for our purposes by creating missingness for covariates. We examine the bias and variance properties of FI compared to two versions of MI as well as CC estimation. Estimation of τ0 under fully observed data is also conducted as a reference point.

4.1 Simulation Setup

Let X = (X1, X2, X3) be confounders associated with the treatment effect. We generate X3 from a Bernoulli(0.2), and we generate (X1, X2) from a bivariate normal N(μX3,ΣX3)conditional on X3 where

μ1=(11),μ0=(11), and Σ1=Σ0=(10.50.51).

We specify the propensity score as e(X,θ)={1+e(0.3+0.2X10.1X20.1X3)}1and generate the treatment indicator Ai from a Bernoulli{e(Xi , θ)}. We choose θ to ensure the starting propensity scores are well behaved (i.e., between 0.1 and 0.9) and satisfy the sufficient overlap assumption.

We generate the outcome Y as

Y=X1+X2X3+2A+0.5A×X1+0.25A×X2+ϵY

where ϵYN(0, 1). The addition of the treatment-covariate interaction terms was implemented to better match the simulation data to what is observed in practice. Note that because X1 and X2 have mean 0, these terms fall out in expectation; however to approximate the true value of τ0, both Y(1) and Y(0) were evaluated for all records so the average difference in potential outcomes could be taken. Simulation results for bias and coverage were calculated for each sample relative to this within-sample τ0.

To simulate our missingness, we generate R from a Bernoulli(Φ) with

Φ=P(R=1A,X1,X3,Y)=1{1+e(0.25+0.25X10.6X3+0.5A+0.4Y)}1.

The coefficient values in Φ approximate a missingness rate of 0.33. Table 1 is a two by two contingency table for the missingness and the treatment assignment, averaging across all simulated data sets.

Table 1

Average Missingness vs Treatment Assignment

ControlTreatmentTotal
Complete Cases0.2940.3890.683
Cases Missing X20.2310.0870.317
Total0.5240.4761.000

A new X2variable is next constructed according to the response indicator, where

X2={X2ifR=1missingelse.

The observed data is Z=(X1,X2,X3,A,Y,R).

In FI, to motivate the proposal distribution for the missing X2records, we regress X2 on X1 and X3 based on the complete cases. The proposal distribution h in Step 0 of the FI algorithm is a non-central t(4) distribution with the mean and variance matched with the regression model. In Steps 1 and 2, the outcome model is specified as a linear regression model with predictors X, A, and their interactions. The propensity score model is the same as the true model e(X, θ). Under these settings, we generate M=200 imputed values for each missing X2.The imputation loop continues until either 250 iterations are performed or all parameters converge within 1x10−6. The FI loop produces final FI weights which can then be used to calculate the IPW and AIPW estimates for τ0 under FI. We compare these estimates versus CC and MI estimates. For CC, the process is straightforward. For MI we use the mi package. We examine MI under two settings, one where the outcome is included when imputing covariates (which we have labeled MI1), and the other where it is excluded (MI2). While the literature surrounding MI indicates utilizing the response in imputing covariates is preferred [43, 44, 45], MI2 has been included for completeness. Lastly, we calculate τ^IPW and τ^AIPWestimates utilizing the full data (as if no missingness had been introduced) for benchmark comparison.

This process was run 2000 times. To estimate variances in each iteration for each method, a leave-10-out jackknife was used.

4.2 Simulation Results

Once all simulations had been run for all methods, the results were as follows:

Table 2

Mean, Coverage, and Execution Results by Method: IPW Results

MethodMADMSEAverage JackKnife SECoverage
FI0.06440.00650.082295.5%
MI10.06370.00650.091197.8%
MI20.08010.00980.091293.4%
CC0.13810.02460.076456.6%
Full0.05250.00430.064394.9%

Similar results were seen for the AIPW estimates:

As expected, both MI and FI do better than CC estimators in all regards. It is worth noting how sizable an impact the exclusion of the response variable Y in the imputation step had on MI. Bias increased by 23%

Table 3

Mean, Coverage, and Execution Results by Method: AIPW Results

MethodMADMSEAverage JackKnife SECoverage
FI0.06440.00650.082095.5%
MI10.06480.00650.091797.6%
MI20.07980.00970.092293.5%
CC0.13950.02500.075455.6%
Full0.05180.00420.063394.8%

when not including Y. Comparing FI to MI1, FI and MI1 had comparable bias, but FI outperformed MI1 in coverage. The FI coverage is much more in-line with the nominal 95% coverage used in the confidence interval calculation. Multiple imputation (as viewed as MI1) saw over-coverage compared to the full data results. This can be attributed to an inflated standard error.

As to performance, average execution times for FI,MI1, and MI2 were similar. The average time of execution for each method (to perform both the estimate of tau0 and to complete the jack-knife estimation of its variance) are presented in table 4. Execution times for CC and Full were negligible and are not reported.

Table 4

Average Execution Times by Method

MethodExecution Time (Mins)
FI24:46
MI123:41
MI223:44

5 Application to the National Health and Nutrition Examination Survey Data

To illustrate the practical usage, we apply our method to a data set from the U.S. National Health and Nutrition Examination Survey. We estimate the effect of cigarette smoking on blood lead levels with age, gender, race, education, and income used as confounders. Of the confounders, only income was subject to missingness at a rate of 8.5% overall (6.0% smokers, 9.2% non-smokers). In the published data set, missing income values were imputed using mean imputation which risks being biased under MAR [10]. We investigate how the MI and FI estimates change from the analysis based on mean imputation. CC was also examined for a common point of reference.

As in our simulation, we will need to model our propensity scores, as well as regress both our Y (lead level) on all confounders as well as our missing confounder (income) on all present confounders. With respect to our regression for income in FI, we built a model of the form

Xinc=α0+α1Xage+α2Xmale+αeduTXedu+αraceTXrace+ϵ

where αedu and Xedu represent the parameter estimates and data for the dummy variables that represent the 6 education levels (similarly for αrace and Xrace for its 5 levels of race) and ϵ is an error term with mean 0. The dummy variables for unknown education and other race were excluded as they were, by construction, linearly dependent on the other columns in their group. This regression gives us an estimate Xinc;i(μ)of the initial mean income vector and of the variance in income in the form of σXinc(init).As before, this was used to create M=200 imputed data values for each of the 285 missing cases drawn as Xinc;ij=Xinc;i(μ)+t˜σXinc(init)where ˜t was drawn from a t(4) distribution. hXincwas calculated similarly as before. A leave-10-out jackknife was still used to estimate variances.

For MI, the same regression as used in FI was performed but with the response and treatement added in as linear terms. The regression equation for income used under MI was then:

Xinc=α0+α1Xage+α2Xmale+αeduXedu+αraceXrace+αsmokeA+αleadY+ϵ

where A is the treatment indicator for smoking and Y the reported lead level in the blood and ϵ is an error term with mean 0. Initial values for the mean and variance of Xinc could then be constructed using the complete cases. Imputed values of income were then drawn from the predicted posterior distribution and then subsequently updated via an MCMC process.

The resulting IPW and AIPW estimates and the accompanying variances for τ can be found in table 5. Additionally, the execution time for each method is provided.

Table 5

Results for estimating the effect of cigarette smoking on blood lead levels: estimate (Est), standard error (SE), and execution time

IPWAIPW
MethodEstSEEstSEExecution Time (Hrs)
FI1.2900.2310.9320.1630:59:05
MI11.2810.2570.9290.1721:30:18
CC1.1630.2280.9420.1840:00:18
Original1.2560.2100.9240.155N/A

While we can not know the true values of income from which we could calculate our bias, an examination of resource utilization does prove useful. It is unsurprising that CC would take minimal time since it does not need to attempt any convergence loop. FI and MI take significantly longer, but we would argue the increase in time is worth the added asymptotic bias advantages over CC. In comparing execution times, MI took about 50% longer to finish the full mean and jack-knife execution process than FI. As to why this resource gain exists for FI over MI, we postulate that since FI does not have to model and redraw from the full distribution of X every iteration, it is able to arrive at estimates of τ0 and standard errors much more quickly than MI.

6 Summary and Future Work

We have demonstrated that FI is an effective method for addressing missingness in covariates when estimating average treatment effects under the condition that covariates are MAR via simulation study and real-world application. Moreover, we were able to demonstrate FI’s superiority over the existing leading methods of MI and CC. FI produces lower bias and better coverage properties than either MI or CC. We also showed that when deployed in a real-world setting, FI is less resource-intensive than MI-based methods, most likely due to the lack of need to estimate the full covariate distribution.

With these results in mind, it is worth noting a comment made by Rubin [46] in his 18 year retrospective of his original work introducing MI. As an initial defense to then contemporary critiques of the method, some of which have been cited here already (see [15] and [23]), Rubin offered up the response that in cases where randomization validity (i.e., actual confidence coverage = nominal interval coverage) is difficult to achieve, statisticians should alternatively seek confidence validity (i.e., actual interval coverage ≥ nominal interval coverage) with decisions between competing methods decided by which method has the shortest interval. Near the end of that defense the reader can find the following comment:

Of course, if we have a procedure that is confidence valid but not randomization valid, there is hope that a better confidence-valid procedure exists (i.e., one with shorter intervals), which is also randomization valid, but in general this is not achievable [46, p. 475].

It is our belief the results above demonstrate FI produces randomization valid inference based on general estimation approaches for the population average treatment effect that MI may lack.

In our own future work, we will extend the FI algorithm in Section 3.1 to the MNAR setting by the inclusion of a model for the missingness and considering the full likelihood from equation (4). When the covariates are MNAR, an important challenge is that the full likelihood function (4) is not identifiable (or recoverable under [9]) in general. To overcome this challenge, one may consider non-response instrumental variable methods (e.g., [47]), missingness graphical models (e.g., [9]), negative controls (e.g., [48]), and sensitivity analysis (e.g., [49]). Once the identification conditions are established, the FI algorithm can be developed similarly based on the full likelihood equation (4). Finally, we would like to explore FI’s potential uses in other causal inference methods for treatment effect estimation beyond weighting methods, particularly as it applies to matching methods.

References

[1] Rosenbaum, P. R. and Rubin, D. B. [1983]. The central role of the propensity score in observational studies for causal effects, Biometrika 70: 41–55.10.21236/ADA114514Search in Google Scholar

[2] Bang, H. and Robins, J. M. [2005]. Doubly robust estimation in missing data and causal inference models, Biometrics 61: 962–973.10.1111/j.1541-0420.2005.00377.xSearch in Google Scholar PubMed

[3] Robins, J. M., Rotnitzky, A. and Zhao, L. P. [1995]. Analysis of semiparametric regression models for repeated outcomes in the presence of missing data, Journal of the American Statistical Association 90: 106–121.10.1080/01621459.1995.10476493Search in Google Scholar

[4] Rotnitzky, A. and Vansteelandt, S. [2014]. Double-robust methods, in G. Molenberghs, G. Fitzmaurice, M. G. Kenward, A. Tsiatis and G. Verbeke (eds), Handbook of missing data methodology Handbooks of Modern Statistical Methods, CRC Press, pp. 185–212.Search in Google Scholar

[5] Lunceford, J. K. and Davidian, M. [2004]. Stratification and weighting via the propensity score in estimation of causal treatment effects: a comparative study, Statistics in Medicine 23: 2937–2960.10.1002/sim.7231Search in Google Scholar PubMed

[6] Rubin, D. B. [1976]. Inference and missing data, Biometrika 63: 581–592.10.1093/biomet/63.3.581Search in Google Scholar

[7] Little, R. J. A. [1988]. A test of missing completely at random for multivariate data with missing values, Journal of the American Statistical Association 83(404): 1198–1202.10.1080/01621459.1988.10478722Search in Google Scholar

[8] Mohan, K. and Pearl, J. [2014]. On the testability of models with missing data, Artificial Intelligence and Statistics pp. 643–650.Search in Google Scholar

[9] Mohan, K. and Pearl, J. [2018]. Graphical models for processing missing data, arXiv preprint arXiv:1801.0358310.1080/01621459.2021.1874961Search in Google Scholar

[10] Council, N. R. [2010]. The Prevention and Treatment of Missing Data in Clinical Trials The National Academies Press, Washington, DC.Search in Google Scholar

[11] White, I. R. and Carlin, J. B. [2010]. Bias and efficiency of multiple imputation compared with complete-case analysis for missing covariate values, Statistics in Medicine 29: 2920–2931.10.1002/sim.3944Search in Google Scholar PubMed

[12] Rubin, D. B. [1987]. Multiple Imputation for Nonresponse in Surveys New York: Wiley.10.1002/9780470316696Search in Google Scholar

[13] Clogg, C. C., Rubin, D. B., Schenker, N., Schultz, B. and Weidman, L.Search in Google Scholar

[14] Binder, D. A. and Sun, W. [1992]. Frequency valid multiple imputation for surveys with a complex design, Proceedings of the Section on Survey Research Methods American Statistical Association.Search in Google Scholar

[15] Fay, R. E. [1992]. When are inferences from multiple imputation valid, Proceedings of the Section on Survey Research Methods American Statistical Association.Search in Google Scholar

[16] Fay, R. E. [1996]. Alternative paradigms for the analysis of imputed survey data, Journal of the American Statistical Association 91: 490–498.10.1080/01621459.1996.10476909Search in Google Scholar

[17] Feodor Nielsen, S. [2003]. Proper and improper multiple imputation, International Statistical Review 71: 593–607.10.1111/j.1751-5823.2003.tb00214.xSearch in Google Scholar

[18] Kim, J. K., Michael Brick, J., Fuller, W. A. and Kalton, G. [2006]. On the bias of the multiple-imputation variance estimator in survey sampling, Journal of the Royal Statistical Society: Series B (Statistical Methodology) 68: 509–521.10.1111/j.1467-9868.2006.00546.xSearch in Google Scholar

[19] Kott, P. [1995]. A paradox of multiple imputation, Proceedings of the Section on Survey Research Methods American Statistical Association, pp. 384–389.Search in Google Scholar

[20] Robins, J. and Wang, N. [2000]. Inference for imputation estimators, Biometrika 87: 113–124.10.1093/biomet/87.1.113Search in Google Scholar

[21] Wang, N. and Robins, J. M. [1998]. Large-sample theory for parametric multiple imputation procedures, Biometrika 85: 935–948.10.1093/biomet/85.4.935Search in Google Scholar

[22] Yang, S. and Kim, J. K. [2016c]. A note on multiple imputation for method of moments estimation, Biometrika 103: 244–251.10.1093/biomet/asv073Search in Google Scholar

[23] Meng, X.-L. [1994]. Multiple-imputation inferences with uncongenial sources of input, Statistical Science 9: 538–558.10.1214/ss/1177010269Search in Google Scholar

[24] Wei, G. C. and Tanner, M. A. [1990]. A monte carlo implementation of the em algorithm and the poor man’s data augmentation algorithms, Journal of the American Statistical Association 85: 699–704.10.1080/01621459.1990.10474930Search in Google Scholar

[25] Booth, J. G. and Hobert, J. P. [1999]. Maximizing generalized linear mixed model likelihoods with an automated monte carlo em algorithm, Journal of the Royal Statistical Society: Series B 61: 265–285. [1991]. Multiple imputation of industry and occupation codes in census public-use samples using bayesian logistic regression, Journal of the American Statistical Association 86: 68–78.10.1111/1467-9868.00176Search in Google Scholar

[26] Kim, J. K. [2011]. Parametric fractional imputation for missing data analysis, Biometrika 98: 119–132.10.1093/biomet/asq073Search in Google Scholar

[27] Yang, S. and Kim, J. K. [2016a]. Fractional imputation in survey sampling: A comparative review, Statist. Sci. 31: 415–432.10.1214/16-STS569Search in Google Scholar

[28] Kim, J. K. and Yang, S. [2014]. Fractional hot deck imputation for robust inference under item nonresponse in survey sampling, Survey Methodology 40: 211–230.Search in Google Scholar

[29] Yang, S. and Kim, J. K. [2016b]. Likelihood-based inference with missing data under missing-at-random, Scandinavian Journal of Statistics 43: 436–454.10.1111/sjos.12184Search in Google Scholar

[30] Yang, S., Kim, J. K. and Shin, D. w. [2013]. Imputation methods for quantile estimation under missing at random, Statistical Interface 6: 369–377.10.4310/SII.2013.v6.n3.a7Search in Google Scholar

[31] Yang, S., Kim, J. K. and Zhu, Z. [2013]. Parametric fractional imputation for mixed models with nonignorable missing data, Statistical Interface 6: 339–347.10.4310/SII.2013.v6.n3.a4Search in Google Scholar

[32] Neyman, J. [1923]. Sur les applications de la thar des probabilities aux experiences Agaricales: Essay de principle. English translation of excerpts by Dabrowska, D. and Speed, T., Statistical Science 5: 465–472.Search in Google Scholar

[33] Rubin, D. B. [1974]. Estimating causal effects of treatments in randomized and nonrandomized studies., Journal of Educational Psychology 66: 688–701.10.1037/h0037350Search in Google Scholar

[34] Rubin, D. B. [1978]. Bayesian inference for causal effects: The role of randomization, The Annals of Statistics 6: 34–58.10.1214/aos/1176344064Search in Google Scholar

[35] Imbens, G. W. and Rubin, D. B. [2015]. Causal Inference in Statistics, Social, and Biomedical Sciences Cambridge University Press, Cambridge UK.10.1017/CBO9781139025751Search in Google Scholar

[36] Zhang, Z., Liu, W., Zhang, B., Tang, L. and Zhang, J. [2016]. Causal inference with missing exposure information: Methods and applications to an obstetric study, Statistical Methods in Medical Research 25: 2053–2066. PMID: 24318273.10.1177/0962280213513758Search in Google Scholar PubMed

[37] Rubin, D. B. [2007]. The design versus the analysis of observational studies for causal effects: parallels with the design of randomized trials, Statistics in Medicine 26: 20–36.10.1002/sim.2739Search in Google Scholar PubMed

[38] Rubin, D. B. [2008]. For objective causal inference, design trumps analysis, The Annals of Applied Statistics 2: 808–840.10.1214/08-AOAS187Search in Google Scholar

[39] Stuart, E. A. [2010]. Matching methods for causal inference: A review and a look forward, Statistical Science 25: 1.10.1214/09-STS313Search in Google Scholar PubMed PubMed Central

[40] Kim, J. and Shao, J. [2013]. Statistical Methods for Handling Incomplete Data A Chapman & Hall book, Taylor & Francis.10.1201/b13981Search in Google Scholar

[41] Louis, T. A. [1982]. Finding the observed information matrix when using the em algorithm, Journal of the Royal Statistical Society: Series B 44: 226–233.10.1111/j.2517-6161.1982.tb01203.xSearch in Google Scholar

[42] Pfeffermann, D., Skinner, C. J., Holmes, D. J., Goldstein, H. and Rasbash, J. [1998]. Weighting for unequal selection probabilities in multilevel models, Journal of the Royal Statistical Society: Series B 60: 23–40.10.1111/1467-9868.00106Search in Google Scholar

[43] Moons, K. G., Donders, R. A., Stijnen, T. and Harrell, F. E. [2006]. Using the outcome for imputation of missing predictor values was preferred, Journal of Clinical Epidemiology 59: 1092–1101.10.1016/j.jclinepi.2006.01.009Search in Google Scholar PubMed

[44] Nguyen, C. D., Carlin, J. B. and Lee, K. J. [2017]. Model checking in multiple imputation: an overview and case study, Emerging Themes in Epidemiology 14: 8.10.1186/s12982-017-0062-6Search in Google Scholar PubMed PubMed Central

[45] Sterne, J. A. C., White, I. R., Carlin, J. B., Spratt, M., Royston, P., Kenward, M. G., Wood, A. M. and Carpenter, J. R. [2009]. Multiple imputation for missing data in epidemiological and clinical research: potential and pitfalls, BMJ 338.10.1136/bmj.b2393Search in Google Scholar PubMed PubMed Central

[46] Rubin, D. B. [1996]. Multiple imputation after 18+ years, Journal of the American Statistical Association 91: 473–489.10.1080/01621459.1996.10476908Search in Google Scholar

[47] Yang, S., Wang, L. and Ding, P. [2019]. Causal inference with confounders missing not at random, Biometrika p. accepted.10.1093/biomet/asz048Search in Google Scholar

[48] Kuroki, M. and Pearl, J. [2014]. Measurement bias and effect restoration in causal inference, Biometrika 101: 423–437.10.21236/ADA557455Search in Google Scholar

[49] Cornfield, J., Haenszel, W., Hammond, E. C., Lilienfeld, A. M., Shimkin, M. B. and Wynder, E. L. [1959]. Smoking and Lung Cancer: Recent Evidence and a Discussion of Some Questions, JNCI: Journal of the National Cancer Institute 22: 173–203.10.1093/ije/dyp289Search in Google Scholar PubMed

[50] Graham, J. W., Olchowski, A. E. and Gilreath, T. D. [2007]. How many imputations are really needed? some practical clarifications of multiple imputation theory, Prevention Science 8: 206–213.10.1007/s11121-007-0070-9Search in Google Scholar PubMed

[51] von Hippel, P. T. [2016]. The number of imputations should increase quadratically with the fraction of missing information, arXiv preprint arXiv:1608.05406 .Search in Google Scholar

A1 Appendix

A1.1 Identifiability of Average Treatment Effect under MAR

Proof of Theorem 1

It is well known that under MAR, the full data distribution is identifiable from the observed data. To show that the average treatment effect τ0 can be identified from the observed data, we express

τ0=E{Y(1)Y(0)}=EAYe(X)(1A)Y1e(X)=EXmis AYeXobs ,xmis (1A)Y1eXobs ,xmis fXmis =xmis Xobs, A,Ydxmis ,

where the second equality follows by Assumption 1 and 2, and the fourth equality follows by Assumption 3.

A1.2 Example FI implementation

Example A1

(Generating a complete data set under FI). For illustration, consider a population where covariates X1and X2, treatment indicator A, and response indicator Y are all binomial variables. Let X1Binomial(0.5). Let X2Binomial(ϕX1)whereϕX1=0.4+0.2X1.Let ABinomial{(1+eX1+X2)1}.

Finally, let YBinomial{(1+eX1+X22A)1}.

Suppose we draw a sample of size n = 10 from the population described above but some observation are missing values for X2with probabilitypR=(1+eX1+A+Y)1,leaving us with the starting data set in Table A1.

Table A1

Pre-imputed data. Observation 2 and observation 3 are both missing values for X2.

IDX1X2AY
11000
21N/A00
30N/A00
40100
50101
60100
71010
80011
91100
100100

To generate imputed values for units 2 and 3, a proposal distribution must be provided. Adopting the decomposition from equation 5, a natural choice is h(X2) = f (X2 | X1; α). Since X2is binomial, we can estimate f (X2 | X1; α) via logistic regression. At this point, α can only be estimated based on the complete cases. This results in the proposal distribution

P(X2=1)=11+e(1.3862.079X1),

withα^=(α^0,α^1)=(1.386,2.079)or put more plainly the probability mass function corresponding to Table A2.

Table A2

PMF for imputation function h().

X1X2f (X2 |X1; α^)
110.3333
100.6667
010.8000
000.2000

With h() in hand, we can generate M new values for each missing X2. If we let M = 5, then this would mean we would generate 5 values(X2,2(j);j=1,,5)for observation 2 drawing from Binomial(0.3333) and 5 values(X2,3(j);j=1,,5)for observation 3 drawing from Binomial(0.8000). At this point in the process, no weights have been updated, so the fractional weight on each of these imputed values is 1/M, leaving us with our initial imputed data set (Table A3) to conclude the I-step. Note, since the imputed valuesX2never change, the valuesh(X2)never change, so it is convenient to include a column holding these values, as they will be used repeatedly when updating the fractional weights in each W-step. In Table A3 the column holding the proposal distribution likelihood for each observation is labeled h0.

Table A3

Initial imputed data set prior to the first W-step.

IDX1X2AYω*h0
1100010.6667
2.110000.20.6667
2.211000.20.3333
2.310000.20.6667
2.411000.20.3333
2.511000.20.3333
3.100000.20.2000
3.200000.20.2000
3.300000.20.2000
3.401000.20.8000
3.501000.20.8000
4010010.8000
5010110.8000
6010010.8000
7101010.6667
8001110.2000
9110010.3333
10010010.8000

To begin the first W-step, we must first estimate the full joint distribution of (X1, X2, A, Y). Again we will return to the decomposition in equation 5, and because both A and Y are binomial we will choose to estimate the distributions e(X1, X2 | θ) and f (Y | X1, X2, A; β) using weighted logistic regression. Similarly, from now on we will be estimating f (X2 | X1; α) using weighted logistic regression to incorporate the fractional weights (whereas before, only the complete cases were used in the I-step). Initial values ofα^(t),β^(t),θ^(t)for t = 1 are as follows:

  1. α^(1)=(α^0,α^1)=(1.0116,1.4171),

  2. β^(1)=(β^0,β^1,β^2,β^A)=(19.7160,39.1894,18.4922,39.3107),

  3. θ^(1)=(θ^0,θ^1,θ^2)=(0.5108,0.8473,20.8207).

Once initial values for(α^,β^,θ^)are estimated, it is convenient to calculate the likelihood of each observation’sX2,A, and Y under their respective current estimated distributions, that is to say calculatingP(X2=X2,i(j)α^(1))for all observations and similarly for A and Y with their distributions and current parameters. For short hand, letfX2(),fA(), and fY () be the likelihood functions for X2, A, and Y respectively under the current parameter valuesα^(t),β(t),θ(t)that were just calculated. For example, fY(Yi)=f(YiAi,X1,i,X2,i(j);β^(t)).With the respective likelihoods calculated, each observation’s fractional weight can be updated as

ωi(j)=fX2(X2,i(j))fA(Ai)fY(Yi)h0,ij.

It is possible at this point thatj=1Mωi(j)1for an individual. If so the weights for that individual will need re-normalized to ensure the conditionj=1Mωi(j)=1is still meti. The updated data set after the first W-step can be seen in Table A4. Note the only data elements to change between Table A3 and Table A4 are the fractional weights in column ω*.

Table A4

Updated data after the first W-step.

IDX1X2AYω*h0
1100010.6667
2.110000.11290.6667
2.211000.25810.3333
2.310000.11290.6667
2.411000.25810.3333
2.511000.25810.3333
3.100000.17140.2000
3.200000.17140.2000
3.300000.17140.2000
3.401000.24290.8000
3.501000.24290.8000
4010010.8000
5010110.8000
6010010.8000
7101010.6667
8001110.2000
9110010.3333
10010010.8000

The M-step, follows naturally from the W-step. Given the updated weights ω*in Table A4, parametersα^(t),β^(t),andθ^(t)are all updated from t = 1 to t = 2 again using weighted logistic regression. In the first M-step that would mean updatingα^(1),β^(1),θ^(1)to

  1. α^(2)=(α^0,α^1)=(1.0860,1.3127),

  2. β^(2)=(β^0,β^1,β^2,β^A)=(19.7179,39.2118,18.4692,39.3238),

  3. θ^(2)=(θ^0,θ^1,θ^2)=(0.6650,0.8686,20.9515).

The updated parameters are checked against some convergence criteria (for instance stopping once the max absolute difference between all parameters is ≤ 0.0001). If the convergence criteria is not met, the process cycles back to the W-step, and weights are recalculated now using parametersα^(2),β^(2),θ^(2).With new weights the M-step will update distributional parametersα^(t),β^(t),θ^(t),and the process continues until the convergence criteria are met. In our sample, the parameters converged on the 18th iteration, producing the final FI-completed data set in Table A5. This data set can now be incorporated into any weighted analysis where a complete-data consistent estimator exists

Table A5

Final FI-completed data set.

IDX1X2AYω*
110001
2.110000.0886
2.211000.2743
2.310000.0886
2.411000.2743
2.511000.2743
3.100000.1334
3.200000.1334
3.300000.1334
3.401000.3000
3.501000.3000
401001
501011
601001
710101
800111
911001
1001001

A1.3 Sensitivity to Imputation Size

In our primary simulation study, the imputation size M was established to be 200 with the assumption that such an M would be sufficiently large to obtain approximately asymptotic results. Traditional recommendations for the size of M are much smaller (M = 2toM = 10)when only inference about the point estimates were of interest. More recent recommendations for the selection of M have occurred when also needing accurately estimate variance of the point estimate are somewhat higher. For our data set with %-missingness approximately equal to 0.33, recommendations for an M size in the context of MI range from 20-40 [50, 51]. To examine if we could lower our M setting for FI we reran the first 500 simulations of our analysis varying M to be M = 5, 10, 20, 50, 100. The results are summarized below in both Figure A1 and in Table A6.

Figure A1 Comparing sensitivity to size of M on bias and coverage among FI and MI implementations when estimating treatment effects via IPW
Figure A1

Comparing sensitivity to size of M on bias and coverage among FI and MI implementations when estimating treatment effects via IPW

Table A6

Table of results for comparing sensitivity to size of M on bias and coverage among FI and MI implementations

IPWAIPW
SettingBiasCoverageBiasCoverage
FI_M50.065100.0%0.064100.0%
MI1_M50.065100.0%0.065100.0%
MI2_M50.08199.6%0.08099.8%
FI_M100.062100.0%0.061100.0%
MI1_M100.06499.4%0.06599.4%
MI2_M100.08298.8%0.08198.8%
FI_M200.06299.4%0.061100.0%
MI1_M200.06498.2%0.06598.2%
MI2_M200.08196.8%0.08197.4%
FI_M500.06497.6%0.06398.0%
MI1_M500.06497.4%0.06597.2%
MI2_M500.08294.8%0.08195.4%
FI_M1000.06495.8%0.06396.2%
MI1_M1000.06497.2%0.06597.0%
MI2_M1000.08294.6%0.08194.6%
FI_M200*0.06495.5%0.06495.5%
MI1_M200*0.06497.8%0.06597.6%
MI2_M200*0.08093.4%0.08093.5%

Our results demonstrate no loss of accuracy between FI and MI1 regardless of the setting of M. There also is no gain in accuracy among any of the methods with increasing M. This matches with existing literature that if only point estimation is of interest then low M settings are sufficient for MI. Our results further suggest FI can also permit low M settings when variance estimation is not of interest. However, all methods perform poorly with respect to coverage until M is at least greater than 20. As Graham, Olchowski, and Gilreath [50] and von Hippel [51] suggest, low M values are not sufficient for accurate variance estimation. Moreover, if focusing on MI1, the coverage gains stop somewhere between M = 20 and M = 50 (again, as expected), but the MI1 coverage is still about 2% higher than the nominal coverage.

As for FI, it takes slightly higher M to surpass the coverage potential of MI1. At M = 100, FI coverage was within 1% of nominal coverage with gains still being seen as M increased to the settings used in our simulations. From these results, we conclude M = 200 was a sufficient setting for our simulations though even better coverage may have been attainable with higher M. Furthermore, if computational resources are limited setting M to only 100 may be a passable setting – still superior to MI but not yet reaching approximate asymptotic properties. It is also important to note that these sensitivity results are only confirmatory for our simulation settings and may differ when FI and MI are applied in more complex settings or when there is a higher level of missingness. As such, we recommend plotting coverage curves like these when deploying either method in future applications to validate M has been set sufficiently high in those situations.

Received: 2018-09-18
Accepted: 2020-08-06
Published Online: 2020-12-31

© 2020 N. Corder and S. Yang, published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 17.9.2025 from https://www.degruyterbrill.com/document/doi/10.1515/jci-2019-0024/html
Scroll to top button