Startseite Prospective and retrospective causal inferences based on the potential outcome framework
Artikel Open Access

Prospective and retrospective causal inferences based on the potential outcome framework

  • Zhi Geng , Chao Zhang , Xueli Wang , Chunchen Liu und Shaojie Wei EMAIL logo
Veröffentlicht/Copyright: 24. Oktober 2024
Veröffentlichen auch Sie bei De Gruyter Brill

Abstract

In this article, we discuss both prospective and retrospective causal inferences, building on Neyman’s potential outcome framework. For prospective causal inference, we review criteria for confounders and surrogates to avoid the Yule–Simpson paradox and the surrogate paradox, respectively. For retrospective causal inference, we introduce the concepts of posterior causal effects given observed evidence to quantify the causes of effects. The posterior causal effects provide a unified framework for deducing both effects of causes in prospective causal inference and causes of effects in retrospective causal inference. We compare the medical diagnostic approaches based on Bayesian posterior probabilities and posterior causal effects for classification and attribution.

MSC 2010: 62D20

1 Introduction

Causal inference has a solid theoretical foundation based on the potential outcome framework, which was first proposed by Neyman (1923) for experimental studies [1] and later extended by Rubin (1974) to observational studies [2]. This framework allows causal concepts and questions to be formally defined and represented mathematically. Without this formal framework, causal relationships are often conflated with correlational relationships, leading to mistaken inferences. By grounding inference in the potential outcome framework, we move beyond simply observing correlations between variables. This allows us to define causal effects more precisely, make essential assumptions to identify these effects from observational data, and develop estimators with desirable statistical properties. In this way, the framework enables rigorous causal inference that aims to uncover genuine causal relationships from both experimental and observational data.

Causal inference involves not only evaluating the effects of causes in a prospective causal inference, but also deducing the causes of effects in a retrospective causal inference. In epidemiology, both prospective and retrospective studies concern the design stage. Prospective and retrospective causal inferences concern the analysis stage. Prospective causal inference is to evaluate effects of causes, which is typically forward-looking, while retrospective causal inference is to deduce causes of effects, which is typically backward-looking [3,4]. For example, it is a prospective causal problem to determine whether a drug will have the effect of lowering blood pressure, while it is a retrospective causal problem when we know that a person died, and we retrospectively ask whether the death was caused by a particular drug. Dawid et al. [5] highlighted an important distinction between effects of causes and causes of effects. Statistical causality emphasizes evaluating the effects of causes rather than the causes of effects [6,7]. Randomized experiments are the gold standard for evaluating causal effects in prospective causal inference. However, for retrospective causal inference, even under randomized experiments, identifying the causes of effects is difficult.

In observational studies, confounding poses a major threat to valid causal inference about effects. The Yule–Simpson paradox provides a striking example of how ignoring a confounder between treatment and outcome can completely reverse an association. Similarly, the surrogate paradox can arise if there is a confounder between surrogate and true endpoints. Unless certain criteria are met, using the surrogate as a substitute for the true endpoint in assessing treatment effects can be misleading. To avoid inferential paradoxes and biases, careful consideration must be given to potential confounders and surrogates when making causal claims from observational data. In this article, we discuss the precise criteria that must be met for a variable to be a confounder or a valid surrogate for reviewing prospective causal inference and highlighting the contributions of the potential outcome framework. Understanding these criteria will allow more rigorous prospective causal inferences to be made from observational studies.

While statistical causality has focused more on prospective causal inference, deducing causes from observed effects is also an important causal reasoning task. Retrospective causal inference aims to determine the causes behind a specific effect or event that has already occurred, based on the observed data and causal assumptions. Dawid and Musio [7] highlighted that counterfactual reasoning is unnecessary for analyzing effects of causes prospectively, but essential for retrospective inference about individual-level causes. For example, whether a particular individual’s lung cancer is caused by smoking requires imagining the counterfactual scenario where the person did not smoke and assessing the probability they would still have developed cancer. Retrospective causal inference is more challenging than prospective inference for several reasons. Confounding can be more complex because conditioning on the occurred effect or outcome may induce additional biases not present in a prospective design [8]. Randomization and “no unobserved confounders” assumptions are often insufficient to eliminate this bias. Moreover, causal effects may be heterogeneous across individuals, so group-level estimates may not apply to a specific individual case with occurred effects. Despite these difficulties, retrospective causal inference has many vital applications including attributing causes in epidemiology and legal cases [5,912].

We discuss posterior causal effects to formally unify prospective and retrospective causal inference problems. Posterior causal effects are causal effects conditioned on observed evidence, which may include observed effect variables [13,14]. They thus measure the effects of causes in a subpopulation restricted by the evidence. Depending on the evidence, posterior effects can be used for both prospective and retrospective inferences. When the evidence excludes effect variables, posterior effects evaluate causes prospectively. For instance, the posterior effect of smoking on lung cancer given age and gender evidence defines the causal effect in that age/gender population. This evaluates the prospective effect in a specific subpopulation. In contrast, when the evidence includes effect variables, posterior effects deduce causes retrospectively. The posterior effect of smoking on lung cancer given occurred lung cancer evidence defines the effect in lung cancer patients. This can judge the possibility that smoking caused lung cancer retrospectively. Posterior causal effects also explain the causal meaning of population attributable risks commonly used in public health and epidemiology. These measure the proportion of cases attributable to an exposure. Posterior effects formally connect attributable risks to causal effects conditioned on observed effects.

In this article, we offer a review of some topics in prospective and retrospective causal inferences, based on the potential outcome framework. The remainder of this article is organized as follows. In Section 2, we review contribution of the potential outcome framework to prospective causal inference, focusing on confounders and surrogate endpoints. We discuss criteria for confounders and surrogate endpoints that help avoid the Yule–Simpson paradox and the surrogate paradox, respectively. Section 3 then covers retrospective causal inference based on the framework. We introduce probabilities of causation and posterior causal effects based on counterfactual reasoning. We also interpret population attributable risks in epidemiology through the lens of posterior causal effects. Finally, we compare medical diagnostic approaches based on Bayesian posterior probabilities versus posterior causal effects.

2 Prospective causal inference

In prospective causal inference, a goal is to evaluate the effect of a cause event that occurred earlier on an outcome event that occurred later. Suppose that all variables presented below are binary, 1 denotes presence and 0 absence. Let X denote an observed cause variable (e.g., smoking) that happened earlier at time t 1 , and Y denote an observed effect variable (e.g., lung cancer) that occurred at time t 2 ( t 1 < t 2 ). The association between smoking and lung cancer can be measured using the observed data of X and Y , such as Pearson’s correlation or relative risks. However, causation between smoking and lung cancer cannot be well defined only by the notation of two observed variables X and Y . To describe the causation, Neyman [1] and Rubin [2] proposed the following notation of potential outcomes. Let Y x denote the potential outcome that would occur at time t 2 if an individual were exposed to the cause X = x at time t 1 . The individual causal effect of cause X on response Y is defined as Y 1 Y 0 , and the average causal effect is E ( Y 1 Y 0 ) . Focusing on the treated population where X = 1 , the average treatment effect on the treated is expressed as E ( Y 1 Y 0 X = 1 ) . Generally, the probabilistic causal effect of X on Y for the treated population can be evaluated by comparing pr ( Y 1 = 1 X = 1 ) to the unobserved counterfactual probability pr ( Y 0 = 1 X = 1 ) , which is not identifiable without any assumption. Randomized experiments are the gold standard approach for identifying pr ( Y 0 = 1 X = 1 ) by the probability pr ( Y = 1 X = 0 ) of observed variables. For observational studies, identification requires some untestable assumptions.

2.1 Confounders

The problem about confounders has been explored for a long time, especially in epidemiology. However, the criteria for assessing confounders and confounding in the epidemiological literature have been inconsistent [1521]. Using Neyman’s potential outcome framework, confounding bias B is defined as the difference between the counterfactual probability of potential outcome without exposure in the exposed population and the probability of observed outcome in the unexposed population [18,22], i.e.,

(1) B = pr ( Y 0 = 1 X = 1 ) pr ( Y 0 = 1 X = 0 ) .

By adjusting the distribution pr ( C = k X = 0 ) of covariate C in the unexposed population to pr ( C = k X = 1 ) , a standardized probability pr Δ ( Y 0 = 1 X = 0 ) is defined as

pr Δ ( Y 0 = 1 X = 0 ) = k = 1 K pr ( Y 0 = 1 X = 0 , C = k ) pr ( C = k X = 1 ) .

A confounder is defined as a risk factor whose control can reduce the confounding bias [15,2326]. Replacing the counterfactual probability pr ( Y 0 = 1 X = 0 ) in (1) by the adjusted pr Δ ( Y 0 = 1 X = 0 ) , Geng et al. [25] defined a confounder as a covariate C for which

pr ( Y 0 = 1 X = 1 ) pr Δ ( Y 0 = 1 X = 0 ) < B .

This definition states that the standardized probability pr Δ ( Y 0 = 1 X = 0 ) adjusted for a confounder C is closer to the counterfactual probability pr ( Y 0 = 1 X = 1 ) than the observed probability pr ( Y = 1 X = 0 ) . For a case of C with multiple covariates, C may be recategorized by a single categorical variate C with the same number of categories as C . VanderWeele and Shpitser [26] considered a similar definition of confounders for the overall effect of the exposure on the whole population rather than the effect of the exposure on the exposed population. Note that these definitions of confounders do not need any assumption such as subpopulation-comparability or a known causal diagram. With this definition, we can determine that a covariate is not a confounder when pr Δ ( Y 0 = 1 X = 0 ) = pr ( Y 0 = 1 X = 0 ) , but we cannot confirm that it is a confounder since pr ( Y 0 = 1 X = 1 ) is not identifiable without further assumptions.

2.2 Surrogate endpoints

In many scientific studies, true endpoint variables cannot be measured or observed due to being expensive, inconvenient, or impractical within a short time span. For example, in clinical trials, CD4 count is used as a surrogate endpoint for survival time in acequired immune deficiency syndrome (AIDS) studies, and bone mass is used as a surrogate endpoint for fracture in osteoporosis studies. However, Fleming and Demets [27] pointed out that in many real clinical trials, surrogates failed to evaluate the treatment effects on true endpoints.

Chen et al. [28] introduced and formulated the surrogate paradox, where a treatment has a positive effect on a surrogate endpoint, which in turn has a positive effect on a true endpoint, but the treatment has a negative effect on the true endpoint. Even by conducting two randomized experiments, we can separately prove probabilistically both that a variate X has a positive causal effect on a variable Y and that the variate Y has a positive causal effect on a variable Z , but we cannot judge that X has a positive causal effect on Z , even if X does not have any direct causal effect on Z . Even if the intermediate variable Y breaks all causal paths from X to Z , the probabilistic causal relationships may not be transitive, although the individual causal relationships may be. The surrogate paradox implies that the sign of treatment effect on the endpoint cannot be predicted by the sign of treatment effect on the surrogate and the sign of causal effect of surrogate on the endpoint. Therefore, the logical reasoning may not be applied to the probabilistic results of causal inference. Jiang et al. [29] also discussed the transitivity of different associations under the conditional independence of X and Z given Y and showed that the finer an association measure is, the stronger its transitivity is.

Prentice [30] proposed a criterion for a statistical surrogate Y , which requires both a strong association between treatment X and the surrogate Y and the conditional independence of the true endpoint Z and treatment X given Y , denoted as X Z Y in the notation by Dawid [31]. The conditional independence means that the surrogate Y can break the association between the treatment X and the endpoint Z , and thus, X Y implies X Z . Frangakis and Rubin [32] presented the criterion for a principal surrogate Y , which should possess the causal necessity: a treatment X has a causal effect on an endpoint Z only if the treatment X has a causal effect on the surrogate Y . Lauritzen [33] proposed the criterion for a strong surrogate Y , which breaks all causal paths from X to Z in a causal diagram (Figure 1). Chen et al. [28] showed that the surrogate paradox cannot be avoided, even by such strong criteria of the statistical surrogates, the principal surrogates, and the strong surrogates. A surrogate Y is an intermediate variable in a causal path from X to Z , and variable X is an instrumental variable when Y is a strong surrogate. A more proper name for the paradox may be “the intermediate variable paradox” to reflect its wider generality. The same paradox applies to other situations. For example, the paradox can be called “the instrumental paradox” in the use of the instrumental variable. The surrogate paradox also points out an issue of the transitivity of causal effects on a causal path. Jiang et al. [34] proposed approaches to identifying the principal stratification causal effects by multiple trials and provided the criteria for surrogates that avoid the surrogate paradox.

Figure 1 
                  Criterion for a strong surrogate.
Figure 1

Criterion for a strong surrogate.

Moore [35] provided a real-world example of the surrogate paradox. Doctors knew that irregular heartbeat was a risk factor for sudden death and presumed that correcting irregular heartbeat would prevent sudden death. Therefore, they used “correction of heartbeat” as a surrogate, and several drugs (Enkaid, Tambocor and Ethmozine) were approved by FDA (Food and Drug Administration). However, the Cardiac Arrhythmia Suppression Trial [36] showed that these drugs did not improve survival times but increased mortality.

Chen et al. [28], Ju and Geng [37], Wu et al. [38], and VanderWeele and Shpitser [26] proposed consistent surrogates and criteria to avoid the surrogate paradox. These criteria apply to single surrogates only. However, in many applications, a treatment may affect the endpoint through multiple pathways, and thus, a single surrogate cannot break all of these pathways. For example, a drug may reduce the risk of death from AIDS through two pathways: by decreasing human immunodeficiency virus type 1 ribonucleic acid (HIV-1 RNA) concentrations and by increasing CD4 count. In this case, a single surrogate may not satisfy any criterion of the statistical, principal, strong, or consistent surrogates. Both HIV-1 RNA concentrations and CD4 count should be used as multiple surrogates for the risk of death from AIDS. Joffe [39] suggested that it is meaningful to generalize the criteria for a single surrogate to multiple surrogates. Luo et al. [40] proposed a criterion for multiple surrogates Y = ( Y 1 , , Y p ) based on stochastic orders of random vectors. All of these criteria for surrogates require some knowledge of causality or associations among the observed variables X and Y and the unobserved variable Z . Therefore, these criteria are not falsifiable without untestable assumptions or observed data for Z .

3 Retrospective causal inference

When evaluating the effects of causes, we prospectively predict the results of an intervention in a population. However, when deducing the causes of effects, we explore the causes of happened effects for a specific individual retrospectively. In doing so, we may have to imagine the potential outcomes if causes would happen in counterfactual scenarios. As will be seen below, counterfactual reasoning can be well described by the potential outcome framework. Retrospective causal inference can be used for causal attribution, medical diagnosis, and blame assignment. For example, scientific studies have evaluated the causal effects of benzene and ionizing radiation exposure on leukemia using data from experimental and observational studies. When we observe that a leukemia patient has been exposed to both benzene and ionizing radiation, we would like to know how much of the patient’s leukemia is attributable to benzene exposure and how much is attributable to ionizing radiation exposure, which is a problem about causes of effects.

Evaluating effects of causes is the main focus of most existing causal inference approaches, while deducing causes of effects is the focus of a few approaches. As Dawid [12] pointed out, assessing causes of effects is more challenging than assessing effects of causes, because the former is mainly a counterfactual inference problem for a single individual. For a counterfactual situation, measures for the probabilities of causation are generally not identifiable even when we use the gold standard approach of randomized experiments and there are no unobserved confounders.

3.1 Probabilities of causation

Pearl [11] provided counterfactual definitions of causation to capture how necessary and/or sufficient a cause is capable of producing a given effect or outcome. Dawid et al. [5] highlighted the distinction between the effects of causes and the causes of effects, and proposed the probability of causation to make inference about the causes of effects. Inferring the causes of effects requires more subtle logic and stronger assumptions than inferring the effects of causes. In the following, we focus on binary variables. First, we introduce probabilities of causation for the case of a single effect variable Y and a single cause X . To measure how possible X is a cause of an effect Y , Dawid et al. [5] defined the probability of causation as

PC ( X Y ) = pr ( Y X = 0 = 0 Y X = 1 = 1 ) .

To measure how necessary X is a cause of an occurred effect Y = 1 , Pearl [41] defined the probability of necessary causation as

PN ( X Y ) = pr ( Y X = 0 = 0 X = 1 , Y = 1 ) .

Lu et al. [13] proposed the posterior causal effects given observed evidence to measure the probabilities of causes and treated evaluating effects of causes and discovering causes of effects from the same perspective. Let C denote a pretreatment variable prior to treatment X (i.e., a covariate). Let O = o denote the observed evidence for the target individual, where o is an observed value of O . For an individual case, we can sometimes observe only a subset O of variables { X , Y , C } . Li et al. [14] defined the posterior total causal effect given O = o as

PostTCE ( X Y O = o ) = E ( Y X = 1 Y X = 0 O = o ) .

For the evidence O = ( C = c ) , the posterior total causal effect E ( Y X = 1 Y X = 0 C = c ) is an average causal effect conditional on C = c ; for the evidence O = ( X = 1 ) , the posterior total causal effect E ( Y X = 1 Y X = 0 X = 1 ) is an average causal effect in a treated subpopulation; for the evidence O = ( X = 1 , Y = 1 ) , the posterior total causal effect is equal to PN :

PostTCE ( X Y X = 1 , Y = 1 ) = 1 pr ( Y X = 0 = 1 X = 1 , Y = 1 ) = PN .

Thus, the posterior total causal effect can be used not only to evaluate effects of causes for a prospective causal inference, but also to assess causes of effects for a retrospective causal inference. For example, let X denote smoking and Y lung cancer. By the posterior total causal effect E ( Y X = 1 Y X = 0 X = 1 , Y = 1 ) , we evaluate the causal effect of smoking on lung cancer in the subpopulation of smokers with lung cancer, which measures the probability that individuals in the subpopulation would not have developed lung cancer if they had not smoked.

The posterior intervention causal effect of X on Y proposed by Zhao et al. [14] given the observed evidence O = o is defined as

PostICE ( X Y O = o ) = E ( Y Y X = 0 O = o ) .

Different from PostTCE, PostICE measures the change of Y ’s expectation if X is removed. When the observed evidence O = o contains X = 1 , we have PostTCE ( X Y X = 1 , ) = PostICE ( X Y X = 1 , ) . When the observed evidence O = o only contains Y = 1 , we have PostICE ( X Y Y = 1 ) = PN(X ⇒ Y ) pr ( X = 1 Y = 1 ) . Therefore, PostICE is different from PN. In disease diagnosis, PostICE considers not only the probability that the disease X is the cause of symptoms Y but also the posterior probability of the disease given the occurrence of symptoms.

Next, we extend the aforementioned case to the case of multiple effect variables Y = ( Y 1 , , Y q ) and multiple cause variables X = ( X 1 , , X p ) , and we define the posterior causal effect of simultaneously intervening on a subset of X on Y . In real applications, available evidence may include multiple observed effect variables, and thus, they can be used simultaneously to more accurately deduce the causes. For example, in medical diagnosis, the more symptoms of a patient are available, the more accurately a doctor can diagnose the patient’s disease. Without loss of generality, we assume that the causes are arranged in a topological order such that X l is not a cause of X k for k < l , and that Y 1 , , Y q subsequent to X are arranged in a topological order such that Y j is not a cause of Y m for m < j . Let g ( y ) be a known function weighting the importance of multiple effects in Y . For example, an additive weighting function is g ( Y ) = i = 1 q a i × Y i , where a i is a weight for Y i . Let X S denote a subvector of X , where S is the subset of indexes { 1 , , p } , and let x S 1 x S 0 denote x i 1 x i 0 for each i S . For a treated group of X S = x S 0 versus a control group of X S = 0 , where 0 denotes ( 0 , , 0 ) , we define the posterior total causal effect of multiple causes X S = x S on multiple effects Y = ( Y 1 , , Y q ) as

PostTCE [ X S ( x S ) Y O = o ] = E [ g ( Y x S ) g ( Y X S = 0 ) O = o ] .

Differing from the conditional counterfactual causal effect, defined by Zhao et al. [42], which restricts x S = ( 1 , , 1 ) , the aforementioned definition does not require this restriction. Comparing PostTCE of X S on Y with different values x S and x S , we can obtain various posterior controlled direct causal effects and interaction effects. For example, given the observed evidence o = ( X 1 = 1 , X 2 = 1 , X 3 = 1 , Y = 1 ) , a controlled direct causal effect of a set ( X 1 , X 2 ) on Y by controlling for X 3 = x 3 can be measured by

PostTCE [ X { 1 , 2 , 3 } ( 1 , 1 , x 3 ) Y O = o ] PostTCE [ X { 1 , 2 , 3 } ( 0 , 0 , x 3 ) Y O = o ] .

Comparing PostTCE across different subsets X S and X S , we contrast whether an event should be attributed more to X S or to X S . For example, given the observed evidence o = ( X 1 = 1 , X 2 = 1 , X 3 = 1 , Y = 1 ) , comparing PostTCE [ X 1 ( 1 ) Y O = o ] , PostTCE [ X 2 ( 1 ) Y O = o ] , and PostTCE [ X 3 ( 1 ) Y O = o ] , we can argue that the event Y = 1 should be attributed most to X 1 , X 2 , or X 3 .

It can be shown that for an additive weighting function g ( Y ) = j = 1 q a j × Y j ,

PostTCE [ X S ( x S ) Y O = o ] = j = 1 q a j × PostTCE [ X S ( x S ) Y j O = o ] .

In terms of PostTCE, we can do the attributions of multiple effects to multiple causes with interaction effects.

The posterior intervention causal effect of X S on Y is defined as

PostICE ( X S Y O = o ) = E [ g ( Y ) g ( Y X S = 0 ) O = o ] .

When the evidence O includes some ( Y j = 1 ) ’s, comparing PostICE ( X S Y O = o ) with PostICE ( X S Y O = o ) , we can retrospectively judge which of X S and X S might make the happened effects more likely. For g ( y 1 , , y q ) = j = 1 q y j , the posterior intervention causal effect PostICE ( X S Y O = o ) measures the expected number of outcomes eliminated by removing risk factors in X S .

3.2 Identification assumptions of posterior causal effects

Let ( X , Y ) = ( V 1 , , V p + q ) be arranged in a causal order and V r : s = ( V r , V r + 1 , , V s ) be a subvector of V for r s . Let ( V s ) v 1 : s 1 denote the potential outcome of V s if V 1 : s 1 were intervened to v 1 : s 1 To identify these posterior causal effects, we need to follow the monotonicity and no-confounding assumptions [13].

Assumption 1

(Monotonicity) For s = 2 , , p + q , the potential outcomes of V s satisfy the monotonicity relation: ( V s ) v 1 : s 1 * ( V s ) v 1 : s 1 whenever v 1 : s 1 * v 1 : s 1 .

This assumption is often expressed as “no prevention” in epidemiology and states that no individual can be helped by exposure to a risk factor. For example, let V 1 , V 2 , and V 3 denote poor diet, high blood pressure and stroke, respectively. The monotonicity assumption means that poor diet and high blood pressure are two potential risk factors for stroke. Exposures to them are not preventive for stroke, and a poor diet is also not preventive for high blood pressure. The validity of monotonicity cannot be directly tested, but this assumption imposes testable restrictions on the probability distribution of observed data in certain cases. Similar assumptions are often made in studies of imperfect compliance of treatment.

Assumption 2

(No confounding)

  1. There is no confounding between V s and V 1 : s 1 , i.e., ( V s ) v 1 : s 1 V 1 : s 1 for all v 1 : s 1 and s = 2 , , p + q ;

  2. The elements in { ( V s ) v 1 : s 1 } s = 2 p + q are mutually independent for any given v 1 : p + q 1 .

Assumption 2 (i) means that the potential outcomes of each variable are independent of its precedent variables arranged in the causal order. If V s has a causal structural model V s = f s ( V 1 : s 1 , ε s ) and an error variable ε s ε 1 : s 1 , then Assumption 2 is equivalent to the absence of latent confounders. Assumption 2 excludes the presence of unobserved confounders between variables X and Y . However, each variable X k may still confound the relationships between Y and X l or between X l and X s for where k < l , s . When there exists a set C containing observed background variables that are not influenced by X , the independence in Assumption 2 can be relaxed to those conditional on C .

When the evidence does not contain any effect variable Y i , the identification of posterior causal effects only requires Assumption 2 of no confounding. But when the evidence contains some effect variables, the identification of posterior causal effects requires both Assumptions 1 and 2. First, consider the case with a single X and a single Y . For the case of a single X and a single Y , Assumption 1 of monotonicity means Y X = 0 Y X = 1 , and PN has the following equation:

PN = pr ( Y = 1 X = 1 ) pr ( Y 0 = 1 X = 1 ) pr ( Y = 1 X = 1 ) .

The numerator is the treatment effect on treated. Under Assumption 2 of no confounding, we have pr ( Y 0 = 1 X = 1 ) = pr ( Y = 1 X = 0 ) , and thus, PN is identifiable. Similarly, under Assumptions 1 and 2 of monotonicity and no-confounding, it can be shown that the aforementioned posterior causal effects defined by intervening on a single cause X k are identifiable [13,14]. When simultaneously intervening on a set X S = x S of multiple causes, for the identification of PostTCE [ X S ( x S ) Y O = o ] , we further need the restriction on the relationship between x S and o . Let X S * = X S O , and x S * and x S * are the value of X S * in x S and o , respectively. Let X S = X S \ X S * , and x S is the value of X S . When q = 1 , i.e., Y = Y 1 , and the evidence contains the effect variable Y = y , we have the following theorem.

Theorem 1

Suppose that Assumptions 1 and 2 hold. PostTCE [ X S ( x S ) Y O = o ] is identifiable if one of the following conditions holds:

  1. x S * x S * and x S = ( 1 , , 1 ) ;

  2. x S * x S * and x S = ( 0 , , 0 ) .

For the case of q > 1 and that the evidence O includes Y k , let X O = O X and Y O = O { Y 1 , , Y k 1 } . The following equality holds from Zhao et al. [42]:

PostTCE [ X S ( x S ) Y k O = o ] = PostTCE [ X S ( x S ) Y k x O , y O , Y k = y ] .

Therefore, for an additive function g ( Y ) , we have

PostTCE [ X S ( x S ) Y O = o ] = j = 1 q a j × PostTCE [ X S ( x S ) Y j x O , y O , Y k = y ] .

When each item of the aforementioned equation is identifiable, PostTCE [ X S ( x S ) Y O = o ] is identifiable.

3.3 Relationship between posterior causal effect and population attributable risk

Greenland [10] pointed out that there are many incorrect equations regarding the probabilities of causation and the population attributable risks. The population attributable risks are used to measure the proportional amounts by which a disease risk would be reduced if risk factors were eliminated from a population [43]. For example, how much of the disease burden due to leukemia in a population could be eliminated if the exposures of benzene and ionizing radiation were eliminated from the population. In the following, we explain the relation of the posterior causal effects to the population attributable risks. For a case of a single Y and multiple causes X = ( X 1 , , X p ) , the population attributable risk is defined by Bruzzi et al. [44] as follows:

AR = pr ( Y = 1 ) pr ( Y = 1 X 1 = 0 , , X p = 0 ) pr ( Y = 1 ) .

It measures the proportional amount by which a disease risk would be reduced if all risk factors were eliminated from a population. Under Assumption 2 of no confounding and a weak monotonicity assumption that Y X = 0 Y X = x Y X = 1 for any x , the population attributable risk is equal to the posterior causal effect of multiple causes X on Y given the evidence of Y = 1 , i.e.,

(2) AR = PostTCE ( X Y Y = 1 ) = E ( Y X = 1 Y X = 0 Y = 1 ) .

AR does not measure how much the disease Y is attributed to a specified risk factor X k . The adjusted attributable risk for X k is defined by adjusting for the remaining risk factors X k = X \ { X k } [44]:

AR ( X k X k ) = pr ( Y = 1 ) x k pr ( Y = 1 X k = 0 , x k ) pr ( x k ) pr ( Y = 1 ) .

Note that the set X k should not contain any intermediate factor between X k and Y since eliminating X k can affect the intermediate factors in the condition X k . Let A k = ( X 1 , , X k 1 ) denote the variable set, which is prior to X k in a topological causal order, and thus, AR ( X k A k ) is a proper adjusted attributable risk. Lu et al. [13] showed that under Assumption 2 of no confounding and a weak monotonicity assumption Y X k = 0 Y X k = 1 , the attributable risk of X k on Y adjusted for the set A k is equal to the posterior total effect of X k on Y given the evidence of Y = 1 :

(3) AR ( X k A k ) = PostTCE ( X k Y Y = 1 ) .

Equations (2) and (3) show the relationships between the posterior causal effects and the population and adjusted attributable risks, and they explain the causal meaning of the attributable risks in terms of the potential outcome framework. The equations also give other identification equations of posterior causal effects PostTCE ( X Y Y = 1 ) and PostTCE ( X k Y Y = 1 ) under the weaker monotonicity assumptions than Assumption 1 of monotonicity.

3.4 Diagnostic approaches based on Bayesian posterior probabilities and posterior causal effects

In the following, we discuss the problem about whether the medical diagnosis should be based on Bayesian posterior probabilities or the posterior causal effects. Bayesian posterior probabilities measure the uncertainty of past events given the later observed evidence, but they do not capture the causal relationships between these events. Posterior causal effects, on the other hand, measure the uncertainty of past events that have causal effects on the later happened evidence.

In the field of medical diagnosis, a probabilistic expert system based on Bayesian posterior probabilities was developed by Lauritzen and Spiegelhalter [45] and Spiegelhalter et al. [46]. This system computes the posterior probabilities of diseases given the observed symptoms, which depend on the prior probabilities of the diseases. The diagnosis based on the maximum posterior probability minimizes the misdiagnosis error [47]. However, this approach does not account for the causal relationships between diseases and symptoms.

As Encyclopaedia Britannica [48] defines, the diagnostic process is the method by which health professionals select one disease over another, identifying one as the most likely cause of a person’s symptoms. Richens et al. [49] pointed out that most existing diagnostic algorithms, including Bayesian model-based and deep learning methods, rely on associative inference, and they identified diseases based on how correlated they are with a patient’s symptoms and medical history. This contrasts with how doctors perform medical diagnosis, selecting the diseases that offer the best causal explanations for the patient’s symptoms. They argued that disease diagnostic reasoning should satisfy three principles concerning not only the posterior probability, but also causality and simplicity. They proposed an approach based on a noisy-operation model, but their model is restricted to the case where neither diseases nor symptoms can affect each other. Li et al. [14] proposed a medical diagnostic approach based on posterior intervention causal effects PostICE, which satisfies the aforementioned principles for medical diagnosis. For disease diagnosis, the evidence O contains some symptoms and backgrounds of a patient, but does not contain the status of diseases X k . Since PostICE ( X k Y O = o , X k = 0 ) = 0 , we can obtain

(4) PostICE ( X k Y o ) = PostICE ( X k Y o , X k = 1 ) × pr ( X k = 1 o ) .

This equation means that the diagnostic approach based on PostICE considers not only Bayesian posterior probability pr ( X k = 1 o ) , but also the posterior causal effect of the disease X k on the symptoms Y in the subpopulation with the disease X k = 1 , which is ignored by the approach based on Bayesian posterior probability. For a patient given the evidence O = o , we diagnose the patient with a disease X k , which has the largest value in { PostICE ( X j Y o ) , j } . It means that the number of symptoms could be eliminated at most if the patient had not gotten the disease X k .

When a patient may have multiple diseases simultaneously, Bayesian approach diagnoses the patient with multiple diseases based on the maximum posterior probabilities pr ( X = x O = o ) . The approach based on posterior causal effects uses PostICE ( X S Y O = o ) for diagnosis. For a patient given the evidence O = o , we diagnose the patient with multiple diseases X S , which has the largest value in { PostICE ( X S Y o ) , S { 1 , , p } } . Bayesian posterior probabilities and posterior intervention causal effects have the following equation:

PostICE ( X S Y O = o ) = x s PostICE ( X S Y X S = x s , O = o ) × pr ( X S = x s O = o ) .

If the diagnostic result is used further for eliminating the symptoms in the future, the Bayesian diagnostic approach may not be optimal, and the diagnostic approach based on posterior causal effects requires an assumption of invariance maybe require the following assumption of invariant relationships between causes and effects in the past and the future.

Assumption 3

(Invariance) Let W and Z denote the future diseases and symptoms, respectively. The potential outcomes of the past and future symptoms are the same (i.e., Y x = Z w ) if the statuses of the past and future diseases are the same (i.e., X = W ).

This assumption can be weakened to that the average causal effects of the diseases on symptoms are invariant across time in the subpopulations of O = o , i.e., E ( Y X = 1 Y X = 0 O = o ) = E ( Z W = 1 Z W = 0 O = o ) . This invariance assumption may hold for many real scenarios, e.g., in a specified room, X denotes a switch on or off, and Y a light on or not. But the assumption may not hold in some real scenarios, e.g., X denotes that Jack drank poison and Y = 1 denotes that he died.

In the following, we use a numerical example of medical diagnosis to compare the approach based on Bayesian posterior probabilities with that based on the posterior causal effects.

Figure 2 
                  Causal diagram of two diseases 
                        
                           
                           
                              
                                 
                                    X
                                 
                                 
                                    1
                                 
                              
                           
                           {X}_{1}
                        
                      and 
                        
                           
                           
                              
                                 
                                    X
                                 
                                 
                                    2
                                 
                              
                           
                           {X}_{2}
                        
                      and a symptom 
                        
                           
                           
                              Y
                           
                           Y
                        
                     .
Figure 2

Causal diagram of two diseases X 1 and X 2 and a symptom Y .

Figure 3 
                  Misclassification probabilities of two approaches.
Figure 3

Misclassification probabilities of two approaches.

Figure 4 
                  PostICEs of two approaches.
Figure 4

PostICEs of two approaches.

Example 1

Let X 1 and X 2 denote two diseases and Y a symptom caused by X 1 . Consider the causal mechanism described by the diagram in Figure 2, where X 1 is the cause of disease X 2 and symptom Y , but X 2 is not the cause of Y . Suppose that the causal diagram has the following probabilities:

pr ( X 1 = 1 ) = 0.400 , pr [ ( X 2 ) x 1 = 0 = 1 ] = 0.550 , pr [ ( X 2 ) x 1 = 1 = 1 ] = 0.622 , pr ( Y x 1 = 0 = 1 ) = 0.401 , pr ( Y x 1 = 1 = 1 ) = 0.500 .

Thus, the observed variables X 2 and Y are generated by

X 2 = X 1 × ( X 2 ) x 1 = 1 + ( 1 X 1 ) × ( X 2 ) x 1 = 0 , Y = X 1 × Y x 1 = 1 + ( 1 X 1 ) × Y x 1 = 0 .

From the probabilities, we can obtain the posterior probabilities and causal effects given the symptom Y = 1 :

pr ( X 1 = 0 , X 2 = 0 Y = 1 ) = 0.246 , pr ( X 1 = 0 , X 2 = 1 Y = 1 ) = 0.300 , pr ( X 1 = 1 , X 2 = 0 Y = 1 ) = 0.171 , pr ( X 1 = 1 , X 2 = 1 Y = 1 ) = 0.283 , PostICE [ ( X 1 , X 2 ) Y Y = 1 ] = 0.272 , PostICE ( X S = ø Y Y = 1 ) = 0.000 , PostICE ( X 1 Y Y = 1 ) = 0.272 , PostICE ( X 2 Y Y = 1 ) = 0.000 .

For Bayesian approach based on posterior probabilities, the diagnostic results based on the maximum joint and marginal posterior probabilities of pr ( x 1 , x 2 Y = 1 ) and pr ( x k Y = 1 ) are ( X 1 , X 2 ) = ( 0 , 1 ) and ( X 2 = 1 ) , respectively. Neither of these results identifies the true cause X 1 of symptom Y . In contrast, by the maximum intervention posterior causal effects PostICE ( X S Y Y = 1 ) , the diagnostic results are X S = ( X 1 , X 2 ) = ( 1 , 1 ) and X S = X 1 = 1 , respectively, since they have the maximum value of 0.182. Either diagnostic result finds the true cause X 1 of symptom Y , and by simplicity, the diagnosis prefers X 1 .

In the following, we first compare the misclassification probabilities of the two diagnostic approaches. The diagnostic results of Bayesian approach are ( X 1 , X 2 ) = ( 0 , 1 ) and ( X 2 = 1 ) . Overall, we consider the diagnostic result to be taking on the disease X 2 , so the individuals without the disease X 2 ( X 2 = 0 ) , which include individuals with ( X 1 = 1 , X 2 = 0 ) and ( X 1 = 0 , X 2 = 0 ) , are misclassified. Thus, the misclassification probability of Bayesian diagnostic approach for the population of Y = 1 is

pr [ ( X 1 , X 2 ) = ( 0 , 0 ) Y = 1 ] + pr [ ( X 1 , X 2 ) = ( 1 , 0 ) Y = 1 ] = pr ( X 2 = 0 Y = 1 ) = 0.417 .

The diagnostic results of posterior causal effect approach are ( X 1 , X 2 ) = ( 1 , 1 ) and ( X 1 = 1 ) . We consider the diagnostic result to be taking on the disease X 1 , and the individuals without the disease X 1 ( X 1 = 0 ) , which include individuals with ( X 1 = 0 , X 2 = 0 ) and ( X 1 = 0 , X 2 = 1 ) , are misclassified. Thus, the misclassification probability of diagnostic approach based on posterior causal effects for the population of Y = 1 is

pr [ ( X 1 , X 2 ) = ( 0 , 0 ) Y = 1 ] + pr [ ( X 1 , X 2 ) = ( 0 , 1 ) Y = 1 ] = pr ( X 1 = 0 Y = 1 ) = 0.546 .

Bayesian diagnostic approach has a lower misclassification probability than the posterior causal effect approach.

Furthermore, we vary the prior probability pr ( X 1 = 1 ) of disease X 1 , and we show the misclassification probabilities of the two diagnostic approaches in Figure 3, where the solid line is the misclassification probability for the posterior causal effect approach and the dotted line is that for Bayesian approach. It can be seen that as pr ( X 1 = 1 ) increases, the misclassification probability of the posterior causal effect approach decreases since it always diagnoses patients with disease X 1 = 1 . For a lower prior probability pr ( X 1 = 1 ) , Bayesian approach diagnoses patients with the disease X 2 = 1 , and its misclassification probability is lower than that of the posterior causal effect approach. When pr ( X 1 = 1 ) increases to a certain extent, Bayesian approach changes the diagnosis X 2 = 1 to X 1 = 1 , and thus, it has the same misclassification probability as the posterior causal effect approach.

Next, we compare the causal effect of treating the diagnosed disease on the elimination of symptoms. To evaluate the causal effects of the treatment after diagnosis, we make Assumption 3 of invariance. Let X k denote the diagnosed disease, and then, the posterior intervention causal effect PostICE ( X k Y Y = 1 ) measures the elimination of symptoms attributed to the intervention on X k . PostICE ( X k Y Y = 1 ) s for the two diagnostic approaches are shown in Figure 4. The posterior causal effect approach always diagnoses patients with disease X 1 = 1 and its PostICE ( X 1 Y Y = 1 ) increases as pr ( X 1 = 1 ) increases. Bayesian approach diagnoses the patients with disease X 2 = 1 for a lower pr ( X 1 = 1 ) , and PostICE ( X 2 Y Y = 1 ) = 0 , which is much less than PostICE ( X 1 Y Y = 1 ) obtained by the posterior causal effect approach. When pr ( X 1 = 1 ) increases to a certain extent, Bayesian approach changes the diagnostic result to disease X 1 = 1 , and then, it has the same value of PostICE ( X 1 Y Y = 1 ) as that of the posterior causal effect approach.

In this example, the posterior causal effect approach represents a white-box method with complete knowledge of the causal mechanisms. It would never diagnose patients with symptom Y = 1 as suffering from disease X 2 , which is non-causative of the symptom. The Bayesian posterior probability approach can be viewed as a black-box method without any knowledge of the underlying causal mechanisms. Although it may diagnose some patients with symptom Y = 1 as having disease X 2 despite its non-causal relationship, this approach has the minimum probability of misdiagnosis overall.

To diagnose possible diseases, regardless of whether or not they are the causes of the occurred symptoms, Bayesian diagnosis based on posterior probabilities always has the minimum misclassification probability [47]. One drawback of the Bayesian diagnostic approach is that it may not identify the causes of occurred symptoms. To identify the causes, we argue that the approach based on posterior causal effects is a better choice. In the aforementioned numerical example, we assume that the probabilities of the causal mechanism are known. A limitation of the approach based on posterior causal effects is that the identifiability of the posterior causal effects requires Assumptions 1 and 2 of monotonicity and no confounding. Under Assumption 1 of monotonicity, patients with symptom Y = 1 are always diagnosed with certain diseases and are not diagnosed with no disease since PostICE ( X S = Y Y = 1 ) = 0 has the least value.

4 Discussion

Prospective and retrospective causal inferences investigate causality from different perspectives. Prospective inference reasons forward from causes to effects. Randomized experiments represent the gold standard for prospective causal inference. In contrast, retrospective inference works backward from observed outcomes to infer their potential causes. However, there is currently no established gold standard methodology for retrospective causal analysis. Posterior causal effects can be utilized for prospective causal inference when the available evidence lacks known outcome details. For example, let X denote smoking and Y denote lung cancer. The average treatment effect on the treated, E ( Y X = 1 Y X = 0 X = 1 ) , evaluates the causal influence of smoking on lung cancer risk in an exposed population. However, it cannot definitively conclude whether smoking causes lung cancer. Conversely, posterior causal effects are employed in retrospective causal analysis when the evidence includes known outcome information. The posterior causal effect E ( Y X = 1 Y X = 0 X = 1 , Y = 1 ) assesses the causal impact of smoking on lung cancer in the subpopulation of smokers diagnosed with lung cancer. It estimates the probability that lung cancer patients in this group would not have developed lung cancer had they not smoked. Similarly, the posterior causal effect E ( Y X = 1 Y X = 0 Y = 1 ) evaluates the causal effect of smoking on lung cancer in the overall lung cancer patient population. It gauges the probability that these patients would not have had lung cancer without smoking. Thus, posterior causal effects can quantify the attributable risks of smoking within specific patient groups.

Confounding poses a challenge in causal inference, as identifying confounders is difficult using only observational data. Without untestable assumptions, we cannot definitively determine if a covariate is a confounder. The surrogate paradox further demonstrates that probabilistic causal effects are generally non-transitive. Specifically, it shows that the signs or directions of causal impacts cannot be logically deduced from the probabilistic outputs of causal analyses. In other words, logistic reasoning does not necessarily apply to the probabilistic results of causal inference.

For a diagnostic problem, Bayesian posterior probability approach may minimize misclassification probability, while the posterior causal effect approach may identify the causes of occurred symptoms. The suitable diagnostic method depends on whether the goal is to predict diseases or uncover the causes of presented symptoms, or even eliminate those symptoms.

Similar to posterior probabilities, posterior causal effects also derive from Bayesian thinking and use potential outcome framework. In retrospective causal analysis, potential outcomes are essential for expressing counterfactual scenarios. However, the posterior causal effects differ from the Bayesian posterior distributions of causal effects. The posterior causal effects are the expectations of the causal effects conditional on the observed evidence, rather than posterior distributions.

Causal graphs may be learned from observed data, which depict causal relationships among variables in a population. But a causal graph cannot be used to deduce the causes of effects for a specific individual because different individuals in the population may have different causes that depend on the evidence.

Many open questions remain regarding retrospective causal inference. It shares numerous topics with prospective causal inference, but may also involve unique considerations specific to reasoning backward from effects to causes. For a given set of evidence about occurred outcomes, there can be different conceptual causes depending on the research objective. In some real-world applications, identifying the root causes of effects may be of greater interest.

Acknowledgement

We would like to thank the editors and the three reviewers for their very helpful and valuable comments, which led to a significant improvement of this manuscript.

  1. Funding information: This research was partially supported by the National Natural Science Foundation of China (No. 12071015), the Disciplinary Funding of Beijing Technology and Business University (No. 50500101002), the Joint Key Research Project funded by the Beijing Municipal Education Commission and the Beijing Municipal Natural Science Foundation (No. 23JA0006), the Research Foundation for Advanced Talents of Beijing Technology and Business University (No. 19008024084), and a joint research project of Alibaba group.

  2. Author contributions: All authors have accepted responsibility for the entire content of this manuscript and approved its submission.

  3. Conflict of interest: The authors state no conflict of interest.

References

[1] Neyman JS. On the application of probability theory to agricultural experiments. Stat Sci. 1923;5:465–80 (1990). http://www.jstor.org/stable/2245382. Suche in Google Scholar

[2] Rubin DB. Estimating causal effects of treatments in randomized and nonrandomized studies. J Educ Psychol. 1974;66:688–701. 10.1037/h0037350. Suche in Google Scholar

[3] Halpern JY. Actual causality. London: MIT Press; 2016. 10.7551/mitpress/10809.001.0001Suche in Google Scholar

[4] Pearl J, Mackenzie D. The book of why: the new science of cause and effect. New York: Basic Books; 2018. Suche in Google Scholar

[5] Dawid AP, Faigman DL, Fienberg SE. Fitting science into legal contexts: Assessing effects of causes or causes of effects? Soc Meth Res. 2014;43(3):359–90. 10.1177/0049124113515188. Suche in Google Scholar

[6] Holland PW. Statistics and causal inference. J Amer Statist Assoc. 1986;81(396):945–60. 10.1080/01621459.1986.10478354. Suche in Google Scholar

[7] Dawid AP, Musio M. Effects of causes and causes of effects. Annu Rev Stat Appl. 2022;9:261–87. 10.1146/annurev-statistics-070121-061120. Suche in Google Scholar

[8] Dawid AP, Faigman DL, Fienberg SE. On the causes of effects: Response to Pearl. Soc Meth Res. 2015;44(1):165–74. 10.1177/0049124114562613. Suche in Google Scholar

[9] Robins J, Greenland S. The probability of causation under a stochastic model for individual risk. Biometrics. 1989;45:1125–38. 10.2307/2531765. Suche in Google Scholar

[10] Greenland S. Relation of probability of causation to relative risk and doubling dose: a methodologic error that has become a social problem. Am J Public Health. 1999;89(8):1166–9. 10.2105/AJPH.89.8.1166. Suche in Google Scholar

[11] Pearl J. Probabilities of causation: three counterfactual interpretations and their identification. Synthese. 1999;121:93–149. 10.1023/A:1005233831499. Suche in Google Scholar

[12] Dawid AP. Causal inference without counterfactuals. J Amer Statist Assoc. 2000;95(450):407–24. 10.2307/2669377. Suche in Google Scholar

[13] Lu Z, Geng Z, Li W, Zhu S, Jia J. Evaluating causes of effects by posterior effects of causes. Biometrika. 2023;110(2):449–65. 10.1093/biomet/asac038. Suche in Google Scholar

[14] Li W, Lu Z, Jia J, Xie M, Geng Z. Retrospective causal inference with multiple effect variables. Biometrika. 2024;111(2):573–89. 10.1093/biomet/asad056. Suche in Google Scholar

[15] Miettinen OS, Cook EF. Confounding: essence and detection. Am J Epidemiol. Oct 1981;114(4):593–603. 10.1093/oxfordjournals.aje.a113225. Suche in Google Scholar PubMed

[16] Boivin JF, Wacholder S. Conditions for confounding of the risk ratio and of the odds ratio. Am J Epidemiol. 1985;121(1):152–8. 10.1093/oxfordjournals.aje.a113977. Suche in Google Scholar PubMed

[17] Grayson D. Confounding confounding. Am J Epidemiol. 1987;126(3):546–53. 10.1093/oxfordjournals.aje.a114687. Suche in Google Scholar PubMed

[18] Greenland S, Holland PW, Mantel N, Wickramaratne PJ, Holford TR. Confounding in epidemiologic studies. Biometrics. 1989;45(4):1309–22. 10.2307/2531783. Suche in Google Scholar

[19] Weinberg CR. Toward a clearer definition of confounding. Am J Epidemiol. 1993;137(1):1–8. 10.1093/oxfordjournals.aje.a116591. Suche in Google Scholar PubMed

[20] Greenland S, Pearl J, Robins JM. Causal diagrams for epidemiologic research. Epidemiology. 1999;10:37–48. https://www.jstor.org/stable/3702180. 10.1097/00001648-199901000-00008Suche in Google Scholar

[21] Greenland S, Pearl J, Robins JM. Confounding and collapsibility in causal inference. Statist Sci. 1999;14:29–46. 10.1214/ss/1009211805. Suche in Google Scholar

[22] Wickramaratne PJ, Holford TR. Confounding in epidemiologic studies: the adequacy of the control group as a measure of confounding. Biometrics. 1987;43(4):751–65. http://www.jstor.org/stable/2531530. 10.2307/2531530Suche in Google Scholar

[23] Kleinbaum DG, Kupper LL, Morgenstern H. Epidemiologic research: principles and quantitative methods. New York: Van Nostrand Reinhold; 1982.Suche in Google Scholar

[24] Greenland S, Robins JM. Identifiability, exchangeability, and epidemiological confounding. Int J Epidemiol. 1986;15:413–9. 10.1093/ije/15.3.413. Suche in Google Scholar PubMed

[25] Geng Z, Guo J, Fung WK. Criteria for confounders in epidemiological studies. J R Stat Soc Ser B (Stat Methodol). 2002;64(1):3–15. 10.1111/1467-9868.00321. Suche in Google Scholar

[26] VanderWeele TJ, Shpitser I. On the definition of a confounder. Ann Statist. 2013;41(1):196–220. 10.1214/12-aos1058. Suche in Google Scholar PubMed PubMed Central

[27] Fleming TR, Demets DL. Surrogate end points in clinical trials: Are we being misled? Ann Intern Med. 1996;125(7):605–13. 10.7326/0003-4819-125-7-199610010-00011. Suche in Google Scholar PubMed

[28] Chen H, Geng Z, Jia J. Criteria for surrogate end points. J R Stat Soc Ser B (Stat Methodol). 2007;69(5):919–32. 10.1111/j.1467-9868.2007.00617.x. Suche in Google Scholar

[29] Jiang Z, Ding P, Geng Z. Qualitative evaluation of associations by the transitivity of the association signs. Statist Sinica. 2015;25(3):1065–79. http://www.jstor.org/stable/24721221. Suche in Google Scholar

[30] Prentice RL. Surrogate endpoints in clinical trials: definition and operational criteria. Stat Med. 1989;8(4):431–40. 10.1002/sim.4780080407. Suche in Google Scholar PubMed

[31] Dawid AP. Conditional independence in statistical theory. J R Stat Soc Ser B (Stat Methodol). 1979;41(1):1–15. 10.1111/j.2517-6161.1979.tb01052.x. Suche in Google Scholar

[32] Frangakis CE, Rubin DB. Principal stratification in causal inference. Biometrics. 2002;58:21–9. 10.1111/j.0006-341X.2002.00021.x. Suche in Google Scholar

[33] Lauritzen S. Discussion on causality. Scand J Stat. 2004;31(2):189–93. 10.1111/j.1467-9469.2004.03-200A.x. Suche in Google Scholar

[34] Jiang Z, Ding P, Geng Z. Principal causal effect identification and surrogate end point evaluation by multiple trials. J R Stat Soc Ser B Stat Meth. Nov 2016;78(4):829–48. 10.1111/rssb.12135. Suche in Google Scholar

[35] Moore T. Deadly medicine: why tens of thousands of patients died in America’s worst drug disaster. New York: Simon & Schuster; 1995. Suche in Google Scholar

[36] Investigators CASTC. Preliminary report: effect of encainide and flecainide on mortality in a randomized trial of arrhythmia suppression after myocardial infarction. The Cardiac Arrhythmia Suppression Trial (CAST) Investigators. New Engl J Med. 1989;321:406–12. 10.1056/NEJM198908103210629. Suche in Google Scholar PubMed

[37] Ju C, Geng Z. Criteria for surrogate end points based on causal distributions. J R Stat Soc Ser B (Stat Methodol). 2010;72(1):129–42. 10.1111/j.1467-9868.2009.00729.x. Suche in Google Scholar

[38] Wu Z, He P, Geng Z. Sufficient conditions for concluding surrogacy based on observed data. Stat Med. 2011;30(19):2422–34. 10.1002/sim.4273. Suche in Google Scholar PubMed

[39] Joffe M. Discussion on “Surrogate measures and consistent surrogates”. Biometrics. 2013;69(3):572–5. https://www.jstor.org/stable/24538121. 10.1111/biom.12074Suche in Google Scholar PubMed

[40] Luo P, Cai Z, Geng Z. Criteria for multiple surrogates. Statist Sinica. 2019;29(3):1343–66. https://www.jstor.org/stable/26706005. 10.5705/ss.202017.0122Suche in Google Scholar

[41] Pearl J. Probabilities of Causation: Three Counterfactual Interpretations and Their Identification. 1st ed. New York: Association for Computing Machinery; 2022. p. 317–72. 10.1145/3501714.3501735. Suche in Google Scholar

[42] Zhao R, Zhang L, Zhu S, Lu Z, Dong Z, Zhang C, et al. Conditional counterfactual causal effect for individual attribution. In: Proceedings of the Thirty-Ninth Conference on Uncertainty in Artificial Intelligence. vol. 216 of Proceedings of Machine Learning Research. PMLR; 2023. p. 2519–28. https://proceedings.mlr.press/v216/zhao23a.html. Suche in Google Scholar

[43] Rockhill B, Newman B, Weinberg C. Use and misuse of population attributable fractions. Am J Public Health. 1998;88(1):15–9. 10.2105/ajph.88.1.15. Suche in Google Scholar PubMed PubMed Central

[44] Bruzzi P, Green SB, Byar DP, Brinton LA, Schairer C. Estimating the population attributable risk for multiple risk factors using case-control data. Am J Epidemiol. 1985;122(5):904–14. 10.1093/oxfordjournals.aje.a114174. Suche in Google Scholar PubMed

[45] Lauritzen SL, Spiegelhalter DJ. Local computations with probabilities on graphical structures and their application to expert systems. J R Stat Soc Ser B (Stat Methodol). 1988;50(2):157–94. 10.1111/j.2517-6161.1988.tb01721.x. Suche in Google Scholar

[46] Spiegelhalter DJ, Dawid AP, Lauritzen SL, Cowell RG. Bayesian analysis in expert systems. Statist Sci. 1993;8(3):219–47. https://www.jstor.org/stable/2245959. 10.1214/ss/1177010888Suche in Google Scholar

[47] Berger JO. Statistical decision theory and Bayesian analysis. New York: Springer Science & Business Media. 2013. Suche in Google Scholar

[48] Rakel RE. Diagnosis. 2023. https://www.britannica.com/science/diagnosis. Suche in Google Scholar

[49] Richens JG, Lee CM, Johri S. Improving the accuracy of medical diagnosis with causal machine learning. Nat Commun. 2020;11(1):1–9. 10.1038/s41467-020-17419-7. Suche in Google Scholar PubMed PubMed Central

Received: 2023-09-30
Revised: 2024-05-19
Accepted: 2024-06-28
Published Online: 2024-10-24

© 2024 the author(s), published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Artikel in diesem Heft

  1. Research Articles
  2. Evaluating Boolean relationships in Configurational Comparative Methods
  3. Doubly weighted M-estimation for nonrandom assignment and missing outcomes
  4. Regression(s) discontinuity: Using bootstrap aggregation to yield estimates of RD treatment effects
  5. Energy balancing of covariate distributions
  6. A phenomenological account for causality in terms of elementary actions
  7. Nonparametric estimation of conditional incremental effects
  8. Conditional generative adversarial networks for individualized causal mediation analysis
  9. Mediation analyses for the effect of antibodies in vaccination
  10. Sharp bounds for causal effects based on Ding and VanderWeele's sensitivity parameters
  11. Detecting treatment interference under K-nearest-neighbors interference
  12. Bias formulas for violations of proximal identification assumptions in a linear structural equation model
  13. Current philosophical perspectives on drug approval in the real world
  14. Foundations of causal discovery on groups of variables
  15. Improved sensitivity bounds for mediation under unmeasured mediator–outcome confounding
  16. Potential outcomes and decision-theoretic foundations for statistical causality: Response to Richardson and Robins
  17. Quantifying the quality of configurational causal models
  18. Design-based RCT estimators and central limit theorems for baseline subgroup and related analyses
  19. An optimal transport approach to estimating causal effects via nonlinear difference-in-differences
  20. Estimation of network treatment effects with non-ignorable missing confounders
  21. Double machine learning and design in batch adaptive experiments
  22. The functional average treatment effect
  23. An approach to nonparametric inference on the causal dose–response function
  24. Review Article
  25. Comparison of open-source software for producing directed acyclic graphs
  26. Special Issue on Neyman (1923) and its influences on causal inference
  27. Optimal allocation of sample size for randomization-based inference from 2K factorial designs
  28. Direct, indirect, and interaction effects based on principal stratification with a binary mediator
  29. Interactive identification of individuals with positive treatment effect while controlling false discoveries
  30. Neyman meets causal machine learning: Experimental evaluation of individualized treatment rules
  31. From urn models to box models: Making Neyman's (1923) insights accessible
  32. Prospective and retrospective causal inferences based on the potential outcome framework
  33. Causal inference with textual data: A quasi-experimental design assessing the association between author metadata and acceptance among ICLR submissions from 2017 to 2022
  34. Some theoretical foundations for the design and analysis of randomized experiments
Heruntergeladen am 7.9.2025 von https://www.degruyterbrill.com/document/doi/10.1515/jci-2023-0063/html
Button zum nach oben scrollen