Home Life Sciences A novel characterization of the generalized family wise error rate using empirical null distributions
Article
Licensed
Unlicensed Requires Authentication

A novel characterization of the generalized family wise error rate using empirical null distributions

  • Jeffrey C. Miecznikowski EMAIL logo and Daniel P. Gaile EMAIL logo
Published/Copyright: March 14, 2014

Abstract

We present a novel characterization of the generalized family wise error rate: kFWER. The interpretation allows researchers to view kFWER as a function of the test statistics rather than current methods based on p-values. Using this interpretation we present several theorems and methods (parametric and non-parametric) for estimating kFWER in various data settings. With this version of kFWER, researchers will have an estimate of kFWER in addition to knowing what tests are significant at the estimated kFWER. Additionally, we present methods that use empirical null distributions in place of parametric distributions in standard p-value kFWER controlling schemes. These advancements represent an improvement over common kFWER methods which are based on parametric assumptions and merely report the tests that are significant under a given value for kFWER.


Corresponding authors: Jeffrey C. Miecznikowski, Department of Biostatistics, SUNY University at Buffalo, 433 Kimball Tower, 3435 Main St., Buffalo, NY 14214, USA, Tel.: 716.881.8953, e-mail: ; and Daniel P. Gaile, Department of Biostatistics, SUNY University at Buffalo, 706 Kimball Tower, 3435 Main St., Buffalo, NY 14214, USA, Tel.: 716.829.2756, e-mail:

Acknowledgments

The authors are very grateful to David Tritchler for sharing his insights and providing helpful comments on an earlier version of this article.

Appendix

Monotonicity of the kFWER estimators

The proposed kFWER estimator is defined as a function of a set Z which is usually an interval. For the left hand side (LHS) kFWER, for a point z, we define Z=(,z), while for the right hand side (RHS) kFWER, we define Z=(z,). In short, the parametric and empirical (LHS or RHS) kFWER estimates are monotone. To see this we can re-examine either equations (21) and (22). Without loss of generality, consider (21) restated here:

(35)kFWER¯(Z)=η=kN(1FB(k1|η,F0¯(Z)))fB(η|N,π0¯). (35)

The only term in (35) involving z or LHS version Z=(,z) is (1FB(k1|η,F0¯(Z))), which is monotone (non-decreasing) as a function of z, hence kFWER¯(Z) is monotone (non-decreasing) as a function of z. A similar argument will show that the (RHS) kFWER is monotone (non-increasing) as a function of z.

The bias of kFWER¯(Z)

Under certain constraints (zero assumption and large π0) we can show E(kFWER¯(Z))=kFWER(Z). To see this, we examine the convexity of kFWER¯(Z) as a function of π0. We assume that F0 is given and so π0 is the only parameter to estimate in kFWER(Z). To emphasize this dependence on π0 we will denote our estimate kFWER¯(Z) as kFWER¯π0(Z). According to (21) in our manuscript we have

(36)kFWER¯π0(Z)=η=kN(1FB(k1|η,F¯0(Z)))fB(η|N,π0). (36)

Figure 8 shows (36) as a function of π0 for certain fixed values of k, N, and F¯0(Z). As shown kFWER¯ appears nearly linear (and constant) for large values of N and π0. This can be further confirmed by examining the second derivative as a function of π0. Computing the second derivative of kFWERπ0(Z) as a function of π0 we have

Figure 8 kFWER¯π0(Z)${\overline {kFWER} _{{\pi _0}}}({\cal Z})$ for various values of N, k, and F0(Z).${F_0}({\cal Z}).$ For large N, the function appears linear and nearly constant for large values of π0.
Figure 8

kFWER¯π0(Z) for various values of N, k, and F0(Z). For large N, the function appears linear and nearly constant for large values of π0.

(37)kFWER¯π0(Z)=η=kN[(Nη)(1FB(k1|η,F0(Z)))]((η1)ηπ0η2(1π0)Nη2(Nη)(1π0)Nη1ηπ0η1+π0η(Nη)(Nη1)(1π0)Nη2) (37)

The second derivative in (37) also depends on k, N, and F0(Z). In Figure 9 we explore the second derivative as a function of π0 and these other variables. We see that for large values of N and π0 the second derivative is near 0 confirming that kFWER is nearly linear. If we assume linearity for large π0, by Jensen’s inequality we have that,

Figure 9 Second Derivative of kFWER¯π0(Z):${\overline {kFWER} _{{\pi _0}}}({\cal Z}):$ for various values of N, k, and F0(Z).${F_0}({\cal Z}).$ The second derivative is nearly zero for large values of N and k suggesting linearity of the kFWER¯$\overline {kFWER} $ estimator.
Figure 9

Second Derivative of kFWER¯π0(Z): for various values of N, k, and F0(Z). The second derivative is nearly zero for large values of N and k suggesting linearity of the kFWER¯ estimator.

(38)E(kFWER¯π0¯(Z))=kFWERE(π0¯)(Z) (38)
(39)=kFWERπ0(Z) (39)
(40)=kFWER(Z) (40)

Note, we assume E(π0¯)=π0. Since F0 is known, we have by definition π0¯=N+(A0)/(NF0(A0)) and so

(41)E(π0¯)=E(N+(A0)NF0(A0)) (41)
(42)=1NF0(A0)E(N+(A0)) (42)
(43)=1NF0(A0)π0NF0(A0)(byzeroassumption) (43)
(44)=π0 (44)

We can also examine kFWER^(Z) with the same assumptions. That is, we assume the following:

  • each zi follows the two group model independently,

  • the zero assumption where f1(z) is near zero for a subset of the sample space near zero, say A0,

  • and that F0 is the standard normal CDF.

Since we are using the nonparametric form kFWER^(Z), we assume that we need to estimate F0 (non parametrically). We use 25,000 Monte Carlo simulations to study the performance of the MLE method to estimate F0 and subsequently π0. The bias of kFWER^π^0,F0^ is shown in Figure 10(A)–(C). Note the bias depends on the direction of significance for the non-null genes. This is reasonable since the direction of the non-null genes will skew F0^ in that direction. That is, if the non-null genes have positive z-values, F0^ will have a positive mean and thus the RHS version of kFWER^ will be an overestimate while the LHS version of kFWER^ will be an underestimate (Figure 10(A)). The opposite will occur with non-null genes with negative z-values (Figure 10(B)). With an equal balance of negative and positive z-scores, F0^ will be similar to a N(0, 1) CDF and thus kFWER^ will be similar kFWER¯, that is, a somewhat conservative estimate of kFWER (Figure 10(C)).

Figure 10 Bias: The bias in kFWER^(Z)$\widehat {kFWER}({\cal Z})$ when assuming k=1, N=500 and π0=0.95. The dotted line is kFWER^(Z)$\widehat {kFWER}({\cal Z})$ with the LHS version for x<0 and the RHS version for x≥0 with the non null genes possessing large positive z-scores (a), large negative z-scores (b) and a (roughly) equal mixture of large positive and negative z-scores (c). The black line is the true kFWER. The bias of kFWER^(Z)$\widehat {kFWER}({\cal Z})$ depends on the direction of the non-null z-scores with kFWER^(Z)$\widehat {kFWER}({\cal Z})$ being conservative when there is roughly an equal number of non-null genes in each direction.
Figure 10

Bias: The bias in kFWER^(Z) when assuming k=1, N=500 and π0=0.95. The dotted line is kFWER^(Z) with the LHS version for x<0 and the RHS version for x≥0 with the non null genes possessing large positive z-scores (a), large negative z-scores (b) and a (roughly) equal mixture of large positive and negative z-scores (c). The black line is the true kFWER. The bias of kFWER^(Z) depends on the direction of the non-null z-scores with kFWER^(Z) being conservative when there is roughly an equal number of non-null genes in each direction.

Sensitivity of estimators to A0 and π0

For our simulations we used the program locfdr default choice of A0 which centers A0 at the median of z1, zs, …, zN, with half-width about 2 times a preliminary estimate of σ0 based on the interquartile range. For our simulations this results in A0(2,2). In Figure 11 we provide a simulation showing the sensitivity of our methods to the choice of A0 and π0. In Table 2 we provide a simulation showing the sensitivity of our parameter estimates to the choice of π0 and N. In short, the empirical kFWER estimator is generally robust to the choice of A0 but does not estimate kFWER well when π0 and N are small.

Figure 11 A0 and π0 sensitivity: The dotted line is the mean kFWER^(Z)$\widehat {kFWER}({\cal Z})$ with Monte Carlo estimated 95 percent confidence interval as the shaded region. The mean kFWER¯(Z)$\overline {kFWER} ({\cal Z})$ is shown as the dashed line with the true kFWER (Z${\cal Z}$) in solid black. In this simulation we examine the LHS version of our estimators with N=500 and the non-null genes with large negative z-scores. In general, the estimators are robust to the choice of A0${{\cal A}_0}$ but the accuracy of kFWER^(Z)$\widehat {kFWER}({\cal Z})$ depends greatly on the choice of π0 with large values of π0 required for accurate F0 estimation. The kFWER¯(Z)$\overline {kFWER} ({\cal Z})$ is an accurate estimator of the true kFWER (Z${\cal Z}$) under all settings for π0 and A0 considered in this simulation.
Figure 11

A0 and π0 sensitivity: The dotted line is the mean kFWER^(Z) with Monte Carlo estimated 95 percent confidence interval as the shaded region. The mean kFWER¯(Z) is shown as the dashed line with the true kFWER (Z) in solid black. In this simulation we examine the LHS version of our estimators with N=500 and the non-null genes with large negative z-scores. In general, the estimators are robust to the choice of A0 but the accuracy of kFWER^(Z) depends greatly on the choice of π0 with large values of π0 required for accurate F0 estimation. The kFWER¯(Z) is an accurate estimator of the true kFWER (Z) under all settings for π0 and A0 considered in this simulation.

In the following subsections we present results related to the adjusted Bonferroni, Šidàk, and Holm procedures. Additionally we present several methods to empirically estimate null distributions.

Adjusted Bonferroni method

Theorem 1Using the two sample mixture model defined in (1), zbon=F01(kα/n)is such that kFWER(zbon)≤α when using the (LHS) kFWER definition wherekFWER(z)=kFWER(Z)whereZ=(,z).

Proof. Let Z=(,zbon). Then,

(LHS)kFWER(zbon)=Pr(N0(Z)k)E(N0(Z))k(byMarkovInequality)=E(E(N0(Z)|N0))/k(Lawoftotalexpectation)=E(N0F0(zbon))/k(sinceN0(Z)|N0~Bin(N0,F0(zbon)))=Nπ0F0(F01(kα/N))/k(sinceN0~Bin(N,π0))=π0αα.

Corollary 9.1Usingzbon=F01(1kα/n)and the (RHS) kFWER definition, we show that (RHS) kFWER(zbon)≤α where (RHS)kFWER(z)=kFWER(Z) whereZ=(z,).

Proof. A similar argument to Theorem 1 establishes the (RHS) kFWER result.□

Šidàk method

Theorem 2Consider the two sample mixture model defined in (1), zsid=F01(psid)where psid is such that

(45)FB(k1|N,psid)=1α. (45)

Then (LHS)kFWER(zsid)≤α when using the (LHS) kFWER definition where kFWER(z)=kFWER(Z) where Z=(,z).

Proof. We can assume that psid=F0(zsid). Then we have

(46)(LHS)kFWER(zsid)=i=kN(1FB(k1|i,psid))fB(i|N,π0). (46)

Note from (45) we can assume that FB(k–1|N, psid)=1–α and so 1–FB(k–1|N, psid)=α. Hence the last term of the sum in (46) is απ0N. Importantly, we further note that FB(k–1|n, p)>FB(k–1|N, p) for any n<N. Thus,

(47)1FB(k1,N,p)>1FB(k1,n,p). (47)

Thus, we can rewrite (46) with the understanding that (LHS)kFWER(zsid)=kFWER(Z) with Z=(,zsid) as follows,

(48)(LHS)kFWER(zsid)=i=kN(1FB(k1|i,psid))fB(i|N,π0)α[fB(k|N,π0))+fB(k+1|N,π0)++   fB(N|N,π0)]α(1FB(k1|N,π0))α(since1FB(k1|N,π0)1). (48)

Corollary 9.2Similarly, if we definezsid=F01(1psid) where psid is such that

(49)FB(k1|N,psid)=1α. (49)

Then (RHS)kFWER(zsid)≤α when using the (RHS) kFWER definition where (RHS) kFWER(z)=kFWER(Z) where Z=(z,).

Proof. Similar to the proof for Theorem 2. □

Holm method

Theorem 3Consider the two sample mixture model defined in (1) and zholm=F01(p(r)) where r is the largest index satisfying (28). Then (LHS) kFWER(zholm)≤α when using the (LHS) kFWER definition with one sided p-values defined by pi=F0(zi).

Proof. If rk, then our α adjustment is the same as the Bonferroni α adjustment and we can use the Markov inequality as employed in the proof for the adjusted Bonferroni argument.

Assume r>k, then we use the technique employed in Lehmann and Romano (2005). Let y1,y2,,yN0 denote the ordered z statistics for the true null hypotheses where y1y2y3yN0. Then let zj=yk where z1z2≤…≤zN denote the ordered z statistics. Thus, the following probability statements hold,

(50)(LHS)kFWER(zholm)=Pr({N0(,zholm)k})=Pr(#ofnullz-values (,zholm)k). (50)

The event {# of null z-values ∈(–∞, zholm)≥k} is equal to the event {yk=zjzholm}. In order to reject at least k true nulls, the largest possible value of j is NN0+k, namely, the situation where the NN0 true alternatives are the smallest z statistics. Hence, we have that yk=zjzNN0+k. Hence, we have Pr[#ofnullz-values(,zholm)k)=Pr({yk=zjzNN0+k}]. Now apply F0 to the event {yk=zjzNN0+k} which is a non decreasing function in order to obtain, F0(yk)=F0(zj)=pjF0(zNN0+k)=pNN0+k. However, since zjzholm we must have that pjαj. Also, αjαNN0+k=kα/N0 since α is an increasing function, see (29). Hence,

(51)(LHS)kFWER(zholm)=Pr(#ofnullz-values(,zholm)k)=Pr({yk=zjzNN0+k})=Pr({F0(yk)=F0(zj)pNN0+k})Pr(F0(yk)=F0(zj)=pjαNN0+k=kα/N0.=Pr(UkkαN0)whereUkF0(yk);thek-thnullp-value=Pr(W>k)whereW~Bin(N0,kα/N0)E(W)/KByMarkovinequality=α (51)

Corollary 9.3Consider the two sample mixture model defined in (1) and zholm=1F01(p(r)) where r is the largest index satisfying (28). Then kFWER(zholm)≤α when using the (RHS) kFWER definition where kFWER(z)=kFWER(Z) where Z=(z,).

Proof. Similar to the proof for Theorem 3.□

Theorem 4We assume that each zi follows the two group model independently and that F0is given as the standard normal CDF. We also assume the zero assumption where f1(z) is near zero for a subset of the sample space near zero, say, A0. Then E(kFWER¯(Z))kFWER(Z).

Proof. Since F0 is given, π0 is the only parameter to estimate in kFWER. To emphasize the dependence on π0, we will denote our estimator kFWER¯(Z) as kFWER¯π0¯(Z). According to (21) we have

(52)kFWER¯π0¯(Z)=η=kN(1FB(k1|η,F0¯(Z)))fB(η|N,π0¯) (52)
(53)=η=kNCfB(η|N,π0¯) (53)
(54)=η=kNC(Nη)ηπ0¯(Nη)1π0¯ (54)
(55)=η=kNCηπ0¯(Nη)1π0¯, (55)

where C and C′ do not depend on π0. By straightforward calculus the function in (55) is convex in terms of π0¯. Thus by Jensen’s inequality we have

(56)E(kFWER¯π0¯(Z))kFWERE(π0¯)(Z) (56)
(57)=kFWERπ0(Z) (57)
(58)=kFWER(Z) (58)

Note, in going from (6) to (7) we assume E(π0¯)=π0. Since F0 is known, we have by definition π0¯=N+(A0)/(NF0(A0)) and so

(59)E(π0¯)=E(N+(A0)NF0(A0)) (59)
(60)=1NF0(A0)E(N+(A0)) (60)
(61)=1NF0(A0)π0NF0(A0)(byzeroassumption) (61)
(62)=π0 (62)

Methods to estimate the null distribution

In this section, we paraphrase two methods described in Efron (2010) for estimating the null distribution. We assume that f0(z) is normal but not necessarily N(0,1) say,

(63)f0(z)~N(δ0,σ02), (63)

and we define fπ0(z)=π0f0(z). This implies that

(64)log(fπ0(z))=[log(π0)12{δ02σ02+log(2πσ02)}]+δ0σ02z12σ02z2, (64)

is a quadratic function of z.

The MLE method for empirical null distribution

This method was first introduced in Efron (2007). The maximum likelihood estimator (MLE) method starts with the zero assumption, where we assume that f1(z) is zero for a certain subset A0 of the sample space. In other words,

(65)f1(z)=0forzA0. (65)

We assume N0 is the number of zi in A0 and 0 their indices, 0={i:ziA0} and N0=#0. We define z0 as the corresponding collection of z-values,

(66)z0={zi,i0}. (66)

Also, let φδ0,σ0(z) be the N(δ0,σ02) density function,

(67)φδ0,σ0(z)=12πσ02exp{12(zδ0σ0)2} (67)

and

(68)H0(δ0,σ0)A0φδ0,σ0(z)dz, (68)

this being the probability that a N(δ0,σ02) variate falls in A0.

We suppose that the N zi values follow the two-group model (1) with f0~N(δ0,σ02) and f1(z)=0 for zA0. Then zo has density and likelihood function

(69)fδ0,σ0,π0(z0)=[(NN0)θN0(1θ)NN0][0φδ0,σ0(zi)H0(δ0,σo)] (69)

when θ=π0H0(δ0,σ0)=Pr({ziA0}).

Computations can produce maximum likelihood estimators (δ^0,σ^0,π^0);fδ0,σ0,π0(z0) is the product of two exponential families which can be solved separately (the two bracketed terms). The binomial term gives θ^=N0/N while δ^0 and σ^0 are the MLEs from a truncated normal family, obtained by familiar iterative calculations, finally yielding

(70)π^0=θ^/Ho(δ^0,σ^0). (70)

The log of (69) is concave in (δ0, σ0, π0) guaranteeing that the MLE solutions are unique. This is described more fully in Section 6.3 of Efron (2010).

The central matching method for empirical null distribution

This method was first introduced in Efron (2004). In this method, we define yk as the number of observations zi in the kth bin,

(71)yk=#{ziZk}, (71)

where we partition the range Z of zi values into K bins of equal width d with

(72)Z=k=1KZk. (72)

Then, with the central matching method, we estimate f0(z) and π0 by assuming that log(f(z)) is quadratic near 0 and equal to (64) with,

(73)log(f(z))β0+β1z+β2z2. (73)

Estimating (β0, β1, β2) can be done using least squares with the histogram counts yk around z=0 and matching coefficients between (64) and (73). In other words, via matching, we obtain,

(74)σ02=1/(2β2), (74)
(75)δ0=β1/(2β2), (75)
(76)logπ0=β0β124β2+log(π/β2). (76)

References

Bahadur, R. (1959): “A representation of the joint distribution of responses to N dichotomous items,” Technical report, Defense Technical Information Center Document.Search in Google Scholar

Cai, G. and S. Sarkar (2006): “Modified Simes’ critical values under positive dependence,” J. Stat. Plann. Infer., 136, 4129–4146.Search in Google Scholar

Dudoit, S., M. Van Der Laan and K. Pollard (2004): “Multiple testing Part I. single-step procedures for control of general type I error rates,” Statistical Applications in Genetics and Molecular Biology 3, Article 13.Search in Google Scholar

Efron, B. (2004): “Large-scale simultaneous hypothesis testing,” J. Am. Stat. Assoc., 99, 96–104.Search in Google Scholar

Efron, B. (2007): “Correlation and large-scale simultaneous significance testing,” J. Am. Stat. Assoc., 102, 93–103.Search in Google Scholar

Efron, B. (2010): Large-Scale Inference: Empirical Bayes Methods for Estimation, Testing, and Prediction, volume 1. Cambridge, United Kingdom: Cambridge Univ Pr.10.1017/CBO9780511761362Search in Google Scholar

Efron, B., B. B. Turnbull and B. Narasimhan (2011): locfdr: Computes local false discovery rates, R package version 1.1-7.Search in Google Scholar

Finos, L. and A. Farcomeni (2010): someKfwer: Controlling the Generalized Familywise Error Rate, R package version 1.1.Search in Google Scholar

Finos, L. and A. Farcomeni (2011): “k-fwer control without p-value adjustment, with application to detection of genetic determinants of multiple sclerosis in Italian twins,” Biometrics, 67, 174–181.10.1111/j.1541-0420.2010.01443.xSearch in Google Scholar

Golub, T. R., D. K. Slonim, P. Tamayo, C. Huard, M. Gaasenbeek, J. P. Mesirov, H. Coller, M. L. Loh, J. R. Downing, M. A. Caligiuri, C. D. Bloomfield and E. S. Lander (1999): “Molecular classification of cancer: class discovery and class prediction by gene expression monitoring,” Science, 286, 531–537.10.1126/science.286.5439.531Search in Google Scholar

Guo, W. and J. Romano (2007): “A generalized Šidàk-Holm procedure and control of generalized error rates under independence,” Statistical Applications in Genetics and Molecular Biology, 6, Article 3.10.2202/1544-6115.1247Search in Google Scholar

Holm, S. (1979): “A simple sequentially rejective multiple test procedure,” Scand. J. Stat., 6, 65–70.Search in Google Scholar

Jin, J. and T. Cai (2007): “Estimating the null and the proportion of nonnull effects in large-scale multiple comparisons,” J. Am. Stat. Assoc., 102, 495–506.Search in Google Scholar

Lehmann, E. (1966): “Some concepts of dependence,” Ann. Math. Stat., 37, 1137–1153.Search in Google Scholar

Lehmann, E. and J. Romano (2005): “Generalizations of the familywise error rate,” Ann. Stat., 33, 1138–1154.Search in Google Scholar

Miecznikowski, J. and D. Gaile (2012): “A novel characterization of the generalized family wise error rate using empirical null distributions,” Technical Report #1203, University at Buffalo, Department of Biostatistics, Buffalo, NY.Search in Google Scholar

Miecznikowski, J., D. Gold, L. Shepherd and S. Liu (2011): “Deriving and comparing the distribution for the number of false positives in single step methods to control k-fwer,” Stat. Probabil. Lett., 81, 1695–1705.Search in Google Scholar

Muralidharan, O. (2010): “An empirical bayes mixture method for effect size and false discovery rate estimation,” Ann Appl Stat 4, 422–438.10.1214/09-AOAS276Search in Google Scholar

Pollard, K. S., Y. Ge, S. Taylor and S. Dudoit: multtest: Resampling-based multiple hypothesis testing, R package version 1.22.0.Search in Google Scholar

Pounds, S. and S. Morris (2003): “Estimating the occurrence of false positives and false negatives in microarray studies by approximating and partitioning the empirical distribution of p-values. Bioinformatics, 19, 1236–1242.10.1093/bioinformatics/btg148Search in Google Scholar

R Core Team (2012). R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria. ISBN 3-900051-07-0.Search in Google Scholar

Romano, J. P. and M. Wolf (2005): “Exact and approximate stepdown methods for multiple hypothesis testing,” J. Am. Stat. Assoc., 100, 94–108.Search in Google Scholar

Romano, J. P., and M. Wolf (2010): “Balanced control of generalized error rates,” Ann. Stat., 38, 598–633.Search in Google Scholar

Roquain, E. and F. Villers (2011): “Exact calculations for false discovery proportion with application to least favorable configurations,” Ann. Stat., 39, 584–612.Search in Google Scholar

Sarkar, S. (2008): “Generalizing Simes’ test and Hochberg’s stepup procedure,” Ann. Stat., 36, 337–363.Search in Google Scholar

Singh, D., P. G. Febbo, K. Ross, D. G. Jackson, J. Manola, C. Ladd, P. Tamayo, A. A. Renshaw, A. V. D’Amico, J. P. Richie., et al. (2002): “Gene expression correlates of clinical prostate cancer behavior,” Cancer cell, 1, 203–209.10.1016/S1535-6108(02)00030-2Search in Google Scholar

Published Online: 2014-3-14
Published in Print: 2014-6-1

©2014 by Walter de Gruyter Berlin/Boston

Downloaded on 6.2.2026 from https://www.degruyterbrill.com/document/doi/10.1515/sagmb-2013-0032/html
Scroll to top button