Home Mathematics Semiparametric Odds Rate Model for Modeling Short-Term and Long-Term Effects with Application to a Breast Cancer Genetic Study
Article Publicly Available

Semiparametric Odds Rate Model for Modeling Short-Term and Long-Term Effects with Application to a Breast Cancer Genetic Study

  • Mengdie Yuan and Guoqing Diao EMAIL logo
Published/Copyright: May 9, 2014

Abstract

The proportional odds model is commonly used in the analysis of failure time data. The assumption of constant odds ratios over time in the proportional odds model, however, can be violated in some applications. Motivated by a genetic study with breast cancer patients, we propose a novel semiparametric odds rate model for the analysis of right-censored survival data. The proposed model incorporates the short-term and long-term covariate effects on the failure time data and includes the proportional odds model as a nested model. We develop efficient likelihood-based inference procedures and establish the large sample properties of the proposed nonparametric maximum likelihood estimators. Simulation studies demonstrate that the proposed methods perform well in practical settings. An application to the motivating example is provided.

1 Introduction

The proportional odds (PO) model [13] is commonly used for analyzing survival data. The PO model specifies that the survival function of the failure time T given the covariates X takes the form

S(t|X)=eβTXH(t)+eβTX,

where H(t) is a strictly increasing function with H(0)=0, and β is a vector of regression parameters. Unlike the Cox proportional hazards model [4], the hazard ratio between two sets of covariates is time-varying under the PO model. Instead, the odds ratio of survival is assumed to be constant over time under the PO model. Consequently, the hazard ratio converges to one as time goes to infinity.

Much attention has been paid to the development of inference procedures under the PO model and its generalizations. Murphy et al. [5] provided statistical inference for the maximum likelihood estimation with right-censored data under the PO model and demonstrated that the estimators of the regression parameters provided by Bennett [1] are efficient. The sieve maximum likelihood estimation and the estimation based on weighted empirical odds functions were developed by Shen [6] and Yang and Prentice [7], respectively. Computational issues in the PO models are discussed in Hunter and Lange [8]. The PO model has also been extensively studied for current status data and interval censored data, see Rossini and Tsiatis [9], Huang and Rossini [10], Rabinowitz et al. [11] and Sun et al. [12], among others. For correlated failure time data, the PO model with random effects has been investigated in literature [13, 14].

All the aforementioned methods are developed under the assumption of constant odds ratio over time. However, in some applications, this assumption is violated. Violation of this assumption may lead to biased and misleading results. A motivating example is from a copy number alteration (CNA) study with breast cancer patients [15]. The objective of the study was to determine whether the CNAs influence the time to develop distant metastasis among breast cancer patients. We expect the log odds curves of survival to be parallel for two groups if the PO assumption is true. Figure 1 presents the Kaplan–Meier curves for two groups partitioned by the median of copy number variation data at two SNP loci. It appears that the PO assumption is reasonable at SNP #2124; however, the two curves cross at SNP #39687 indicating a violation of the PO assumption. The PO model failed to capture the phenomenon of the crossing survival at SNP #39687.

Figure 1 The Kaplan–Meier survival curves of time to develop distant metastasis for the CNA study at SNP #2124 and SNP #39687
Figure 1

The Kaplan–Meier survival curves of time to develop distant metastasis for the CNA study at SNP #2124 and SNP #39687

To test the PO assumption, Dauxois and Kirmani [16] developed a formal analytical test for two-sample problems. Resampling approaches using observed score processes for the regression coefficients have been developed by Scheike [17] to check the goodness-of-fit of the PO model. Liu et al. [18] provided graphical diagnostics to check model misspecification for the PO model. To alleviate the PO assumption, Scheike [17] proposed a flexible transformation model which allows for time-varying coefficients for some covariates in the model. They developed estimation and inference procedures based on martingale estimating equations. Additionally, they proposed to use a robust sandwich-type covariance matrix for the estimates of the regression coefficients. More recently, Chen, Hu, Cheng, Musoke, and Zhao [19] proposed an extended PO model by incorporating “external” time-varying covariates and developed appropriate inference procedures for the regression parameters. Note that both Liu et al. [18] and Chen et al. [19] do not incorporate time-varying coefficients.

In this article, we propose an alternative approach for relating covariates of interest to the odds of survival for right-censored data. Specifically, we propose a novel semiparametric odds rate model to incorporate the short-term and long-term covariate effects. This model includes the PO model as a special case, leading to a new diagnostic tool for checking the PO assumption. We establish efficient likelihood-based estimation and inference procedures and derive the asymptotic properties of the proposed estimators. Results from numerical studies demonstrate that the proposed methods perform well in practical settings.

The remaining of the article is organized as follows. In Section 2, we propose the new model and derive the nonparametric maximum likelihood estimation. The model assumptions and the large sample properties are presented in Section 3. In Section 4, we conduct extensive simulation studies to investigate the finite sample properties of the proposed estimators. An application to the motivating CNA study is provided in Section 5 and we conclude with a brief discussion in Section 6. Theoretical proofs are provided in the Appendix.

2 Method

Suppose there are n independent subjects in a random sample. For the ith subject, let Ti be the failure time, Ci be the censoring time, and Xi be a p×1 vector of time-independent covariates. The observed data contain {(Yi=min(Ti,Ci),Δi=I(TiCi),Xi),i=1,,n}, where I() is the indicator function.

Let F(t|X) and S(t|X) be the conditional distribution function and survival function of the failure time T given the covariates X, respectively. We propose the following generalized odds model

(1)F(t|X)S(t|X)=H(t)eβTXF(t)+eγTXS(t),

where H(t) is a non-decreasing function with H(0)=0. In model (1), H(t) is the baseline odds function of failure, S(t) is the baseline survival function which is the survival function for a subject with X=0, and F(t)=1S(t) is the baseline cumulative distribution function.

Under the proposed model (1), the regression coefficients β and γ have attractive and intuitive interpretations. For two sets of covariates X and X, we can show that

limt0F(t|X)/S(t|X)F(t|X)/S(t|X)=eγT(XX),limtF(t|X)/S(t|X)F(t|X)/S(t|X)=eβT(XX).

Therefore, β and γ can be interpreted as the long-term and short-term log-odds ratios, respectively. When β=γ, model (1) reduces to the PO model. Hence, this model provides a useful diagnostic tool for testing the PO model. To be more specific, we can check the PO model by testing the null hypothesis H0:β=γ. The test can be performed by using Wald, score or likelihood ratio test statistics.

We now derive the likelihood function for the unknown parameters (β,γ,H). It can be shown that

S(t)=1H(t)+1,F(t)=H(t)H(t)+1.

It follows that

S(t|X)=H(t)+1eβTXH2(t)+eγTX+1H(t)+1.

Consequently, the hazard function is given by

λ(t|X)=H(t)eβTXH2(t)+2eβTXH(t)+eγTXH(t)+1eβTXH2(t)+(eγTX+1)H(t)+1,

where H(t) is the first derivative of H(t). Under the conditional independent censoring assumption, the likelihood function for (β,γ,H) has the form

(2)Ln(β,γ,H)=i=1nλΔi(Yi|Xi)S(Yi|Xi)=i=1nH(Yi)eβTXiH2(Yi)+2eβTXiH(Yi)+eγTXi{H(Yi)+1}eβTXiH2(Yi)+(eγTXi+1)H(Yi)+1Δi×H(Yi)+1eβTXiH2(Yi)+(eγTXi+1)H(Yi)+1.

We would like to make inference based on the likelihood function. Note that the maximum of the likelihood function does not exist since we can always choose H(Yi)=0 for some Yi with Δi=1. Therefore, following the techniques of nonparametric maximum likelihood estimation in the Cox proportional hazards model [20, 21], the PO model [5, 14], and the semiparametric transformation model [22], we allow H to be a right continuous function. Therefore, we allow the estimator of H to be a step function with positive jumps only at the observed failure time points. Let H{Yi} denote the jump size of H(t) at Yi. We replace H(Yi) with H{Yi} in eq. (2) to obtain the nonparametric likelihood function, still denoted by Ln(β,γ,H) for simplicity,

Ln(β,γ,H)=i=1nH{Yi}eβTXiH2(Yi)+2eβTXiH(Yi)+eγTXi{H(Yi)+1}eβTXiH2(Yi)+(eγTXi+1)H(Yi)+1Δi×H(Yi)+1eβTXiH2(Yi)+(eγTXi+1)H(Yi)+1.

Assume the data have m unique observed failure time points Y(1)<Y(2)<<Y(m). Easily we have H(t)=k:Y(k)tH{Y(k)}. Then

S(t)=1k:Y(k)tH{Y(k)}+1,F(t)=k:Y(k)tH{Y(k)}k:Y(k)tH{Y(k)}+1.

In this case, the odds function H(t|X)=F(t|X)/S(t|X) is given by

k:Y(k)teβTXF(Y(k))+eγTXS(Y(k))H{Y(k)}.

Hence, to estimate the unknown parameters, we maximize the nonparametric likelihood function with respect to the regression parameters β and γ, and jump sizes H{Y(1)}, H{Y(2)}, …, H{Y(m)}. The resulting nonparametric maximum likelihood estimators (NPMLEs) are denoted by (βˆn, γˆn, Hˆn).

The maximization procedure can be performed by using the quasi-Newton optimization algorithm in Press et al. [23]. Based on Fisher information theory, the covariance matrix for (βˆn, γˆn, Hˆn) can be obtained by inverting the observed information matrix, which is the negative second derivatives of the nonparametric likelihood function with respect to the regression parameters β and γ, and the jump sizes H{Y(1)}, H{Y(2)}, …, H{Y(m)}. The justification follows the arguments in theorem 2 of Parner [24].

3 Asymptotic properties

Before we derive the asymptotic properties of the NPMLEs, we impose the following regularity conditions:

  1. There exists some positive constant δ0 such that Pr(Ciτ|Xi)δ0 almost everywhere, where τ is the end of the study. In addition, Pr(Ti>τ|Xi)>0.

  2. Conditional on the vector of covariates Xi, the censoring time Ci is independent of Ti for i=1,,n.

  3. Xi are bounded for all i=1,,n. Furthermore, aTXi=0 almost everywhere if and only if a=0.

  4. The true values β0 and γ0 for parameters β and γ belong to the interior of a known compact set B0 in R2p.

  5. The true baseline odds function H0 belongs to

    H0={H:Hisstrictlyincreasingin[0,τ]andiscontinuouslydifferentiablewithH(0)=0andH(τ)<}.
Remark. Condition (C3) implies that Xi is linearly independent, which ensures the identifiability of the model. The proof of identifiability of the model is provided in Appendix A. Condition (C1) implies that any subject who survives at the end of the study is considered censored at time τ, and every subject has a positive probability to relapse after the study. Conditions (C4) and (C5) guarantee the existence of the NPMLEs.

We first establish the consistency of the NPMLEs.

Theorem 1

Under conditions (C1)–(C5), the NPMLEs (βˆn,γˆn,Hˆn) exist. Furthermore||βˆnβ0||0, ||γˆnγ0||0andsupt[0,τ]|Hˆn(t)H0(t)|0almost surely asn, where||||is the Euclidean norm.

The existence of (β,γ) can be proved from the compactness of the parameter space and the boundedness of covariates. Suppose for some Yi with Δi=1, we have Hˆn{Yi}=. In that case, the likelihood function must be 0, which implies Hˆn{Yi} must be finite. Therefore, the NPMLEs exist. The detailed proof for Theorem 1 is provided in Appendix B. The key part of the proof is to establish the boundedness of Hˆn(τ). Then we can construct a limit, (β,γ,H), for the sequences (βˆn,γˆn,Hˆn) and prove that the limit is actually (β0,γ0,H0) by using the Kullback–Leibler information.

We next establish the asymptotic normality of the NPMLEs. Let l[0,τ] be a space consisting of all uniformly bounded functions on [0,τ] and the norm on it is defined as the supremum absolute value on [0,τ].

Theorem 2

Under conditions (C1)–(C5), n(βˆnβ0,γˆnγ0,HˆnH0)Tconverges weakly to a Gaussian process with mean zero in the metric spacel(A)where

A=(θ1,θ2,g):θ1Rp,θ2Rp,gisafunctionon[0,τ],and||θ1||1,||θ2||1,|g|V1,where|g|Visthetotalvariationofgon[0,τ]}.

Detailed proofs for Theorem 2 are provided in Appendix C. Moreover, we can show that (βˆn,γˆn) is an asymptotic linear estimator for (β0,γ0) and the corresponding influence functions are in the spanned space via score functions, which implies (βˆn,γˆn) is semiparametrically efficient [25].

4 Simulation studies

We conducted simulation studies to examine the finite-sample performance of our proposed methods. We generate the failure time from the following model

F(t|X)=eβXt2+eγXteβXt2+eγXt+t+1,

where X was set to be uniform(0.5,0.5). Under the above model, H(t)=t.

Results were obtained under four scenarios of regression parameters (β,γ): (0.5, 0.5), (0, 0.5), (–0.5, 0.5), and (–0.5, 0). The long-term and short-term effects are equal under the first scenario where the model reduces to the PO model; there is no long-term effect under scenario 2; the long-term and short-term effects are in opposite directions under scenario 3; and there is no short-term effect under scenario 4. We generated the censoring time from an exponential distribution with mean 6. The censoring time was truncated at 4. The censoring rates were about 30% under all scenarios. For each setting, we considered the sample sizes of 100 and 200 and generated 10,000 replicates. The variance–covariance matrix of the NPMLEs was approximated by the inverse of the observed information matrix.

Table 1 presents the summary statistics of the NPMLEs of the unknown regression parameters and the baseline odds functions at the time points 0.3, 0.8, 2, and 4, the first three of which corresponds to the approximate first quartile, median and third quartile of the failure time, and the last one is the end of the study. In the table, Bias is the sampling average of the biases of the estimates; SEE denotes the average of the standard error estimates; SE denotes the empirical standard deviations of the parameter estimates; CP is the coverage probability of 95% confidence interval estimates; and PO denotes the average of the parameter estimates under the PO model.

Table 1

Summary statistics of the proposed NPMLEs

ParameterBiasSEESECPPO
Scenario 1: (β,γ)=(0.5,0.5)
n=100β0.0241.2771.4040.9550.498
γ−0.0241.0681.1320.941
H(0.3)−0.0130.0710.0730.908
H(0.8)−0.0310.1720.1730.921
H(2.0)−0.0540.4860.5010.913
H(4.0)−0.0281.2731.3530.898
n=200β0.0120.8490.8680.9590.499
γ−0.0120.7390.7410.952
H(0.3)−0.0060.0510.0510.937
H(0.8)−0.0140.1200.1190.938
H(2.0)−0.0200.3350.3370.934
H(4.0)−0.0010.8710.8930.929
Scenario 2: (β,γ)=(0,0.5)
n=100β−0.0101.2751.4110.9560.268
γ−0.0151.0641.1300.941
H(0.3)−0.0130.0720.0730.911
H(0.8)−0.0310.1720.1740.919
H(2.0)−0.0530.4870.5040.917
H(4.0)−0.0221.2781.3780.901
n=200β0.0000.8490.8690.9590.270
γ−0.0090.7370.7390.951
H(0.3)−0.0050.0510.0510.936
H(0.8)−0.0130.1200.1200.938
H(2.0)−0.0200.3360.3380.934
H(4.0)0.0000.8730.9020.928
Scenario 3: (β,γ)=(0.5,0.5)
n=100β−0.0491.2851.4190.9540.038
γ−0.0051.0651.1250.940
H(0.3)−0.0130.0720.0730.909
H(0.8)−0.0300.1730.1760.919
H(2.0)−0.0530.4910.5090.916
H(4.0)−0.0381.2771.3570.898
n=200β−0.0160.8560.8780.9580.041
γ−0.0060.7360.7390.949
H(0.3)−0.0060.0510.0510.935
H(0.8)−0.0130.1210.1210.938
H(2.0)−0.0190.3400.3440.934
H(4.0)−0.0040.8770.9050.932
Scenario 4: (β,γ)=(0.5,0)
n=100β−0.0161.2771.4160.955–0.234
γ−0.0121.0661.1310.939
H(0.3)−0.0130.0710.0720.908
H(0.8)−0.0320.1720.1740.918
H(2.0)−0.0560.4860.5040.914
H(4.0)−0.0391.2681.3450.896
n=200β−0.0040.8490.8710.957–0.231
γ−0.0070.7380.7410.948
H(0.3)−0.0060.0510.0510.934
H(0.8)−0.0140.1200.1190.938
H(2.0)−0.0210.3360.3390.936
H(4.0)−0.0040.8710.8970.930

From Table 1, we observed that the biases of the NPMLEs were small under all simulation settings. The standard deviation estimate exhibited the true variation. The coverage probabilities of the 95% confidence interval estimates were close to the nominal level. As expected, as sample size increases, the coverage probabilities improve. The PO model yielded biased results when the PO assumption is violated. Particularly, when the long-term and short-term effects are in opposite directions, the PO model had little power to detect covariate effects.

The next set of simulation studies were concerned with the performance of our proposed procedures for testing the covariate effects and the PO assumption. We used the same simulation settings as the previous simulation study except that the censoring time was from a uniform distribution in (0,20), yielding approximate censoring rates of 20% under all scenarios. We considered five hypothesis testings here: (H1) H0:β=0; (H2) H0:γ=0; (H3) H0:β=γ=0; (H4) H0:β=γ; (H5) H0:βPO=0. The first four tests were conducted based on our proposed model. The last one was conducted based on the PO model. We used Wald test statistics for testing (H1), (H2), (H4), and (H5) and likelihood ratio test statistics for testing (H3). The data were generated under the same model as above with n=400.

Table 2 presents the sizes/powers at the nominal significance level of 0.05 based on 10,000 replicates. Our method provides accurate control of type I error rates. We also observed reasonable powers of the proposed tests for testing (H3) compared to the test using the PO model when the PO assumption is true; the powers for testing the covariate effect under both scenarios β=γ=1 and β=γ=1 are approximately 84.2% and 90.7% using the proposed model and the PO model, respectively. However, when the PO assumption is violated, the proposed tests tend to be more powerful than the test based on the PO model, especially when the long-term and short-term effects are in opposite directions. For example, when (β,γ)=(1,1), the powers for testing (H1), (H2), and (H3) were 0.4553, 0.5213, and 0.5804, respectively, compared to the power of 0.0569 from the PO model. The proposed test for the PO assumption also yielded reasonable powers; for example, we obtained powers of 0.6755 and 0.6713 for (β,γ)=(1,1) and (β,γ)=(1,1), respectively.

Table 2

Empirical sizes/powers at significance level of 0.05 based on 10,000 replicates

βγH0:β=0γ=0β=γ=0β=γβPO=0
−1−10.47890.51260.84570.04940.9065
−100.48120.05370.42120.22490.3548
−110.45530.52130.58040.67550.0569
0−10.05110.52230.45340.23610.3932
000.05020.05210.05330.04880.0518
010.04880.51690.45320.23580.3952
1−10.44970.52590.57870.67130.0568
100.47270.05360.41810.22610.3586
110.47670.50220.84220.04920.9065

In the last set of simulation studies, we compared the performance of our method with the method by Scheike [17]. We considered the simulation settings similar to those in Scheike [17]. Specifically, we generated data from the following model

F(t|X,Z)S(t|X,Z)=A0(t)+XA1(t)exp(Zβ),

where Z is a standard normal random variable, X is a log-normal random variable truncated at 30 with mean 0.3Z and standard deviation 0.4, A0(t)=0.1t, and A1(t)=0.05t2. We varied the value of β from –0.2 to 0.2. The failure times were censored at 12 yielding approximate censoring rates of 30%. For each setting, we considered the sample sizes of 200 and 400. We fit the proposed model with the constraint of equal short-term and long-term effects of Z and compared its performance with that of Scheike [17].

Table 3 summarizes the results for estimating the effect of Z. The NPMLE of the effect of Z under the proposed model appears to have little bias suggesting that the proposed method is robust to moderate model mis-specifications. Compared to the method of Scheike [17], the NPMLE has slightly larger biases but the empirical standard deviations are smaller. Additionally, the NPMLE appears to be more efficient in terms of the mean squared error since the method of Scheike [17] involves the estimation of two unknown functions A0(t) and A1(t) instead of one unknown function H(t) in the proposed model.

Table 3

Comparison between the proposed method and the method of Scheike [17]1

NewScheike [17]
TrueBiasSEMSEBiasSEMSERE
n=200−0.20.0090.1220.0150.0040.1330.0181.170
−0.10.0050.1220.0150.0020.1310.0171.169
0.00.0010.1210.0150.0010.1310.0171.172
0.1−0.0030.1220.0150.0000.1320.0171.172
0.2−0.0070.1230.015−0.0020.1330.0181.170
n=400−0.20.0120.0860.0070.0030.0940.0091.196
−0.10.0070.0850.0070.0030.0940.0091.211
0.00.0020.0850.0070.0020.0940.0091.217
0.1−0.0030.0850.0070.0010.0940.0091.216
0.2−0.0090.0850.0070.0000.0940.0091.203

5 Example

In this Section, we apply the proposed methods to the motivating CNA study [15] described in Section 1. In this study, the copy number data for 313 lymph node-negative breast cancer patients who were selected from the tumor bank at the Erasmus Medical Center are available. Survival analysis on time to develop distant metastasis (DM) was conducted at 56859 SNP loci in the breast cancer genome. The follow-up time ranged from 20 to 169 months with a median of 99 months. Among the 313 breast cancer patients, 114 patients developed distant metastasis yielding a censoring rate of 63.6%.

There are two types of estrogen receptor (ER) for tumors, ER positive or ER negative. Among 313 patients, 199 tumors were ER positive; and 161 patients experienced menopause. The ages (years) of patients at the time of surgery were recorded ranging from 26 to 83 with a median of 54. As suggested by Zhang et al. [15], we conduct the analysis for ER-positive group and ER-negative group separately. For each group, we include surgery age and menopause status as confounders and evaluate the long-term and short-term effects of the CNV data at each of the 56859 SNP loci. As a common practice, we standardize all the covariates in the numerical implementation.

Using the proposed method, the PO assumption holds at about 99.5% of SNP loci for age and about 66.4% for menopause status. Table 4 presents the results for the ER-positive group. The first part of Table 4 presents the top 10 SNPs with the smallest P-values for testing the PO assumption. It evidently shows that the PO model is inappropriate when the long-term and short-term effects are in opposite directions. At these loci, the proposed methods are able to detect significant short-term and/or long-term genetic effects missed by the PO model. The second part of Table 4 presents the top 10 SNPs with the most significant copy number effect under the PO model. The results demonstrate that our proposed model can also detect significant long-term and/or short-term effects at those SNP loci. In contrast to the first part, we present the first 10 SNPs with the largest P-values for testing the PO assumption in the third part of Table 4. The proposed methods yield short-term and long-term effects similar to the estimates obtained from the PO model, suggesting that the proposed methods work well when the PO assumption is true.

Table 4

Results for the CNA study1

SNP loci with the smallest P-values for testing the PO assumption
SNPβˆ3se(βˆ3)γˆ3se(γˆ3)β3=0γ3=0β3=γ3=0β3=γ3βˆ3POβ3PO=0
396871.4820.249−1.1300.2122.55E–091.01E–072.93E–101.82E–120.0060.967
397341.4810.249−1.1300.2122.55E–091.01E–072.94E–101.82E–120.0060.967
360911.4630.249−1.1390.2124.52E–098.01E–083.02E–102.48E–120.0010.996
112221.4040.246−1.1180.2091.14E–088.88E–089.13E–104.75E–12−0.0220.885
27435−2.4100.3410.7080.1951.68E–122.84E–044.21E–075.12E–12−0.1720.222
78791.4110.251−1.1220.2091.82E–087.88E–089.19E–106.54E–12−0.0280.851
79341.4110.251−1.1220.2091.82E–087.89E–089.19E–106.54E–12−0.0280.851
524931.3560.235−1.1510.2218.15E–092.00E–071.10E–097.42E–120.0290.844
550321.3550.235−1.1520.2228.43E–092.00E–071.12E–097.63E–120.0290.844
543351.3550.235−1.1520.2228.43E–092.00E–071.12E–097.63E–120.0290.844
SNP loci with the most significant effects in the PO model
SNPβˆ3se(βˆ3)γˆ3se(γˆ3)β3=0γ3=0β3=γ3=0β3=γ3βˆ3POβ3PO=0
381360.8560.2010.4260.1882.07E–052.38E–021.39E–060.1140.6235.04E–06
9940−0.9740.5310.9020.1686.63E–028.37E–088.48E–060.0020.6205.64E–06
9732−0.9780.5300.9020.1686.49E–028.14E–088.35E–060.0020.6205.64E–06
9889−0.9750.5310.9020.1686.60E–028.34E–088.47E–060.0020.6205.65E–06
24812−0.3730.364−0.7890.2183.05E–013.04E–045.69E–060.399−0.6496.79E–06
84950.9840.2870.4180.2356.05E–047.51E–022.28E–060.1940.6116.93E–06
259400.9680.3190.4390.2452.43E–037.35E–023.01E–060.2750.6147.04E–06
313530.9690.3190.4380.2462.41E–037.48E–023.06E–060.2740.6137.16E–06
314120.9690.3190.4380.2462.42E–037.48E–023.06E–060.2740.6137.17E–06
312530.9690.3190.4380.2462.42E–037.48E–023.07E–060.2740.6137.18E–06
SNP loci with the largest P-values for testing the PO assumption
SNPβˆ3se(βˆ3)γˆ3se(γˆ3)β3=0γ3=0β3=γ3=0β3=γ3βˆ3POβ3PO=0
192520.1080.3540.1080.2430.7600.6560.4371.0000.1130.420
330940.0570.4290.0570.2060.8940.7800.6761.0000.0530.699
192000.1080.3540.1080.2430.7600.6560.4371.0000.1130.420
330460.0570.4290.0570.2050.8940.7800.6761.0000.0530.699
38039−0.3220.445−0.3220.2300.4700.1600.0271.000−0.3260.025
38133−0.3220.446−0.3220.2300.4710.1610.0271.000−0.3260.026
184460.2190.2510.2190.2100.3820.2970.0951.0000.2190.096
8897−0.1280.304−0.1280.2310.6740.5780.3731.000−0.1270.378
2124−0.2030.285−0.2030.2420.4760.4010.1721.000−0.1930.196
263660.2240.2790.2240.1910.4210.2410.0751.0000.2170.083

As described in Section 1, the PO assumption appears to be true at SNP #2124, but is violated at SNP #39687. In Table 4, at SNP #2124, the estimates for the long-term and short-term effects were both −0.203 with standard error estimates of 0.285 and 0.242, respectively. The estimate for the effect under the PO model was –0.193, which is close to the estimates under the proposed model. At SNP #39687, we obtained significant long-term and short-term effects in opposite directions with estimates 1.482 and −1.130. However, the PO model yielded insignificant results with regression parameter estimate of 0.006 (P-value =0.967). These results suggest that in this case the proposed model works as well as the PO model when the PO assumption is true and outperforms the PO model when the assumption is violated.

We also calculated the adjusted odds ratio comparing patients with copy number value of 1 against 0 while fixing the two confounders at their sample means. Figure 2 displays the estimated curves of the log odds ratios with their pointwise 95% confidence bands at SNP #2124 and SNP #39687. It appears that the log odds ratio curve at SNP #2124 is close to a horizontal line, whereas the log odds ratio curve at SNP #39687 is non-decreasing and crosses 0 around 60 months suggesting violation of the PO assumption.

Figure 2 Log odds ratios (solid curves) and the 95% pointwise confidence bands (dashed curves) at SNP #2124 and SNP #39687
Figure 2

Log odds ratios (solid curves) and the 95% pointwise confidence bands (dashed curves) at SNP #2124 and SNP #39687

6 Discussion

We have proposed a novel odds rate model incorporating short-term and long-term covariate effects and have established efficient likelihood-based nonparametric estimation and inference procedures. The proposed model allows one to explicitly model the long-term and short-term covariate effects on the odds of survival without the use of smoothing. The simulation studies demonstrate that our model can detect the short-term and long-term effects when the PO assumption is violated, while the PO model is inappropriate in this circumstance. We have implemented our new method in an efficient computer program. This program is freely available upon request.

One limitation of the proposed method is that we assume inexplicitly that the odds ratio function is monotonic over time. Consider two univariate covariates X>X. It can be shown that

(3)F(t|X)/S(t|X)F(t|X)/S(t|X)=eβ(XX)F(t)+eγ(XX)S(t)=eβ(XX)F(t)+eγ(XX){1F(t)}.

When β<γ, the odds ratio in eq. (3) is decreasing, whereas when β>γ, the odds ratio is increasing. On the other hand, the transformation model proposed by Scheike [17] postulates that

F(t|X,Z)S(t|X,Z)=XTA(t)exp(ZTβ),

where Z is another set of covariates which satisfy the PO assumption, A() is a vector of functions, and H(t|X)XTA(t) is assumed to be increasing for all X. This model is more general than the proposed model in the sense that it does not impose any assumption on the odds ratio functions A(t). However, it is not clear whether the estimators of A(t) are constrained to be monotonic (see equation (7) in Scheike [17]). Alternatively, one may consider another model described by Scheike [17]

F(t|X,Z)S(t|X,Z)=H(t)expZTβ+XTα(t).

This model is not identifiable as t0. Furthermore, it is not clear how one can ensure that the estimator of S(t|X,Z) is non-increasing.

We assume in this article that the failure time and the censoring time are conditionally independent given the covariates. Violation of this assumption may lead to biased results. For example, in the analysis of time to develop distant metastasis in the CNA study, censoring is likely to be informative because some local recurrences might censor the observations while being driven by the same tumor. To account for such dependent censoring, one may assume certain dependence structures between the failure time and the censoring time (e.g., see Tanaka and Rao [26] and Li et al. [27]). Alternatively, we can adopt the technique proposed by Othus et al. [28] through the use of the conditional crude hazard function of the censoring time given the covariates. Future research along this direction is warranted.

In this article, we assume independent observations. In several applications, we may encounter clustered or family data. One interesting future work is to extend our proposed method by incorporating random effects to account for the correlations within a cluster/family. Suppose Xij is the set of covariates for the jth subject in the ith family. We can incorporate random effects in the following model

F(t|Xij,ωi)S(t|Xij,ωi)=H(t)ωieβTXijF(t)+eγTXijS(t),

where ωi denotes the shared frailty for the ith family. We are currently investigating the theoretical and computational aspects of this model.

Yang and Prentice [29] proposed a two-sample model that accommodates the short-term and long-term covariate effects on the hazard function. More recently, Diao et al. [30] extended the above two-sample model to a regression setting accommodating possibly external time-dependent covariates. While the models of Yang and Prentice [29] and Diao et al. [30] are extensions of the Cox proportional hazards model, the proposed model renders direct interpretations of short-term and long-term covariate effects on the survival function and includes the PO model as a special case. It is well known that both the proportional hazards model and the PO model belong to a general class of semiparamtric transformation models [22]. One may be interested in investigating the short-term and long-term covariate effects under this class of transformation models. Future research is warranted.

The flat lines after about 80 months in Figure 1 suggest that a proportion of patients in the CNA study may never develop distant metastasis, i.e., there exists a cured sub-population. To account for the presence of a cured sub-population, we can combine the proposed model and the standard mixture cure model. Specifically, given the covariates Zi and Xi, the mixture cure model postulates that the population survival function takes the form

Spop(t|Zi,Xi)=1θ(Zi)+θ(Zi)S(t|Xi),

where

θ(Zi)=exp(βTZi)1+exp(βTZi)

and S(t|Xi) is the conditional survival function in the un-cured sub-population. Note that 1θ(Zi) is the cure fraction. We assume the proposed model for S(t|Xi). The likelihood function is then given by

i=1nθ(Zi)f(Yi|Xi)Δi1θ(Zi)+θ(Zi)S(Yi|Xi)1ΔiI(Yi<)1θ(Zi)I(Yi=),

where f(t|Xi) is the conditional density function of the failure time in the un-cured sub-population. We can perform nonparametric likelihood estimation and inference using techniques similar to those in Zeng et al. [31] and Diao and Yin [32].

In this study, we performed hypothesis testing at each SNP locus to search for the significant location at which the CNV impacts the time to develop distant metastasis, leading to the so-called large p small n problem. Appropriate procedures are needed to control for the family-wise error rate introduced by the multiple comparisons. Many statistical methods have been proposed in literature to account for multiple testing. Particularly, Lin [33] developed a Monte Carlo approach to approximate the joint distribution of the test statistics along the genome. It would be interesting to adopt this approach in our model setting to adjust for multiple testing.

An interesting area is focused on the selection of important variables that impact the failure time. Recently, Lu and Zhang [34] proposed variable selection procedures using adaptive LASSO for the PO model. It would be desirable to extend the adaptive LASSO procedure to select variables with significant long-term and/or short-term effects.

7 Software

Program in the form of C code is available on request from the corresponding author.

Acknowledgments

This work was supported by National Institutes of Health [CA150698].

Appendix A

Proof of identifiability

Suppose two sets of parameters (β,γ,H) and (β˜,γ˜,H˜) give the same likelihood function for any observed data. That is,

(4)H(Y)eβTXH2(Y)+2eβTXH(Y)+eγTX{H(Y)+1}eβTXH2(Y)+(eγTX+1)H(Y)+1Δ×H(Y)+1eβTXH2(Y)+(eγTX+1)H(Y)+1=H˜(Y)eβ˜TXH˜2(Y)+2eβ˜TXH˜(Y)+eγ˜TX{H˜(Y)+1}eβ˜TXH˜2(Y)+(eγ˜TX+1)H˜(Y)+1Δ×H˜(Y)+1eβ˜TXH˜2(Y)+(eγ˜TX+1)H˜(Y)+1.

Since we assume eq. (4) holds for any Y and X, setting Δ=1, Y=0 and X=0, we have H(0)=H˜(0). And for any X=0 with Δ=1, Y=0, we have eγTX=eγ˜TX, i.e.

(γγ˜)TX=0.

By (C3), we obtain γ=γ˜. With Δ=0 and X=0, for any Y=y, we have H(y)=H˜(y). It follows that eβTX=eβ˜TX, which implies β=β˜ with (C3).

Appendix B

Proof of consistency

By constructing a new step function H˜n(t)=Hˆn(t)/Hˆn(τ), we show that its corresponding likelihood will be larger than the likelihood of the NPMLEs if Hˆn(τ) goes to infinity.

In fact, let ln(β,γ,H) denote the log-likelihood with sample size n. We have

1n[ln(β^n,γ^n,H^n)ln(β^n,γ^n,H˜n)]=1n[ln(β^n,γ^n,H^n(τ)H˜n)ln(β^n,γ^n,H˜n)]=1ni=1n{Δilog[H^n(τ){eβ^nTXiH^n2(τ)H˜n2(Yi)+2eβ^nTXiH^n(τ)H˜n(Yi)+eγ^nTXi}{H^n(τ)H˜n(Yi)+1}{eβ^nTXiH^n2(τ)H˜n2(Yi)+(eγ^nTXi+1)H^n(τ)H˜n(Yi)+1}{H˜n(Yi)+1}{eβ^nTXiH˜n2(Yi)+(eγ^nTXi+1)H˜n(Yi)+1}{eβ^nTXiH˜n2(Yi)+2eβ^nTXiH˜n(Yi)+eγ^nTXi}]+log[H^n(τ)H˜n(Yi)+1eβ^nTXiH^n2(τ)H˜n2(Yi)+(eγ^nTXi+1)H^n(τ)H˜n(Yi)+1eβ^nTXiH˜n2(Yi)+(eγ^nTXi+1)H˜n(Yi)+1H˜n(Yi)+1]}.

Given the compactness of the parameter space and the boundedness of covariates and H˜n, it is easy to show when Hˆn(τ) goes to infinity, the first log term goes to a bounded limit, and the second log term goes to . That is,

01nln(βˆn,γˆn,Hˆn)ln(βˆn,γˆn,H˜n),asn.

By the contradiction, Hˆn(τ) is bounded almost everywhere. In addition, H˜n(t) is bounded by 1. This holds for arbitrary n.

Hence, it follows from compactness theorems that there exist convergent subsequences of (βˆn, γˆn, Hˆn) with limits (β, γ, H).

We next show that β=β0, γ=γ0, and H=H0. We begin with constructing a new sequence of step functions Hˉn(t) with jumps at those Yi’s for which Δi=1. We define the jump size as

ΔiHˉn{Yi}=k=1nI(YkYi)ΔkAk(β0,γ0,H0)+eβ0TXkH02(Yk)+2eβ0TXkH0(Yk)+eγ0TXk2Bk(β0,γ0,H0),

where

Ak(β,γ,H)=e2βTXkH4(Yk)+4e2βTXkH3(Yk)+(2e2βTXk+4eβTXk+γTXk)H2(Yk)+(2eβTXk+γTXk2eβTXk+2e2γTXk+2eγTXk)H(Yk)2eβTXk+e2γTXk+2eγTXk

and

Bk(β,γ,H)=(H(Yk)+1)eβTXkH2(Yk)+(eγTXk+1)H(Yk)+1×eβTXkH2(Yk)+2eβTXkH(Yk)+eγTXk.

The above equation is derived similarly from the score functions of ln(β,γ,H) with respect to H{Yi} as follows:

ΔiHˆn{Yi}=k=1nI(YkYi)ΔkAk(βˆn,γˆn,Hˆn)+eβˆnTXkHˆn2(Yk)+2eβˆnTXkHˆn(Yk)+eγˆnTXk2Bk(βˆn,γˆn,Hˆn),

where Hˉn(t)=i=1nHˉn{Yi}I(Yit) and Hˆn(t)=i=1nHˆn{Yi}I(Yit).

Let

Mk(β,γ,H)=ΔkAk(β,γ,H)+eβTXkH2(Yk)+2eβTXkH(Yk)+eγTXk2Bk(β,γ,H).

We have

(5)Hˆn(t)=0tk=1nI(YkYi)Mk(β0,γ0,H0)k=1nI(YkYi)Mk(βˆn,γˆn,Hˆn)dHˉn(t).

Therefore Hˆn(t) is absolutely continuous with respect to Hˉn(t). If we could verify that Hˉn(t) converges uniformly to H0(t) in [0,τ] with probability 1, then in eq. (5), letting n, we obtain

Hn(t)=0tk=1nI(YkYi)Mk(β0,γ0,H0)k=1nI(YkYi)Mk(β,γ,H)dH0(t).

Therefore Hn(t) is differentiable with respect to H0(t). It follows that dHˆn(t)/dHˉn(t) converges uniformly to dH(t)/dH0(t) in [0,τ]. Here we introduce a lemma before the following proof.

Lemma 1

The class

F={I(Y1y)M1(β,γ,H):y[0,τ],(β,γ)B,HH}
is a bounded p-Donsker class, therefore a Glivenko–Cantelli class, whereBis a compact set and
H={h:hisincreasingin[0,τ]withh(0)=0,h(τ)<B0whereB0isapositiveconstantsuchthatHˆn(τ)B0withprobabilityone}.

Detailed proof of Lemma 1 is provided in Appendix D. By the Glivenko–Cantelli theorem given by van der Vaart and Wellner [35],

Hˉn(t)EI(Y1t)Δ1μ(Y1)uniformly,whereμ(y)=EI(Y1y)M1(β0,γ0,H0).

Let Sc(|X) be the survival function for the censoring time. We can show that

μ(y)=E[H(y){eβ0TXH02(y)+2eβ0TXH0(y)+eγ0TX}Sc(y|X){H0(y)+1}{eβ0TXH02(y)+(eγ0TX+1)H0(y)+1}].

Then we have

EI(Y1t)Δ1μ(Y1)=E0tH(y)eβ0TXH02(y)+2eβ0TXH0(y)+eγ0TXSc(y|X)μ(y){H0(y)+1}eβ0TXH02(y)+(eγ0TX+1)H0(y)+1dH0(y)=0tdH0(y)=H0(t).

Thus, Hˉn(t) converges uniformly to H0(t) in [0,τ] with probability 1. By the argument above, dHˆn(t)/dHˉn(t) converges uniformly to dH(t)/dH0(t) in [0,τ].

On the other hand,

1n[ln(β^n,γ^n,H^n)ln(β0,γ0,H¯n)]=1ni=1n{Δilog[H^n{Yi}{eβ^nTXiH^n2(Yi)+2eβ^nTXiH^n(Yi)+eγ^nTXi}{H^n(Yi)+1}{eβ^nTXiH^n2(Yi)+(eγ^nTXi+1)H^n(Yi)+1}{H¯n(Yi)+1}{eβ0TXiH¯n2(Yi)+(eγ0TXi+1)H¯n(Yi)+1}H¯n{Yi}{eβ0TXiH¯n2(Yi)+2eβ0TXiH¯n(Yi)+eγ0TXi}]+log[H^n(Yi)+1eβ^TXiH^n2(Yi)+(eγ^TXi+1)H^n(Yi)+1eβ0TXiH¯n2(Yi)+(eγ0TXi+1)H¯n(Yi)+1H¯n(Yi)+1]}0.

Letting n, given X, we have

E[1ni=1nlimn{ln,i(β^n,γ^n,H^n)ln,i(β0,γ0,H¯n)}]=E{log[{H*'(Y1){eβ*TXH*2(Y1)+2eβ*TXH*(Y1)+eγ*TX}(H*(Y1)+1)P1(β*,γ*,H*)}Δ1{H*(Y1)+1P1(β*,γ*,H*)}/{H0(Y1){eβ0TXH02(Y1)+2eβ0TXH0(Y1)+eγ0TX}(H0(Y1)+1)P1(β0,γ0,H0)}Δ1{H0(Y1)+1P1(β0,γ0,H0)}]}0,

where ln,i(β,γ,H) is the log-likelihood contributed from the ith subject and

P1(β,γ,H)=eβTXH2(Y1)+(eγTX+1)H(Y1)+1.

This expectation is the negative Kullback–Leibler information, which is non-positive. As a result, it equals zero. Consequently, we have

H(Y1)eβTXH2(Y1)+2eβTXH(Y1)+eγTX(H(Y1)+1)P1(β,γ,H)Δ1H(Y1)+1P1(β,γ,H)=H0(Y1)eβ0TXH02(Y1)+2eβ0TXH0(Y1)+eγ0TX(H0(Y1)+1)P1(β0,γ0,H0)Δ1H0(Y1)+1P1(β0,γ0,H0).

Using the same argument in the proof of identifiability, we have β=β0, γ=γ0, and H=H0.

Therefore, ||βˆnβ0||0, ||γˆnγ0||0 and |Hˆn(t)H0(t)|0 pointwise, almost surely as n. Since H0 is a continuous function, we have further uniform convergence supt[0,τ]|Hˆn(t)H0(t)|0 almost surely as n.

Appendix C

Proof of asymptotic normality

First, we introduce some notation in the context of empirical process. Let n and be the empirical measure and the population distribution of n i.i.d. observations O1,O2,,On. Let Gn=n(n) denotes the empirical process. Then for any measurable function h,

nh=1ni=1nh(Oi),h=E[h],

and

Gnh=n(nhh).

By the consistency theorem, ||βˆnβ0||+||γˆnγ0||+supt[0,τ]|Hˆn(t)H0(t)|<ε with probability 1 when sample size n is large enough. Therefore we consider a neighborhood of (β0,γ0,H0),

U=(β,γ,H):||ββ0||+||γγ0||+supt[0,τ]|H(t)H0(t)|<ε

for a small constant ε, and define a sequence of functions Kn mapping U to l(A) as

Kn(β,γ,H)[θ1,θ2,g]=1nddεln(β+εθ1,γ+εθ2,H(t)+ε0tgdH(s))|ε=0.

Particularly, consider a single observation (Y,Δ,X), we have

(6)K(β,γ,H)[θ1,θ2,g]=Δ[g(Y)+eβTX{θ1TXH2(Y)+2θ1TXH(Y)+2(H(Y)+1)EHg(Y)}eβTXH(Y)(H(Y)+2)+eγTX+eγTXθ2TXeβTXH(Y)(H(Y)+2)+eγTX](Δ+1)[eβTX(θ1TXH2(Y)+2H(Y)EHg(Y))eβTXH2(Y)+(eγTX+1)H(Y)+1+eγTX(θ2TXH(Y)+EHg(Y))+EHg(Y)eβTXH2(Y)+(eγTX+1)H(Y)+1](Δ1)EHg(Y)H(Y)+1,

where EHg(Y)=0Yg(s)dH(s).

Essentially, the empirical process GnK also maps from U to l(A), and

nK(βˆn,γˆn,Hˆn)[θ1,θ2,g]=0,K(β0,γ0,H0)[θ1,θ2,g]=0.

This is implied from the definition of Kn and the consistency of NPMLEs.

To prove the asymptotic normality, we verify three Properties (a)(c) below. Then it follows from theorem 3.3.1 of van der Vaart and Wellner [35] that n(βˆnβ0,γˆnγ0,HˆnH0)T converges in distribution to a Gaussian process in l(A).

  1. GnK(β^n,γ^n,H^n)[θ1,θ2,g]GnK(β0,γ0,H0)[θ1,θ2,g]=o[1+n(||β^nβ0||+||γ^nγ0||+supt[0,τ]|H^n(t)H0(t)|)].

  2. GnK(β0,γ0,H0)[θ1,θ2,g] converges to a tight random element ξ.

  3. K(β,γ,H)[θ1,θ2,g] is Frechet differentiable at (β0,γ0,H0) with a continuously invertible derivative (˙K)(β0,γ0,H0).

Note that we assume that β and γ belong to the interior of a known compact set B0 in R2p, H(τ)<, and X are bounded. Therefore, it is easy to see from the explicit expression (6) where (θ1,θ2,g)A that K(β,γ,H) is continuously differentiable with respect to (β,γ) and the norm of the derivative is bounded. Furthermore, K(β,γ,H1)K(β,γ,H2) is dominated by

|H1(Y)H2(Y)|+0Yd|H1(s)H2(s)|.

Therefore, we have

sup(θ1,θ2,g)AK(β,γ,H)[θ1,θ2,g]K(β0,γ0,H0)[θ1,θ2,g]20,

as ||ββ0||+||γγ0||+supt[0,τ]|H(t)H0(t)|0.

Using the arguments in Appendix D, we prove that the class

K(β,γ,H)[θ1,θ2,g]K(β0,γ0,H0)[θ1,θ2,g]:(β,γ,H)U,(θ1,θ2,g)A

is p-Donsker. Therefore, by applying lemma 3.3.5 of van der Vaart and Wellner [35], Property (a) holds.

Similarly, we can prove that the class {K(β0,γ0,H0)[θ1,θ2,g]:(θ1,θ2,g)A} is p-Donsker. Then Property (b) holds. In fact, ξ is a Gaussian process indexed by (θ1,θ2,g)A. The covariance between ξ(θ1,θ2,g) and ξ(θ1,θ2,g) is denoted by

K(β0,γ0,H0)[θ1,θ2,g]×K(β0,γ0,H0)[θ1,θ2,g].

The smoothness of (β,γ,H) implies the Frechet differentiability in Property (c).

According to the proof of theorem 2 in Zeng and Lin [36], to show the continuously invertibility of (˙K) can be implemented by showing that (˙K) is one-to-one; that is,

K(β0,γ0,H0)[θ1,θ2,g]=0

implies θ1=0,θ2=0,g=0. In fact, to prove the continuously invertibility, it suffices to prove the Fisher information along the path β0+εθ1,γ0+εθ2,H0+εEH0g is nonsingular. This can be proved by using the idea of proving the identifiability in Appendix A.

Let Δ=1,Y=0, we have θ2TX+g(0)=0. By Condition (C3), we have θ2=0 and g(0)=0. Let Δ=0,X=0, we obtain EHg=0. While when Δ=1,X=0, g2EHgH+1=0. Thus, g=0. Plugging in θ2=0,g=0, we have

K(β0,γ0,H0)[θ1,0,0]=Δeβ0TX(θ1TXH2+2θ1TXH)eβ0TXH(H+2)+eγ0TX(Δ+1)eβ0TXθ1TXeβ0TXH2+(eγ0TX+1)H+1.

By letting Δ=0, we immediately have θ1=0 by Condition (C3). Therefore, Property (c) is proved.

It follows from theorem 3.3.1 of van der Vaart and Wellner [35] that n(βˆnβ0,γˆnγ0,HˆnH0)T converges in distribution to a tight Gaussian random element (˙K)1ξ in l(A).

Appendix D

Proof of p-Donsker of F

Recall that

F=I(Y1y)M1(β,γ,H):y[0,τ],(β,γ)B,HH.

First we prove that

F={M1(β,γ,H):y[0,τ],(β,γ)B,HH}

is a bounded p-Donsker class. And {I(Y1y),y[0,τ]} is uniformly bounded and monotone on the real line, therefore is p-Donsker. By the preservation of Donsker property under product, we can prove Lemma B.1.

It remains to prove that F is a bounded p-Donsker class. Despite the complex expression for M1(β,γ,H), it is easy to know that it is bounded and continuously differentiable with respect to (β,γ) for any (β,γ)B and the norm of the derivative are also bounded. It is also dominated by H. Therefore, by the mean-value theorem, |M1(β,γ,H)M1(β˜,γ˜,H˜)| is dominated by ||ββ˜||+||γγ˜||+supt(0,τ)|H(t)H˜(t)|. Then F is a p-Donsker class [35]. It follows that F is a p-Donsker class.

References

1. Bennett S. Analysis of survival data by the proportional odds model. Stat Med 1983;2:273–7.10.1002/sim.4780020223Search in Google Scholar PubMed

2. Bennett S. Log-logistic regression-models for survival data. Appl Stat – J R Stat Soc Ser C 1983;32:165–71.10.2307/2347295Search in Google Scholar

3. Pettitt AN. Proportional odds models for survival data and estimates using ranks. Appl Stat – J R Stat Soc Ser C 1984;33:169–75.10.2307/2347443Search in Google Scholar

4. Cox DR. Regression models and life-tables. J R Stat Soc: Ser B (Methodological) 1972;34:187–220.10.1007/978-1-4612-4380-9_37Search in Google Scholar

5. Murphy SA, Rossini AJ, van der Vaart AW. Maximum likelihood estimation in the proportional odds model. J Am Stat Assoc 1997;92:968–76.10.1080/01621459.1997.10474051Search in Google Scholar

6. Shen X. Proportional odds regression and sieve maximum likelihood estimation. Biometrika 1998;85:165–77.10.1093/biomet/85.1.165Search in Google Scholar

7. Yang S, Prentice RL. Semiparametric inference in the proportional odds regression model. J Am Stat Assoc1999;94:125–36.10.1080/01621459.1999.10473829Search in Google Scholar

8. Hunter DR, Lange K. Computing estimates in the proportional odds model. Ann Inst Stat Math 2002;54:155–68.10.1023/A:1016126007531Search in Google Scholar

9. Rossini AJ, Tsiatis AA. A semiparametric proportional odds regression model for the analysis of current status data. J Am Stat Assoc 1996;91:713–21.10.1080/01621459.1996.10476939Search in Google Scholar

10. Huang J, Rossini A. Sieve estimation for the proportional-odds failure-time regression model with interval censoring. J Am Stat Assoc 1997;92:960–7.10.1080/01621459.1997.10474050Search in Google Scholar

11. Rabinowitz D, Betensky RA, Tsiatis AA. Using conditional logistic regression to fit proportional odds models to interval censored data. Biometrics 2000;56:511–18.10.1111/j.0006-341X.2000.00511.xSearch in Google Scholar PubMed

12. Sun JG, Sun LQ, Zhu C. Testing the proportional odds model for interval-censored data. Lifetime Data Anal 2007;13:37–50.10.1007/s10985-006-9029-6Search in Google Scholar PubMed

13. Lam KF, Lee YW, Leung TL. Modeling multivariate survival data by a semiparametric random effects proportional odds model. Biometrics 2002;58:316–23.10.1111/j.0006-341X.2002.00316.xSearch in Google Scholar

14. Zeng D, Lin DY, Yin G. Maximum likelihood estimation for the proportional odds model with random effects. J Am Stat Assoc 2005;100:470–83.10.1198/016214504000001420Search in Google Scholar

15. Zhang Y, Martens JW, Yu JX, Jiang J, Sieuwerts AM, Smid M, et al. Copy number alterations that predict metastatic capability of human breast cancer. Cancer Res 2009;69:3795–801.10.1158/0008-5472.CAN-08-4596Search in Google Scholar PubMed

16. Dauxois JY, Kirmani S. Testing the proportional odds model under random censoring. Biometrika 2003;90:913–22.10.1093/biomet/90.4.913Search in Google Scholar

17. Scheike TH. A flexible semiparametric transformation model for survival data. Lifetime Data Anal 2006;12:461–80.10.1007/s10985-006-9021-1Search in Google Scholar PubMed

18. Liu I, Mukherjee B, Suesse T, Sparrow D, Park SK. Graphical diagnostics to check model misspecification for the proportional odds regression model. Stat Med 2009;28:412–29.10.1002/sim.3386Search in Google Scholar PubMed

19. Chen YQ, Hu N, Cheng SC, Musoke P, Zhao LP. Estimating regression parameters in an extended proportional DDDS model. J A Stat Assoc 2012;107:318–30.10.1080/01621459.2012.656021Search in Google Scholar PubMed PubMed Central

20. Murphy SA. Consistency in a proportional hazards model incorporating a random effect. Ann Stat 1994;22:712–31.10.1214/aos/1176325492Search in Google Scholar

21. Murphy SA. Asymptotic theory for the frailty model. Ann Stat 1995;23:182–98.10.1214/aos/1176324462Search in Google Scholar

22. Zeng D, Lin DY, Lin X. Semiparametric transformation models with random effects for clustered failure time data. Stat Sin 2008;18:355–77.Search in Google Scholar

23. Press WH, Teukolsky SA, Vetterling WT, Flannery BP. Numerical recipes in C: the art of scientific computing, 2 ed. New York: Cambridge University Press, 1992.Search in Google Scholar

24. Parner E. Asymptotic theory for the correlated gamma-frailty model. Ann Stat 1998;26:183–214.10.1214/aos/1030563982Search in Google Scholar

25. Bickel PJ, Klaassen CAJ, Ritov Y, Wellner JA. Efficient and adaptive estimation for semiparametric models. Baltimore, MD: Johns Hopkins University Press, 1993.Search in Google Scholar

26. Tanaka Y, Rao PV. A proportional hazards model for informatively censored survival times. J Stat Plan Inference 2005;129:253–62.10.1016/j.jspi.2004.06.051Search in Google Scholar

27. Li Y, Tiwari RC, Guha S. Mixture cure survival models with dependent censoring. J R Stat Soc Ser B (Statistical Methodology) 2007;69:285–306.10.1111/j.1467-9868.2007.00589.xSearch in Google Scholar

28. Othus M, Li Y, Tiwari RC. A class of semiparametric mixture cure survival models with dependent censoring. J Am Stat Assoc 2009;104:1241–50.10.1198/jasa.2009.tm08033Search in Google Scholar PubMed PubMed Central

29. Yang S, Prentice R. Semiparametric analysis of short-term and long-term hazard ratios with two-sample survival data. Biometrika 2005;92:1–17.10.1093/biomet/92.1.1Search in Google Scholar

30. Diao G, Zeng D, Yang S. Efficient semiparametric estimation of short-term and long-term hazard ratios with right-censored data. Biometrics 2013;69:840–9.10.1111/biom.12097Search in Google Scholar PubMed PubMed Central

31. Zeng D, Yin G, Ibrahim JG. Semiparametric transformation models for survival data with a cure fraction. J Am Stat Assoc 2006;101:670–84.10.1198/016214505000001122Search in Google Scholar

32. Diao G, Yin G. A general transformation class of semiparametric cure rate frailty models. Ann Inst Stat Math2012;64:959–89.10.1007/s10463-012-0354-0Search in Google Scholar

33. Lin DY. An efficient Monte Carlo approach to assessing statistical significance in genomic studies. Bioinformatics 2005;21:781–7.10.1093/bioinformatics/bti053Search in Google Scholar PubMed

34. Lu WB, Zhang HH. Variable selection for proportional odds model. Stat Med 2007;26:3771–81.10.1002/sim.2833Search in Google Scholar PubMed

35. van der Vaart AW, Wellner JA. Week convergence and empirical processes: with applications to statistics. New York: Springer, 1996.10.1007/978-1-4757-2545-2Search in Google Scholar

36. Zeng D, Lin DY. Semiparametric transformation models with random effects for recurrent events. J Am Stat Assoc 2007;102:167–80.10.1198/016214506000001239Search in Google Scholar

Published Online: 2014-5-9
Published in Print: 2014-11-1

© 2014 by Walter de Gruyter Berlin / Boston

Downloaded on 31.12.2025 from https://www.degruyterbrill.com/document/doi/10.1515/ijb-2013-0037/html
Scroll to top button