Abstract
In studies of clinical phenotypes, such as dementia, disability, and frailty, participants are typically assessed at in-person clinic visits. Thus, the precise time of onset for the phenotype is unknown. The discreteness of the clinic visits yields grouped event time data. We investigate how to perform a risk factor analysis in the case of grouped data. Since visits can be months to years apart, numbers of ties can be large, causing the exact tie-handling method of the Cox model to be computationally infeasible. We propose two, new, computationally efficient approximations to the exact method: Laplace approximation and an analytic approximation. Through extensive simulation studies, we compare these new methods to the Prentice-Gloeckler model and the Cox model using Efron’s and Breslow’s tie-handling methods. In addition, we compare the methods in an application to a large cohort study (N = 3,605) on the development of clinical frailty in older adults. In our simulations, the Laplace approximation has low bias in all settings, and the analytic approximation has low bias in settings where the regression coefficient is not large in magnitude. Their corresponding confidence intervals also have approximately the nominal coverage probability. In the data application, the results from the approximations are nearly identical to that of the Prentice-Gloeckler model.
1 Introduction
Older adults are at increased risk of experiencing poor health outcomes, such as frailty, disability, and dementia. Identifying risk factors for these outcomes is a major goal of prospectively conducted epidemiological studies of aging. Many aging-related phenotypes can only be assessed at in-person study visits. For these phenotypes, a participant’s time of onset cannot be known precisely. It is known up to the time interval between the study visit in which the outcome was first observed and the previous visit. This type of data in which the true time of onset is only known to be in a certain group is called grouped event time data, and is the focus of our paper. Grouped data arises whenever the outcome of interest is assessed at discrete study visits, so it is common not only in aging research but also in many other fields. Some examples of epidemiological studies that have collected grouped event data include the Cardiovascular Health Study, the Multiethnic Study of Atherosclerosis, and the Women’s Health and Aging Study (Fried et al. 1991; Bild et al. 2002; Onder et al. 2005).
The structure of event times for grouped data differs from that of interval censored event times. In the grouped data setting, the intervals are fixed by the number and timing of visits given by the study design, while in the interval-censored data setting the intervals are random. In the grouped data setting, all the observed event times will cluster into a small number of groups, which would correspond to the study visits, whereas in the case of interval censored data the number of intervals can be quite large.
We are interested in examining how the grouped nature of the data affects risk factor analysis in prospective cohort studies. One modeling approach is the Prentice-Gloeckler (PG) model, which accounts for the grouped nature of the data (Prentice and Gloeckler 1978). A straightforward, but less principled approach is to use the Cox proportional hazards (PH) model, the workhorse of survival analysis, substituting the unknown precise onset of the outcome with the visit time at which the outcome is first observed. The Cox model estimates the impact of covariates on the hazard function, which is based on continuous event time, whereas the observed event times in the grouped data setting are discrete. Due to the grouped nature of the event time, there will be a substantial number of ties in event times, and this needs to be addressed while using the Cox PH model. There are three well-known tie-handling techniques: the exact method, the method by Breslow, and the method by Efron. A challenge of the exact method is that it can be computationally intractable when there is a large number of ties in the data. We propose two new approximations to the exact method that are computationally efficient, even for a very large number of tied failure times. The first approximation is the Laplace approximation to an integral (Erdelyi 1956, p. 36–37), and the second is an analytic approximation to the integral.
Through simulation, we compare our two approximations to the PG model, and to the Cox PH model using Breslow’s and Efron’s tie-handling schemes. In general, the proposed methods yield estimators with low bias, and confidence intervals with the desired coverage probability. Also, we apply the proposed methods to the Cardiovascular Health Study (Fried et al. 1991), a cohort study on the development of clinical frailty in older adults. The methods give comparable results to the Prentice-Gloeckler model, in our data application. The R code for the proposed methods is provided online. The link can be found in the Supplementary Materials section at the end of this paper. Issues of competing events, although relevant given the type of events considered in our motivating example, are not addressed in this paper.
Currently, Breslow’s approximation is the default method for handling tied data in most software packages, such as Stata and SAS. Hertz-Picciotto and Rockhill (1997) and Scheike and Sun (2007) found that Breslow’s approximation tends to underestimate the coefficients in the Cox model and that Efron’s approximation performs better than Breslow’s approximation. The simulation results of this paper are consistent with those findings.
This paper is organized as follows. The assumptions are stated in Section 2. The PG model and the Cox PH Model, including the exact tie-handling method and the tie-handling methods by Breslow and Efron, are introduced in Section 3. We present our proposed approximations to the exact method in Section 4. The data application to the Cardiovascular Health Study and the simulation studies are described in Sections 5 and 6, respectively. Concluding remarks are given in Section 7.
2 Context and assumptions
We consider a population whose members have not experienced the event of interest at baseline (t = 0), which is entry into the study. For each member, we assume her/his hazard function
which is the Cox proportional hazards assumption. In (1),
For any given participant i, the observed data vector is
A common attribute of grouped data is heavy ties in the observed event times. This occurs because adjacent visits can be months to years apart. For example, in the Cardiovascular Health Study, the first follow-up assessment of the clinical frailty outcome occurred three years after the baseline visit. There were 307 participants whose observed event time was the first follow-up visit, and 187 participants whose observed event time was the second follow-up visit. Define
We compare the performance of various approaches to estimate
3 Existing approaches
3.1 Prentice-Gloeckler model
Prentice and Gloeckler (1978) apply maximum likelihood to estimate
We refer to
where
The PG model is equivalent to using a generalized linear model of the binomial family with a complementary log-log link (Allison 1982). The Cox model and PG model both make the proportional hazards assumption and they do not require the user to assume a specific form for the baseline hazard
3.2 Cox model with existing tie-handling methods
Let
However, ties are common in our context of grouped data, as discussed in Section 2. We present the existing methods for tie-handling using a simple example. Suppose that participants 1 and 2 are tied at Visit k. Their contribution to the partial likelihood is either equal to:
if participant 1 experienced the event before participant 2, or equal to:
if participant 2 experienced the event before participant 1. Both (7) and (8) have the term
In the exact approach, one takes the average of (7) and (8) (Kalbfleisch and Prentice 2002). The exact approach is computationally feasible when the number of tied subjects
Commonly used tie-handling methods in the Cox PH model are those proposed by Breslow and Efron (Therneau and Grambsch 2010). Rather than taking the average of the
In Efron’s method, the full sum is in the first denominator, while the second denominator uses the average of
More generally, if there are
4 Proposed approximations to the exact tie-handling method
As discussed in Section 3, computing the exact likelihood is difficult when the tie sizes are not small. For any given participant i and Visit k, where
For conciseness, we refer to this as
Our proposed method is to approximate (14), using two new approaches: the Laplace approximation and a simple analytic approximation. We also considered a Gaussian quadrature approach, the Gauss-Laguerre quadrature (Davis and Rabinowitz 1984, p. 222–223). However, it did not work as well as the other approaches discussed here, mainly due to the difficulty in choosing the appropriate number of quadrature points (results not shown).
4.1 Laplace approximation
Our first approximation is based on the Laplace approximation to an integral (Erdelyi 1956) and will be referred to as “Laplace" throughout the paper. Here, we present the procedure for Laplace. The mathematical derivations are presented in Appendix A. Denote the integrand of (14) as
Let
Define
Let
This is equivalent to finding the maximizer of the function
which is obtained by taking the logarithm of (15) and omitting the terms that do not depend on
4.2 Analytic approximation
We propose another approximation that will be referred to as “Analytic,” in which (14) is approximated as
which is a function of
which is the product of (16) over
which results from taking the logarithm of (17) and omitting the terms that do not depend on
5 Application to the cardiovascular health study
The Cardiovascular Health Study (CHS) was a study of 5,201 older adults enrolled between 1989 and 1990, and followed up at annual clinic visits until 1999 (Fried et al. 1991). One of the outcomes collected in CHS was clinical frailty, which is a syndrome defined as the presence of three or more of the following: unintentional weight loss, exhaustion, low grip strength, slow walking speed, and low physical activity (Fried et al. 2001). Frailty status was assessed at the baseline visit and at two follow-up visits, occurring three and seven years after the baseline visit, respectively. We fitted the Cox and PG models to time-to-frailty data collected in CHS. It should be noted that, in addition to the cohort of 5,201 participants, a second cohort of 687 participants was recruited in 1992. We focus on the first cohort in our analysis. This is because the follow-up schedule for frailty differs between the two cohorts. The second cohort had only a single follow-up assessment of frailty.
In our analysis of time-to-frailty, five covariates were used: age, gender (0 = female, 1 = male), whether the person had ever smoked (0 = no, 1 = yes), whether the person had received education beyond high school (0 = no, 1 = yes), and C-reactive protein (CRP) level. These covariates were collected at the baseline visit. We standardized both age and CRP level. Age, in units of years, was standardized by subtracting 65 and dividing by 10. CRP level, in units of mg/L, was standardized by dividing by the interquartile range (2.15 mg/L). The fitted models included main terms for the covariates and no interaction terms. We let δ be an indicator of whether the participant was assessed as frail during her/his follow-up (0 if no, 1 if yes). If δ = 0, we let V be the last follow-up visit at which the participant’s frailty status was assessed. If δ = 1, we let V be the first follow-up visit at which the participant was assessed as frail.
We included only participants who satisfied the following conditions: (i) were non-frail at the baseline visit, (ii) frailty status was assessed at the first follow-up only, or at both follow-ups, (iii) covariates were non-missing. The resulting sample size was n = 3,605. The purpose for (i) was, since we sought to study the time-to-frailty development after the baseline visit, we focused on those who were not yet frail at the baseline visit. Criterion (ii) ensured that there was some follow-up on the frailty status after the baseline visit because those without any such follow-up would not affect the model fits. Also, criterion (ii) excluded those whose frailty status was observed at the second follow-up but not the first follow-up. If the status at the second follow-up was frail, then δ = 1 but it could not be deduced whether V should be 2 or 1 (i. e. the person was actually frail by the first follow-up). If the status at the second follow-up was non-frail, there was also some ambiguity since the person might have been non-frail through the second follow-up, or have been frail at the first follow-up and transitioned to being non-frail by the second follow-up. A potential impact of excluding participants who did not attend the first follow-up visit is that the sample used in our analysis may under-represent a specific subtype of participants. For example, the participants who missed the first follow-up visit may be more likely to have severe health problems, e. g. some of them might have missed their study visit due to their health problems. Criterion (iii) was required since participants without the full set of covariates would not impact the model fits.
Web Figure 1 in the Supplementary Materials presents the empirical marginal distributions of age, gender, smoking status, education level, and CRP level, after excluding participants as described above. The mean age is 72 (range 64–95) and the proportion of female participants is 0.58. Table 1 presents the empirical joint distribution on (δ,V), after excluding participants as described above. Each cell gives the number of participants, followed by the corresponding sample proportion in parentheses. We evaluated the validity of the proportional hazards assumption by testing the null hypothesis that the correlation between the scaled Schoenfeld residuals and time is zero, separately for each covariate. The results supported the proportional hazards assumption; the p-value was 0.10 for age, 0.77 for education level, 0.48 for gender, 0.33 for smoking status, and 0.80 for CRP level.
Distribution of (δ,V) in Cardiovascular Health Study Application.
V | |||
---|---|---|---|
1 | 2 | ||
δ | 0 | 884 (0.25) | 2227 (0.62) |
1 | 307 (0.09) | 187 (0.05) |
Table 2 presents the regression coefficient estimates in the top panel, with the corresponding standard error estimates in the middle panel. The values are shown up to four decimal places. The results in this table and those throughout the paper were computed using R. When fitting the Cox model, we applied our proposed tie-handling methods Laplace and Analytic, as well as the existing methods by Efron and Breslow. When assessing each of these methods, we use PG as a reference since it accounts for the discreteness of the data directly. All of the methods were computationally efficient; in this data application, the running time was 0.01 seconds for Efron’s method, 0.02 seconds for Breslow’s method, 0.3 seconds for PG, 1.2 seconds for Analytic, and 1.4 seconds for Laplace.
Regression Coefficient Estimates, Standard Error Estimates, and 95 % Confidence Intervals in the Cardiovascular Health Study Application. Panel (a) presents the regression coefficient estimates and Panel (b) presents the standard error estimates. Panel (c) presents both the estimates of and the 95 % confidence intervals for the exponentiated regression coefficients. Each confidence interval is obtained by taking the corresponding regression coefficient estimate (in Panel A), adding and subtracting the quantity of 1.96 multiplied by the standard error estimate (in Panel B), and then exponentiating the results. The covariate age, in units of years, has been standardized by subtracting 65 and dividing by 10. The covariate CRP level, in units of mg/L, has been standardized by dividing by the interquartile range (2.15 mg/L).
(a) Regression coefficient estimates. | |||||
---|---|---|---|---|---|
PG | Cox PH | ||||
Laplace | Analytic | Efron | Breslow | ||
age | 1.1127 | 1.1127 | 1.1069 | 1.1063 | 1.0263 |
education | |||||
gender | |||||
smoke | 0.1856 | 0.1856 | 0.1845 | 0.1849 | 0.1708 |
CRP | 0.0418 | 0.0418 | 0.0417 | 0.0418 | 0.0405 |
(b) Standard error estimates. | |||||
---|---|---|---|---|---|
PG | Cox PH | ||||
Laplace | Analytic | Efron | Breslow | ||
age | 0.0793 | 0.0793 | 0.0792 | 0.0786 | 0.0775 |
education | 0.0924 | 0.0924 | 0.0924 | 0.0922 | 0.0922 |
gender | 0.1003 | 0.1003 | 0.1003 | 0.1002 | 0.1000 |
smoke | 0.0949 | 0.0949 | 0.0949 | 0.0948 | 0.0947 |
CRP | 0.0123 | 0.0123 | 0.0123 | 0.0123 | 0.0125 |
(c) Point estimates and 95 % confidence ntervals for exponentiated regression coefficients. | |||||
---|---|---|---|---|---|
PG | Cox PH | ||||
Laplace | Analytic | Efron | Breslow | ||
age | 3.04 [2.60, 3.55] | 3.04 [2.60, 3.55] | 3.02 [2.59, 3.53] | 3.02 [2.59, 3.53] | 2.79 [2.40, 3.25] |
education | 0.75 [0.63, 0.90] | 0.75 [0.63, 0.90] | 0.75 [0.63, 0.90] | 0.75 [0.63, 0.90] | 0.77 [0.64, 0.92] |
gender | 0.58 [0.48, 0.71] | 0.58 [0.48, 0.71] | 0.59 [0.48, 0.71] | 0.59 [0.48, 0.71] | 0.60 [0.50, 0.73] |
smoke | 1.20 [1.00, 1.45] | 1.20 [1.00, 1.45] | 1.20 [1.00, 1.45] | 1.20 [1.00, 1.45] | 1.19 [0.99,1.43] |
CRP | 1.04 [1.02, 1.07] | 1.04 [1.02, 1.07] | 1.04 [1.02, 1.07] | 1.04 [1.02, 1.07] | 1.04 [1.02,1.07] |
In the upper two panels of Table 2, the results for Laplace perfectly match those for PG. For all covariates, the absolute difference between the Laplace estimate and the corresponding PG estimate was on the order of
The bottom panel of Table 2 shows the 95 % confidence intervals for the exponentiated regression coefficients computed using each approach. For interpretability, we present the confidence intervals for the exponentiated regression coefficients rather than for the regression coefficients themselves. Each confidence interval is computed by taking the corresponding regression coefficient estimate, adding and subtracting the quantity 1.96 multiplied by the standard error estimate, and exponentiating both values. Beside each 95 % confidence interval, we also present the corresponding regression coefficient estimate (after exponentiation). Although there are some differences between the confidence intervals computed by the different approaches, the qualitative conclusions are the same. The various approaches all find that a person’s hazard for frailty increases with higher age, holding the other factors constant. Also, the hazard for frailty is lower for those who have education beyond high school than for those who have a high school education or less. Females have a higher hazard than males, holding other factors constant. All of the approaches concluded that the hazard for frailty is not statistically different between those who have ever smoked versus those who have never smoked, but the point estimate suggests that smoking raises the hazard. The hazard for frailty increases with higher CRP level, holding other factors constant.
6 Simulation studies
Through simulation, we evaluate the Cox and PG models in the grouped data setting. For the Cox model, the Laplace, Analytic, Efron’s, and Breslow’s tie-handling methods are separately applied. We compare the regression coefficient estimators with respect to bias, standard error, and root-mean-square error. Also, we evaluate the actual coverage probability of the nominal 95 % confidence interval for each method.
Each simulation study includes a large number of simulations,
where
where
Each simulation follows n participants according to a schedule which includes a baseline visit (Visit 0) and m follow-up visits (Visits
6.1 Single covariate simulations
In these simulation studies, each simulation is a randomized controlled trial with m = 3 follow-up visits. The single covariate Z represents the treatment assignment (1 if control, 0 if treatment). Treatment assignment is random, with equal probability of being assigned to treatment or control, i. e. P(Z = z) = 0.5 for z ∈ { 0,1}.
The simulations also generate the censoring time C and the grouped event time
The grouped event time
We formulated (22) so that there would be approximately 25 % with observed event times, 25 % dropout, and 50 % administrative censoring in each randomized trial, if β = 0. The baseline hazard function, i. e. (22) under Z = 0, is
By definition, the censoring indicator δ and visit number V are as follows:
Note that, although
Single covariate simulation studies: average number of ties at each visit, for each setting of β and sample Size n.
n = 200 | n = 500 | n = 2000 | |||||||
---|---|---|---|---|---|---|---|---|---|
β | Visit 1 | Visit 2 | Visit 3 | Visit 1 | Visit 2 | Visit 3 | Visit 1 | Visit 2 | Visit 3 |
0 | 7 | 18 | 22 | 18 | 45 | 56 | 72 | 178 | 222 |
0.2 | 8 | 20 | 24 | 20 | 49 | 60 | 80 | 196 | 240 |
0.4 | 9 | 22 | 26 | 22 | 54 | 64 | 89 | 216 | 257 |
0.6 | 10 | 24 | 28 | 25 | 60 | 69 | 101 | 239 | 275 |
0.8 | 11 | 27 | 29 | 29 | 66 | 73 | 114 | 265 | 293 |
1 | 13 | 29 | 31 | 33 | 73 | 77 | 131 | 294 | 307 |
2 | 27 | 45 | 28 | 68 | 112 | 70 | 271 | 448 | 282 |
Table 4 presents the absolute bias, standard error, root-mean-square error, and coverage probability at n = 500. Web Table 1 and Table 2 of the Supplementary Materials present the results for sample sizes n = 200 and 2,000. For a visual representation, the tabulated results are also plotted in Web Figures 2–4 in the Supplementary Materials. These figures exclude the results for Breslow’s method since it has larger bias than the other methods when
Bias, Standard Error (SE), Root-Mean-Square Error (RMSE), and Coverage Probability (CP) in Single Covariate Simulation Study with n = 500. For each value of β, the number of simulations is 1,0000.
CoxPH | PG | |||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Laplace | Analytic | Efron’s method | Breslow’s method | |||||||||||||||||
β | Bias | SE | RMSE | CP | Bias | SE | RMSE | CP | Bias | SE | RMSE | CP | Bias | SE | RMSE | CP | Bias | SE | RMSE | CP |
0 | 0.002 | 0.189 | 0.189 | 0.951 | 0.002 | 0.187 | 0.187 | 0.953 | 0.002 | 0.188 | 0.188 | 0.951 | 0.002 | 0.176 | 0.176 | 0.965 | 0.002 | 0.189 | 0.189 | 0.951 |
0.2 | –0.002 | 0.179 | 0.179 | 0.949 | –0.004 | 0.178 | 0.178 | 0.952 | –0.003 | 0.179 | 0.179 | 0.949 | –0.017 | 0.166 | 0.167 | 0.965 | –0.002 | 0.180 | 0.180 | 0.948 |
0.4 | 0.003 | 0.173 | 0.173 | 0.947 | –0.001 | 0.172 | 0.172 | 0.949 | 0.002 | 0.173 | 0.173 | 0.947 | –0.031 | 0.159 | 0.162 | 0.961 | 0.004 | 0.174 | 0.174 | 0.947 |
0.6 | 0.003 | 0.168 | 0.168 | 0.950 | –0.005 | 0.166 | 0.166 | 0.953 | 0.001 | 0.168 | 0.168 | 0.950 | –0.054 | 0.153 | 0.162 | 0.958 | 0.004 | 0.169 | 0.169 | 0.950 |
0.8 | 0.004 | 0.164 | 0.164 | 0.949 | –0.011 | 0.161 | 0.161 | 0.952 | 0.000 | 0.164 | 0.164 | 0.949 | –0.080 | 0.148 | 0.168 | 0.939 | 0.005 | 0.164 | 0.164 | 0.949 |
1 | 0.000 | 0.161 | 0.161 | 0.950 | –0.025 | 0.156 | 0.158 | 0.951 | –0.006 | 0.161 | 0.161 | 0.951 | –0.115 | 0.144 | 0.185 | 0.903 | 0.001 | 0.162 | 0.162 | 0.950 |
2 | 0.006 | 0.157 | 0.158 | 0.952 | –0.177 | 0.135 | 0.222 | 0.803 | –0.033 | 0.153 | 0.157 | 0.945 | –0.378 | 0.130 | 0.400 | 0.277 | 0.010 | 0.158 | 0.158 | 0.951 |
Laplace and PG have absolute bias close to zero that also improves with higher sample size. At n = 200, the magnitude of the absolute bias is
Although it has the best standard error and negligible bias when β = 0, Breslow’s method has the worst bias of all the methods when
This first set of simulation studies mimics a randomized controlled trial setting. To mimic a typical epidemiological setting without randomization to the exposures, we perform simulations based on the Cardiovascular Health Study.
6.2 Data-based simulations
In the data-based simulation studies, we generate each data set by sampling n participants with replacement from the Cardiovascular Health Study data set, discussed in Section 5. We focus on the outcome frailty and the covariates age, gender, education, CRP level, and smoking. The true
Data-Based Simulation Studies: Average Number of Ties at Each Visit, for Each Sample Size n.
n | Visit 1 | Visit 2 |
---|---|---|
500 | 42 | 26 |
1000 | 85 | 52 |
2000 | 170 | 104 |
Table 6 presents the absolute bias, standard error, root-mean-square error, and coverage probability for n = 500. Web Table 3 and Table 4 of the Supplementary Materials present the results for sample sizes n = 1,000 and 2,000. The tabulated results are plotted in Web Figures 5–7 in the Supplementary Materials. For all three sample sizes, Laplace, Analytic, Efron’s method, and PG have comparable results for bias, standard error, root-mean-square error, and coverage probability. Their bias is small. For all four methods, at n = 500, the magnitude of the bias is
Bias, Standard Error (SE), Root-Mean-Square Error (RMSE), and Coverage Probability (CP) in Data-Based Simulation Study with n = 500. The number of simulations was 10,000.
CoxPH | PG | |||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Laplace | Analytic | Efron’s method | Breslow’s method | |||||||||||||||||
Bias | SE | RMSE | CP | Bias | SE | RMSE | CP | Bias | SE | RMSE | CP | Bias | SE | RMSE | CP | Bias | SE | RMSE | CP | |
age | 0.004 | 0.224 | 1.181 | 0.949 | –0.004 | 0.221 | 1.174 | 0.952 | –0.005 | 0.220 | 1.173 | 0.951 | –0.087 | 0.197 | 1.099 | 0.956 | 0.006 | 0.225 | 1.183 | 0.948 |
education | –0.008 | 0.258 | 0.739 | 0.951 | –0.006 | 0.256 | 0.737 | 0.952 | –0.006 | 0.256 | 0.737 | 0.951 | 0.013 | 0.241 | 0.721 | 0.962 | –0.008 | 0.258 | 0.739 | 0.951 |
gender | –0.012 | 0.281 | 0.910 | 0.949 | –0.009 | 0.279 | 0.907 | 0.950 | –0.010 | 0.280 | 0.908 | 0.949 | 0.022 | 0.263 | 0.880 | 0.960 | –0.013 | 0.282 | 0.911 | 0.949 |
smoke | –0.004 | 0.264 | 0.626 | 0.949 | –0.005 | 0.262 | 0.625 | 0.950 | –0.004 | 0.262 | 0.626 | 0.950 | –0.018 | 0.246 | 0.618 | 0.964 | –0.003 | 0.264 | 0.626 | 0.948 |
CRP | 0.006 | 0.043 | 0.568 | 0.936 | 0.006 | 0.043 | 0.568 | 0.938 | 0.005 | 0.042 | 0.568 | 0.936 | 0.003 | 0.039 | 0.568 | 0.960 | 0.006 | 0.043 | 0.568 | 0.936 |
7 Discussion
The exact method of tie-handling is computationally intractable when there is a large number of ties in the observed event times. Hence, we propose two, new and computationally efficient approximations, Laplace and Analytic. Both approximations performed well in the simulation studies and a real data application to the Cardiovascular Health Study. The Laplace approximation yielded results that were almost identical to that of the PG model for all scenarios. Laplace and PG had low bias and generated confidence intervals with approximately the nominal coverage probability. Our results provide strong evidence that the Cox model, using the Laplace approximation for tie-handling, can serve as a substitute for the PG model in the grouped data setting. The Analytic approximation, given its remarkable simplicity, performed reasonably well in most scenarios, although it tended to have substantial bias for large β, as shown in the Single Covariate Simulations when β = 2. Hence, Analytic may be an attractive option in grouped event time data where the relative risks are not expected to be large. In fact, in the CHS application, the Analytic approximation provided results very similar to PG and Laplace. Our new approximations use a different approach than Breslow’s and Efron’s approximations because the former approximate the exact likelihood with ties, while the latter approximate the actual likelihood. A potential advantage of our approach compared to using the PG model is that fitting the PG model for high-dimensional data might be computationally intensive, relative to our simple approximations.
Both PG and Cox models make the proportional hazards assumption. Although the Cox model estimates the baseline hazard nonparametrically, and the PG model estimates it as the intercept terms, these two are essentially equivalent (with a complementary log-log link in the PG model). Thus, the essential comparison is in terms of different ways of handling the regression part, i. e. the covariate part of the model.
A potential area of future research is to examine the accuracy of our methods under cases with more covariates. Another potential area of future research is to determine the impact of a large number of visits on the results. In this paper, we considered the case of a small number of follow-up visits, e. g. up to 3 follow-up visits in the simulation studies and 2 in the CHS data example.
We would not recommend the use of Breslow’s tie-handling method in the grouped data setting. In simulation, it was observed that Breslow’s method had larger bias than the other methods and can have poor coverage probability, as low as 0.5. In the CHS application, the results of Breslow’s method were also quite different from the other methods.
For those who use the R programming language, we note that the exact method implemented in the “survival” package of R is not the exact method for tie-handling discussed in this paper. The exact method in the “survival” package corresponds to the discrete logistic model proposed by Cox (1972). The discrete logistic model has a different parameter of interest than the models considered in this paper. This is because, rather than making the proportional hazards assumption given in (1), the discrete logistic model assumes that the odds ratios of the hazards are proportional, i. e.
The strong performance of the Cox PH model, using Laplace, Analytic, and Efron’s tie-handling methods, was an impressive and highly surprising finding, given the extremely discrete nature of the event times that we considered. There were 307 and 187 tied event times, respectively, at the first and second follow-up visits of frailty in the CHS study. One would not think that the Cox PH model would be appropriate for estimating relative risks in such a situation. To our knowledge, this type of result has not been reported in the literature before.
Funding source: U.S. Food and Drug Administration
Award Identifier / Grant number: U01 FD004977-01
Funding statement: This work was supported by the U.S. Food and Drug Administration (Funder Id: http://doi.org/10.13039/100000038, Grant Number: U01 FD004977-01)
Funding source: National Institute on Aging
Award Identifier / Grant number: R01AG023629
Funding statement: National Institute on Aging (Funder Id: http://doi.org/10.13039/100000049, Grant Number: R01AG023629)
Funding source: National Institute on Aging
Award Identifier / Grant number: T32AG000247
Funding statement: National Institute on Aging (Funder Id: http://doi.org/10.13039/100000049, Grant Number: T32AG000247)
Appendix
A Derivation of the Laplace Approximation
Denote the integrand of (14) as
We have
Let
Define
Using a Taylor expansion of
Thus, (14) is approximately equal to
where
which is the product of (25) over
which is obtained by taking the logarithm of (26) and omitting the terms that do not depend on
B Derivation of the “Analytic” Approximation
Without loss of generality, we assume that
We approximate the term in the square brackets by
Using this approximation, we have
References
Allison, P. D. (1982). Discrete-time methods for the analysis of event histories. Sociological Methodology, 13:61–98.10.2307/270718Search in Google Scholar
Bild, D. E., Bluemke, D. A., Burke, G. L., Detrano, R., Diez Roux, A. V., Folsom, A. R., et al. (2002). Multi-ethnic study of atherosclerosis: objectives and design. American Journal of Epidemiology, 156:871–881.10.1093/aje/kwf113Search in Google Scholar
Cox, D. R. (1972). Regression models and life-tables. Journal of the Royal Statistical Society: Series B (Methodological), 34:187–202.10.1111/j.2517-6161.1972.tb00899.xSearch in Google Scholar
Davis, P., and Rabinowitz, P. (1984). Methods of Numerical Integration. New York: Dover Publications Mineola.10.1016/B978-0-12-206360-2.50012-1Search in Google Scholar
DeLong, D., Guirguis, G., and So, Y. (1994). Efficient computation of subset selection probabilities with application to Cox regression. Biometrika, 81:607–611.10.1093/biomet/81.3.607Search in Google Scholar
Erdelyi, A. (1956). Asymptotic Expansions. New York: Dover Publications New York.10.21236/AD0055660Search in Google Scholar
Fried, L., Borhani, N., Enright, P., Furberg, C., Gardin, J., Kronmal, R., et al. (1991). The Cardiovascular health study: Design and rationale. Annals of Epidemiology, 1:263–276.10.1016/1047-2797(91)90005-WSearch in Google Scholar
Fried, L., Tangen, C., Walston, J., Newman, A., Hirsch, C., Gottdiener, J. et al. (2001). Frailty in older adults: Evidence for a phenotype. The Journals of Gerontology Series A: Biological Sciences and Medical Sciences, 56:M146–M157.10.1093/gerona/56.3.M146Search in Google Scholar
Hertz-Picciotto, I., and Rockhill, B. (1997). Validity and efficiency of approximation methods for tied survival times in Cox regression. Biometrics, 53:1151–1156.10.2307/2533573Search in Google Scholar
Kalbfleisch, J. D., and R. L. Prentice. (2002). The Statistical Analysis of Failure Time Data. Hoboken, New Jersey: John Wiley & Sons.10.1002/9781118032985Search in Google Scholar
Onder, G., Penninx, B. W., Ferrucci, L., Fried, L. P., Guralnik, J. M., and Pahor, M. (2005). Measures of physical performance and risk for progressive and catastrophic disability: results from the women’s health and aging study. The Journals of Gerontology Series A: Biological Sciences and Medical Sciences, 60:74–79.10.1093/gerona/60.1.74Search in Google Scholar PubMed
Prentice, R., and Gloeckler, L. (1978). Regression analysis of grouped survival data with application to breast cancer data. Biometrics, 34:57–67.10.2307/2529588Search in Google Scholar
Scheike, T., and Sun, Y. (2007). Maximum likelihood estimation for tied survival data under cox regression model via em-algorithm. Lifetime Data Analysis, 13:399–420.10.1007/s10985-007-9043-3Search in Google Scholar PubMed
Therneau, T., and Grambsch, P. (2010) Modeling Survival Data: Extending the Cox Model. New York: Springer-Verlag.Search in Google Scholar
Supplementary Material
The online version of this article offers supplementary material (DOI:https://doi.org/10.1515/em-2018-0011).
© 2019 Walter de Gruyter GmbH, Berlin/Boston
Articles in the same Issue
- Articles
- Analysing Interrupted Time Series with a Control
- Instrumental Variable Estimation with the R Package ivtools
- Identification of Spikes in Time Series
- Modeling of Clinical Phenotypes Assessed at Discrete Study Visits
- The Magnitude and Direction of Collider Bias for Binary Variables
- Causal Mediation Analysis in the Presence of a Misclassified Binary Exposure
Articles in the same Issue
- Articles
- Analysing Interrupted Time Series with a Control
- Instrumental Variable Estimation with the R Package ivtools
- Identification of Spikes in Time Series
- Modeling of Clinical Phenotypes Assessed at Discrete Study Visits
- The Magnitude and Direction of Collider Bias for Binary Variables
- Causal Mediation Analysis in the Presence of a Misclassified Binary Exposure