Abstract
In recent years, the growing availability of biomedical datasets featuring numerous longitudinal covariates has motivated the development of several multi-step methods for the dynamic prediction of survival outcomes. These methods employ either mixed-effects models or multivariate functional principal component analysis to model and summarize the longitudinal covariates’ evolution over time. Then, they use Cox models or random survival forests to predict survival probabilities, using as covariates both baseline variables and the summaries of the longitudinal variables obtained in the previous modelling step. Because these multi-step methods are still quite new, to date little is known about their applicability, limitations, and predictive performance when applied to real-world data. To gain a better understanding of these aspects, we performed a benchmarking of these multi-step methods (and two simpler prediction approaches) using three datasets that differ in sample size, number of longitudinal covariates and length of follow-up. We discuss the different modelling choices made by these methods, and some adjustments that one may need to do in order to be able to apply them to real-world data. Furthermore, we compare their predictive performance using multiple performance measures and landmark times, assess their computing time, and discuss their strengths and limitations.
1 Introduction
Predicting the probability that individuals may experience adverse events (such as the onset of a disease, or death) is an important task of modern medicine. Risk prediction models (RPMs, [1]) are statistical and machine learning models that can be employed to predict the probability that an individual (or patient) will experience an event of interest over time. They can be used by clinicians to monitor disease progression, provide patients with information about their health status and risks, and guide decisions about hospitalization, surgeries and initiating, adjusting or terminating a certain treatment.
The development of a RPM requires the estimation of the probability that individuals will not experience the event over time (called survival probability) as a function of their risk factors (covariates or predictors). Often, all covariates are time-independent (i.e., they do not change over time) or measured at a single point in time. The main advantages of this static modelling approach are that it is relatively easy to implement, and it does not require gathering patient information over time. However, static RPMs do not incorporate any information on individual changes over time that may be predictive of the event of interest. Dynamic RPMs make it possible to include longitudinal (“time-dependent“) covariates (i.e., covariates measured at multiple points in time on the same individual), empowering more accurate predictions in situations where changes in the level of one or more covariates over time are associated with the risk of experiencing the event. Moreover, dynamic pRPMs can be used to dynamically update individual predictions of survival each time new (more recent) measurements become available.
Traditional approaches to the dynamic prediction of survival outcomes include last observation carried forward (LOCF) landmarking [2] and joint modelling [3]. The major advantages of the LOFC approach are its simplicity and ease of implementation; possible disadvantages include the fact that the last observation taken before the landmark may be outdated, and the lack of a model to describe the evolution over time of the longitudinal covariates and to account for measurement error (which may be considerable in the case of biomarkers). Unlike LOCF landmarking, joint models (JMs) enable efficient use of the longitudinal data, and to account for measurement error. However, the estimation of JMs is computationally difficult and demanding, particularly when attempting to include many longitudinal covariates in the JM. To date, high computing times and frequent convergence problems have limited the applicability of JMs to problems with a limited number of longitudinal covariates [4], 5], forcing researchers working on data with numerous longitudinal covariates to preselect a subset of those covariates before fitting the JM.
Over the last decades, technological advancements have led to a growing availability of studies where many longitudinal covariates are measured alongside a survival outcome. Examples include observational studies where patients are monitored for several years to study the occurrence of a slowly progressing disease such as cancer [6], dementia [7] or Alzheimer’s disease [8], as well as studies that seek to identify biomarkers associated with disease progression and the occurrence of disease milestones among a large number of longitudinal omic variables [9], 10]. Recently, this increasing availability of repeated measurement data in biomedical studies has driven the development of new statistical methods [11], [12], [13], [14], [15] and software [16], 17] for dynamic prediction that can better handle the availability of a large number of longitudinal covariates. These methods approach the dynamic prediction problem through a multi-step modelling approach: first, they employ either mixed-effects models (MEMs) or multivariate functional principal component analysis (MFPCA) to describe the evolution over time of the longitudinal covariates and to obtain time-independent summaries of the repeated measurement data; then, they model the relationship between the survival outcome and the available covariates using either Cox models or random survival forests (RSFs).
Because these multi-step methods have been introduced very recently, to date little is known about their applicability to, and performance with, real-world datasets. Practical modelling questions include how well these methods can accommodate real-world datasets with numerous longitudinal covariates and unbalanced measurement times, and how computationally intensive they are (especially when compared to a computationally inexpensive approach such as landmarking, and to a computationally inefficient one such as JM). From an applied perspective, it is also unknown what the predictive performance of the different multi-step methods may be when applied to real-world datasets.
The goal of this study is to investigate these open questions using publicly available real-world datasets. We gathered data from three longitudinal studies that differ substantially in terms of sample size, number of longitudinal covariates and length of follow-up. We employed four multi-step dynamic prediction methods, alongside two simpler prediction approaches, to predict a relevant time-to-event outcome in each dataset, and proceeded to evaluate the predictive performance of each method according to three different metrics for survival data. Although an additional comparison with the JM approach would be interesting, JMs are not included in this benchmarking study because their estimation through state of the art statistical software for JMs (such as joineRML and JMbayes2) repeatedly failed when attempting to include more than three to five longitudinal predictors in the JM.
The remainder of this article is organized as follows: in Section 2 we briefly describe the dynamic prediction problem and the methods considered in the benchmarking study. In Section 3 we present the datasets used for the benchmarking, and provide detailed information on how each prediction method was applied to each dataset. Section 4 presents the results of the benchmarking, which are further discussed in Section 5.
2 Methods
2.1 Dynamic prediction of survival outcomes
We consider the setup of a longitudinal study where n subjects are enrolled, and interest lies in predicting a time-to-event outcome. Each subject i ∈ {1, 2, …, n} is followed from time t = 0 (study entry/baseline visit) until they either experience the event of interest at time t
i
, or they are right-censored at time c
i
. We denote by T
i
and C
i
the random variables corresponding to the event and censoring times, respectively, and let
Between times 0 and
Let
At any point in time
using all the baseline and longitudinal covariate information
When developing dynamic RPMs, it can often be of interest to compute predictions of survival
2.2 Strict and relaxed data landmarking
Equation (1) states that the goal of dynamic prediction is to predict survival probabilities over time for all individuals who are event-free up until the landmark time ℓ, using as predictors all covariate values measured up to time ℓ.
In practice, often repeated measurement data gathered after ℓ are available when estimating a statistical model, raising the question of whether such data should be used for model estimation. Gomon et al. [15] named “relaxed landmarking“ the approach where longitudinal data gathered after the landmark are used to estimate the model on the training set, contrasting it to a “strict landmarking“ method which discards any repeated measurement taken after ℓ in line with equation (1). They showed that using relaxed data landmarking led to a considerable worsening of the predictive performance of MFPCCox, when compared to the same model estimated using strict data landmarking.
At first sight, relaxed data landmarking may appear attractive, because it allows to use more data to estimate the model. However, this approach is problematic for two reasons: first, it uses future information (collected after t > ℓ) to predict the conditional survival probability P(T > t|T > ℓ). Note that such future information is not available at the landmark time, and thus in reality it would not be possible to make use of it to make predictions at ℓ. Furthermore, relaxed data landmarking introduces a selection bias in the modelling of the longitudinal trajectories: because individuals who survive longer are likely to have more repeated measurements taken after ℓ than individuals with shorter survival, the available measurements after the landmark are not a representative sample from the population of individuals who survived up until ℓ. Whereas this is not an issue with joint models, the multi-step approaches considered in this article do not correct for this source of bias.
Despite these issues, in the literature relaxed data landmarking is often used to estimate multi-step dynamic prediction models. Based on the aforementioned remarks and the results of Gomon et al. [15], hereafter we describe and implement the dynamic prediction models included in our comparison following a strict data landmarking approach.
2.3 Dynamic prediction methods
2.3.1 Static Cox model
Before describing the dynamic prediction methods compared in this work, we consider a static Cox model [18] that only uses baseline information to predict
where h i (t) denotes the hazard for subject i at time t, h0(t) is a non-parametric baseline hazard, and α and γ are vectors of regression coefficients. The predicted conditional survival probability of equation (1) can be estimated by computing
where
Admittedly, model (2) does not make use of the available longitudinal measurements to update predictions of survival: nevertheless, it is useful to evaluate the predictive performance of this model to quantify how accurate our predictions of survival would be if we were to ignore any information collected after baseline. By comparing the performance of this static prediction model to that of dynamic methods, which incorporate more and more repeated measurements as the landmark time increases, we can effectively quantify the extent to which adding longitudinal information to a RPM may improve predictive performance. Although the discrepancy between static and dynamic methods will be typically data-dependent, in general we can expect the difference between these two approaches to increase as the landmark time increases, as the information provided to model (2) becomes more and more outdated, while dynamic methods continue being fed more and more up to date patient information.
2.3.2 Landmarking
Landmarking [2] is a pragmatic approach to dynamic prediction that summarizes the trajectory of each longitudinal covariate up until the landmark time into a single summary measure per covariate. Two commonly-used summary measures are the last available observation (up to ℓ) of the longitudinal covariate, an approach referred to as last observation carried forward (LOCF), and the average of all repeated measurements taken up until ℓ. The summaries of the longitudinal covariates are then included as predictors in a Cox model alongside the baseline covariates.
Let
where α and
2.3.3 Multivariate functional principal component analysis Cox model (MFPCCox)
The Multivariate Functional Principal Component Analysis Cox (MFPCCox) model [11], 15] uses Multivariate Functional Principal Component Analysis (MFPCA) to summarize the Q longitudinal covariates into MFPCA scores, which are then used as predictors of survival in a Cox model.
First, MFPCA [19] is employed to approximate the longitudinal trajectory for the q-th longitudinal predictor as the sum of a mean process μ q (t) shared by all subjects, and a finite sum of subject-specific MFPCA scores ρ ki that are shared across the Q longitudinal covariates:
where
This is achieved by assuming that the observed Y q (t) is a noisy realization from an underlying random process R q (t) with mean μ q (t) = E[R q (t)] and covariance function C q (t, t′) = Cov(R q (t), R q (t′)): Y q (t) = R q (t) + ϵ q (t), where ϵ q (t) are i.i.d. Gaussian errors. First, one needs to estimate the mean μ q (t) and covariance C q (t, t′) functions using non-parametric estimators. Then, C q (t, t′) is decomposed to obtain the eigenfunctions, and the corresponding univariate FPCA scores are computed using the Principal component Analysis by Conditional Expectation (PACE) algorithm [20]. Lastly, estimates of the multivariate FPCA scores ρ ki and eigenfunctions ψ kq (t) in equation (5) are computed from the univariate FPCA scores [19].
To predict survival, MFPCCox considers a Cox model where the estimated MFPCA scores are included in the linear predictor alongside the baseline covariates:
where
2.3.4 Penalized regression calibration (PRC)
The Penalized Regression Calibration (PRC) method [12], 16] uses mixed-effects models to summarize the Q longitudinal covariates into predicted random effects, which are then used as predictors of survival in a Cox model.
First, PRC models each longitudinal covariate (or a transformed version of it, for example its log-transform) using a linear mixed effects model (LMM, [21]):
where β q is a vector of population parameters shared across all subjects called fixed effects, u qi is a vector of subject-specific parameters called random effects that are assumed to follow a multivariate normal distribution, ɛ qi is a vector of Gaussian errors, and W qi (t ij ) and Z qi (t ij ) are design matrices containing the covariates relevant for β q and u qi , respectively. Notice that [12] also considered a more complex multivariate mixed model that is not relevant for the datasets considered in this paper.
Specifically, in this article we consider an LMM with a random intercept u qi0 and a random slope u qi1 associated to the time variable t ij :
where
After model (7) has been estimated, PRC proceeds to compute the vector with the predicted random effects
To predict survival, PRC uses a Cox model where the predicted random effects are used as predictors of survival together with the baseline covariates:
where
2.3.5 Functional random survival forest (FunRSF)
The Functional Random Survival Forest (FunRSF) approach [13] uses MFPCA to summarize the trajectories described by the longitudinal covariates, and a random survival forest (RSF, [22]) to predict survival.
The MFPCA modelling step is analogous to the one used by MFPCCox and described in Section 2.3.3. This step yields K estimated MFPCA scores
The estimation of the RSF involves sampling B bootstrap samples from a dataset that comprises information on the survival outcome, the baseline covariates and the MFPCA scores. For each boostrap sample, a random survival tree is estimated using as splitting variables both the baseline covariates, and the MFPCA scores. At each node of the tree, a fixed number F of candidate splitting variables is selected among the union of the P baseline covariates and K MFPCA scores. Each node is split by identifying the candidate variable whose splitting maximizes the log-rank test statistic, and the splitting is continued until a terminal node is reached. A node is terminal if further splitting would lead to child nodes that contain less than a fixed number s of subjects. Each terminal node k in the b-th tree is then summarized by its cumulative hazard function (CHF)
which is then used to estimate
2.3.6 Dynamic random survival forest (DynForest)
The Dynamic Random Survival Forest (DynForest) method [14], 17] makes use of LMMs to summarize the Q longitudinal covariates into predicted random effects that are used as predictors of survival within an RSF.
The estimation of the RSF starts by drawing B bootstrap samples from a dataset that comprises information on the survival outcome, the baseline covariates and the longitudinal covariates. For each bootstrap sample, a survival tree is estimated as follows: at each node, a fixed number F of candidate splitting variables is selected among the union of the P baseline covariates and the Q longitudinal covariates. For each candidate longitudinal covariate, an LMM of the form given in equation (7) is estimated, and the longitudinal trajectory of covariate q for subject i, y
qi
, is then summarized by the predicted random effects
which is then used to estimate
Notice that although the construction of the RSF in DynForest shares similarities with FunRSF, an important difference is that FunRSF uses a single MFPCA computation to summarize the longitudinal covariates into MFPCA scores prior to estimating the RSF, whereas DynForest estimates new LMMs within each tree and node (meaning that the same longitudinal covariate may be summarized by different predicted random effects across trees and nodes). Differently from MFPCCox, PRC and FunRSF, DynForest can handle competing risks (which we do not address this in this article, since most multi-step methods have not yet been extended to the competing risks setting).
2.3.7 Software implementation
All computations presented in this study were performed using R on an AMD EPYC 7662 processor with 2 GHz CPU. After properly subsetting the data as appropriate for the static Cox model and LOCF landmarking model, we proceeded to implement such models in R, relying on the coxph function from R package survival [23] for model estimation. We used on the R packages pencal [16] and DynForest [17] to implement and estimate the PRC and DynForest models. Lastly, to estimate MFPCCox and FunRSF we proceeded to adapt as appropriate the code provided in [11] and [13]; respectively. All scripts created to perform the analyses presented in this paper are publicly available at https://github.com/mirkosignorelli/comparisonDynamicPred.
3 Datasets and implementation
Sections 3.1–3.3 present the three datasets used for the benchmarking. We describe how the data were gathered, the length of follow-up and the frequency of repeated measurements, define the survival outcome that we want to predict, and list the baseline and longitudinal covariates that are employed to predict it. The datasets differ substantially in sample size, length of the follow-up, frequency and number of repeated measurements per subject, and number of longitudinal covariates employed as predictors (Supplementary Table 1), making them representative of different scenarios that one may encounter when dealing with real-world longitudinal studies.
Lastly, in Section 3.4 we provide information about the implementation of the dynamic prediction methods for these datasets, including information about landmark times and prediction horizons, performance measures, and modelling and computational details.
3.1 The ADNI dataset
The Alzheimer’s Disease Neuroimaging Initiative (ADNI) study [8] is an ongoing study started in 2004 that was designed to identify and validate biomarkers related to the progression of AD. By April 2023, the study had enrolled 2,428 individuals, each scheduled to undergo an initial visit and subsequent assessments scheduled at 3, 6, 12, 18 and 24 months from baseline, followed by annual visits thereafter. At each follow-up visit, participants undergo comprehensive evaluations for dementia, including a variety of cognitive assessments, biospecimen sampling and brain imaging analysis.
Differently from the ROSMAP study that will be introduced in Section 3.2, where the cause of dementia (AD or not AD) is reported, the ADNI study only reports dementia diagnoses, without attributing them to AD or other causes. At each visit, ADNI participants are classified into one of the following categories: cognitive normal (CN), mildly cognitive impaired (MCI), or affected by dementia. Moreover, both baseline information about the participant and longitudinal measurements of cognitive, imaging and biochemical markers are collected.
Our goal is to predict the time until a dementia diagnosis for subjects that entered the study as CN or MCI. To predict this survival outcome we employ 5 baseline covariates (age, gender, baseline diagnosis, number of apolipoprotein ϵ4 alleles, and number of years of education) and 21 longitudinal covariates that are listed in Supplementary Table 2 (these covariates were selected based on their relevance as predictors of dementia, provided that they did not have too many missing values). After removing subjects already diagnosed with dementia at baseline, subjects without any follow-up information after baseline, and subjects with missing covariate values at the baseline visit, 1,643 subjects were retained for analysis. The number of visits per subject varies from 1 to 22, with a mean of 6.2 visits. The follow-up length ranges from 3 months to 15.5 years, with a mean of 3.4 years. Supplementary Figure 1 illustrates how the number of individuals at risk evolves over time, and the Kaplan-Meier estimator of the probability to be free from dementia in this dataset.
3.2 The ROSMAP dataset
The Religious Orders Study and Rush Memory and Aging Project (ROSMAP) study [7] is an ongoing project designed to study the onset of Alzheimer’s disease (AD). Started in 1994, by May 2023 ROSMAP had enrolled 3,757 participants, collecting longitudinal information about their cognitive status and possible risk factors for AD.
Every year, subjects enrolled in ROSMAP undergo a clinical assessment that leads to an evaluation of their cognitive status, which is classified into 6 classes: no cognitive impairment (NCI); mild cognitive impairment (MCI); MCI and another condition contributing to cognitive impairment (MCI+); Alzheimer’s dementia (AD); AD and other condition contributing to CI (AD+); other primary cause of dementia, without clinical evidence of AD (Other). In addition, longitudinal information about a wide range of genomic, experiential, psychological and medical risk factors is collected alongside information about the participant’s background (such as age, gender and education).
We aim to predict the time until AD is diagnosed (i.e., cognitive status AD or AD+) for subjects that entered the study in the NCI, MCI, MCI+ and Other categories. To do so, we employ 5 baseline covariates (age, gender, years of education received, baseline cognitive status and presence of a cancer diagnosis) and a set of 30 longitudinal covariates, listed in Supplementary Table 3, which measure aspects related to several different domains (these covariates were selected from a wider list of longitudinal covariates based on their relevance as risk factors for AD, provided that they did not have too many missing values). After removing subjects already diagnosed with AD at baseline, subjects without any follow-up information after baseline, and subjects with missing covariate values at baseline, 3,293 subjects were retained for analysis. The number of visits per subject varies from 2 to 30, with a mean of 9.3 visits. The follow-up period ranges from 1 to 29 years, with a mean of 8.3 years. Supplementary Figure 2 illustrates how the number of individuals at risk evolves over time, and the Kaplan-Meier estimator of the probability to be free of AD in the ROSMAP dataset.
3.3 The PBC2 dataset
The PBC2 dataset contains data from a clinical trial on primary biliary cholangitis (PBC) conducted between 1974 and 1984 at the Mayo Clinic [24]. The randomized trial involved 312 participants who were randomized between a placebo group, and a treatment group where patients received the drug D-penicillamine. The trial recorded the occurrence of the first of two survival outcomes: liver transplantation, and death. Because most of the multi-step methods considered in our comparison cannot deal with competing risks, for the purpose of our analysis we focus on predicting time to death, treating patients who underwent liver transplantation as right-censored at the date of transplantation.
To predict time to death, we employ 3 baseline covariates (age, gender and treatment group) alongside the 8 longitudinal covariates listed in Supplementary Table 4. Patients enrolled in the trial underwent visits upon study entry, 6 and 12 months after baseline, and yearly visits after that. The number of follow-up visits per patient ranged from 1 to 22, with an average of 6.1; the follow-up period ranged from 0.1 to 14.1 years, with a mean of 4.6 years. Supplementary Figure 3 shows the number of patients at risk and the Kaplan-Meier estimator of the survival probability for the PBC2 dataset.
3.4 Application of the dynamic prediction methods to the ADNI, ROSMAP and PBC2 datasets
3.4.1 Definition of the landmark and horizon times
The datasets considered in our study differ substantially in their sample size, frequency of the repeated measurements (visits) and length of the follow-up. All these aspects have an influence on the number of individuals still at risk at a given landmark, and the number of repeated measurements per subject available up to the landmark time. Lastly, the length of the follow-up across patients affects the possibility to evaluate predictive performance at a given horizon time.
Given all these differences, we determined the landmark and horizon times at which predictions of survival are evaluated separately for each dataset, striving to balance the need to have a good number of repeated measurements to model the evolution of the longitudinal covariates with MFPCA or LMMs, and that of having a good number of individuals still at risk to be able to estimate the Cox/RSF model and reliably assess predictive performance. Clearly, this is easier for the ADNI and ROSMAP datasets, whose sample sizes are considerably bigger than that of PBC2.
The landmarks considered in our analysis are the following:
ADNI: 2, 3 and 4 years from baseline;
ROSMAP: 2, 3, 4, 5 and 6 years from baseline;
PBC2: 2.5, 3 and 3.5 years from baseline.
The accuracy of predictions was evaluated for every year after the landmark up to 10 years from baseline for ADNI, up to 15 years from baseline for ROSMAP, and up to 8 years from baseline for PBC2.
3.4.2 Performance measures and validation of predictive performance
To evaluate predictive performance, we consider the three most commonly-used measures of predictiveness in survival analysis, namely the Brier score [25], the time-dependent AUC [26], and the C index [27].
We employed repeated cross-validation (RCV) to obtain unbiased estimates of the predictive performance of the different models. For the ROSMAP and ADNI datasets we employed 10-fold cross-validation (CV) with 10 repetitions; for the PBC2 dataset, where using 10-fold CV led to very small validation folds, we employed 5-fold CV with 20 repetitions.
3.4.3 Implementation details
To ensure that every method was trained on data from the same set of subjects, in each dataset we excluded subjects for which (a) the value of one or more baseline covariates was missing or (b) the value of one of the longitudinal predictors at the baseline visit was missing. Condition (a) is needed by all methods to ensure that there are no missing values in the baseline covariates, whereas (b) aims to ensure that are no missing values when estimating the static Cox model (where we use the measurements of the longitudinal covariates at the baseline visit as baseline covariates). Besides this data selection step, the four multi-step methods required some additional preprocessing and modelling choices that are specific to each method.
MFPCCox and FunRSF rely on MFPCA to summarize the longitudinal covariates into MFPCA scores. In principle, MFPCA can handle longitudinal data where the measurement times differ across subjects. However, highly-irregular measurement times complicate the estimation and inversion of the univariate covariance functions C q (t, t′), often leading to numeric errors when attempting to perform the MFPCA step as implemented in MFPCCox and FunRSF. Given that visits in the ADNI and ROSMAP datasets follow a regular measurement grid, but in a few instances the actual visit takes place a few days before or after the planned date, we proceeded to align the measurement times across subjects so as to obtain a regular measurement grid that allowed to solve most of the problems with the estimation of the MFPCA scores. Furthermore, the MFPCA step within MFPCCox and FunRSF requires the specification of the minimum percentage of variance explained (PVE) to retain when reducing the longitudinal covariates to univariate FPCA scores (PVE1), and then from univariate to multivariate FPCA scores (PVE2). We proceeded to set PVE1 = PVE2 = 90 % for both methods.
DynForest and PRC employ LMMs to summarize the longitudinal covariates into predicted random effects. When the response variable is strongly skewed, this can often lead to estimation (convergence) problems. To reduce the frequency of such convergence problems and the need to address convergence issues with ad-hoc solutions each time they arise during the RCV, we proceeded to apply a log-transformation to those longitudinal covariates whose skewness index was larger than 1, and a cubic transformation to those with skewness index below −1.
PRC allows to choose between the ridge, elasticnet and lasso penalty for the estimation of the Cox model in equation (9). In this study we used the ridge penalty, which is the default option in the pencal package and the recommended choice in [12].
Lastly, the RSF algorithms used by FunRSF and DynForest require the specification of multiple parameters. As concerns FunRSF, the random forests were trained using an ensemble of B = 1,000 trees. The minimum number of subject in a terminal node s was required to be at least 15 for all datasets. The number of predictors F to be randomly selected at each node was set to the square root of the total number of predictors available in the given dataset. The random forests within DynForest were trained using B = 200 trees because estimating DynForest is much more time-consuming than estimating FunRSF (see Section 4.4). Choosing the F, s and e parameters for this method is more complicated than for FunRSF, because the need to re-estimate the LMMs after each node split often leads to convergence problems within DynForest. To reduce the frequency of convergence problems during the RCV, multiple values of s (listed in Supplementary Table 5) were used sequentially in case of convergence problems, progressively increasing s to reduce the chance of encountering convergence errors towards the end of the tree. The value of F was set in the same way as for FunRSF; e was set equal to 5 for ROSMAP and ADNI, and to 4 for PBC2.
4 Results
We now present the results of our benchmarking study. In Sections 4.1–4.3 we compare the predictive performance of the multi-step dynamic prediction methods on the ADNI, ROSMAP and PBC2 datasets, respectively. In Section 4.4, we turn our attention to computing time.
4.1 Predictive performance on the ADNI data
The RCV estimates of the C index, tdAUC and Brier score for the dynamic prediction of time to dementia in the ADNI study are presented in Table 1 and Figures 1 and 2.
Cross-validated C index estimates (and standard deviation) for the ADNI dataset. For each landmark, the top two methods are highlighted in bold.
| Landmark | |||
|---|---|---|---|
| Method | 2 | 3 | 4 |
| Static Cox | 0.901 (0.002) | 0.885 (0.006) | 0.856 (0.008) |
| Landmarking | 0.906 (0.001) | 0.890 (0.004) | 0.855 (0.009) |
| MFPCCox | 0.889 (0.003) | 0.872 (0.009) | 0.859 (0.008) |
| PRC | 0.913 (0.001) | 0.908 (0.003) | 0.904 (0.003) |
| FunRSF | 0.873 (0.004) | 0.858 (0.011) | 0.845 (0.012) |
| DynForest | 0.891 (0.003) | 0.883 (0.005) | 0.871 (0.011) |

Cross-validation estimates of the time-dependent AUC for the prediction of time to dementia in the ADNI dataset. The corresponding numeric values can be found in Supplementary Table 6.

Cross-validated Brier score estimates for the prediction of time to dementia in the ADNI dataset. The corresponding numeric values can be found in Supplementary Table 7.
C index. As shown in Table 1, for all landmark times PRC consistently achieves the highest performance, and FunRSF the lowest one. The relative performance of the remaining methods varies slightly across landmarks: at landmarks 2 and 3, landmarking and static Cox have the second and third highest C index, and are followed by DynForest and MFPCCox. Instead, at landmark 4 DynForest is the second best method. Overall, the relative performance of DynForest and MFPCCox improves as the landmark time increases, whereas that of landmarking and static Cox worsens.
Time-dependent AUC (Figure 1). The results for the tdAUC align closely with the C index. PRC is again the best performing method, and FunRSF the worse, at each landmark time. Landmarking is the second best at landmark 2, and DynForest at 4; the two methods have similar performance at landmark 3. MFPCCox and FunRSF typically show lower accuracy. Once again we can observe that when compared to the multi-step methods, the performance of LOCF landmarking and of the Static Cox model tends to worsen as the landmark increases.
Brier score (Figure 2). The differences across methods appear to be smaller when looking at the Brier score. At all landmarks, the static Cox model performs quite well for predictions until year 7; from year 8, its performance starts to worsen in comparison to other methods. PRC is usually the best or second best performing method, followed by landmarking, DynForest and MFPCCox. Once again, FunRSF performs worse than all other methods at every landmark time.
Effect of sample size and number of predictors on predictive performance. To investigate the effect of the sample size n on predictive performance, we proceeded to fit the same models on a third and two thirds of the available observations, considering the first landmark (Supplementary Table 12 and Supplementary Figures 4 and 5). Moreover, we investigated the effect of the number of predictors P and Q on predictive performance by refitting the same models using a third and two thirds of the covariates considered in the full analysis (Supplementary Table 13 and Supplementary Figures 6 and 7).
Overall, the sample size does not affect substantially the cross-validated values of the C index, time-dependent AUC and Brier score, however we can notice that the standard deviation over the repetitions of the CV decreases substantially with n; this reflects the fact that as the sample size increases, more information is available for accurate evaluation of the predictive performance. On the other hand, increasing the number of predictors leads to higher predictive performance, producing slight increases in the C index and tdAUC, and reductions in the Brier score. This is in line with the expectation that using more predictors can lead to gains in predictive performance (although in this specific dataset the gains appear to be rather small, suggesting that it is already possible to achieve a high predictive performance with a rather small number of predictors).
Summary. The performance results for the ADNI dataset point out some interesting trends:
landmarking and the static Cox model have a good predictive performance at earlier landmark times (especially at ℓ = 2). Surprisingly, these methods often outperform MFPCCox, and systematically outperform FunRSF;
the two multi-step methods that rely on mixed-effects models (PRC and DynForest) typically outperform those that use MFPCA (MFPCCox and FunRSF);
conditionally on the method (LMM, or MFPCA) used to summarize the longitudinal predictors, the methods that use a Cox model for the survival outcome (PRC and MFPCCox) outperform the corresponding methods that use an RSF (DynForest and FunRSF);
overall, the approach used to model the longitudinal predictors has a stronger impact on predictive performance than the approach used to model the survival outcome.
4.2 Predictive performance on the ROSMAP data
The RCV estimates of the C index, tdAUC and Brier score for the application to the ROSMAP data are presented in Table 2 and Figures 3 and 4. At landmarks 2 and 5, no results were produced for DynForest due to the fact that estimation of this model failed on all or most (100 and 97/100, respectively) folds during the repeated CV. Similarly, at landmark 5 no results are presented for MFPCCox and FunRSF as computation of the MFPCA scores failed on 97 of the 100 folds. These convergence issues are due to a higher percentage of missing visit/values in this dataset than in the ADNI dataset, which complicates the modelling of the longitudinal predictors.
Cross-validated C index estimates (and standard deviation) for the ROSMAP dataset. For each landmark, the top two methods are highlighted in bold.
| Landmark | |||||
|---|---|---|---|---|---|
| Method | 2 | 3 | 4 | 5 | 6 |
| Static Cox | 0.824 (0.003) | 0.812 (0.003) | 0.811 (0.003) | 0.799 (0.005) | 0.786 (0.004) |
| Landmarking | 0.844 (0.002) | 0.849 (0.002) | 0.851 (0.001) | 0.854 (0.001) | 0.860 (0.001) |
| MFPCCox | 0.837 (0.003) | 0.845 (0.007) | 0.843 (0.027) | – | 0.842 (0.026) |
| PRC | 0.846 (0.001) | 0.857 (0.001) | 0.864 (0.002) | 0.862 (0.001) | 0.868 (0.002) |
| FunRSF | 0.806 (0.002) | 0.815 (0.005) | 0.814 (0.024) | – | 0.808 (0.022) |
| DynForest | – | 0.835 (0.027) | 0.836 (0.018) | – | 0.853 (0.003) |

Cross-validation estimates of the time-dependent AUC for the prediction of time to AD in the ROSMAP dataset. The corresponding numeric values can be found in Supplementary Table 8.

Cross-validated Brier score estimates for the prediction of time to AD in the ROSMAP dataset. The corresponding numeric values can be found in Supplementary Table 9.
C index (Table 2). PRC and LOCF landmarking appear to be the best performing methods at all landmark times. The difference in performance between the two methods is negligible at landmark 2, and somewhat larger at all other landmarks. MFPCCox ranks third at landmarks 2, 3 and 4, and fourth at landmark 6; DynForest ranks fourth at landmarks 3 and 4, and third at landmark 6. FunRSF and the Static Cox model exhibit the worse predictive performance at all landmarks.
Time-dependent AUC (Figure 3). When looking at the tdAUC estimates, we notice that the performance difference across methods is larger at the later landmark times. Differently from the ADNI data, here we can observe that all multi-step methods and landmarking improve over the static Cox model, suggesting that updating prediction based on the longitudinal data may be more beneficial for the ROSMAP data. We can observe that PRC, DynForest and LOCF landmarking are the top performing methods across all landmarks, with PRC usually outperforming the other two methods on the later horizon times. MFPCCox is usually the fourth best performing method, followed by FunRSF and the static Cox model.
Brier score (Figure 4). Similarly to the tdAUC, also for the Brier score estimates the difference in performance across methods increases with the landmark. Moreover, the relative performance of the methods is fairly consistent across landmarks. PRC and LOCF landmarking exhibit the lowest Brier scores, followed by MFPCCox and DynForest. Once again, FunRSF and the static Cox model have the worst predictive performance
Summary. On the ROSMAP data, incorporating repeated measurements data in the RPM seems to improve predictive performance more than in the ADNI dataset. The relative performance of the prediction models is somewhat similar to the one observed with the ADNI data, with PRC and LOCF landmarking performing particularly well, followed by DynForest and MFPCCox. Once again, FunRSF appears to be the worst performing method among the multi-step methods considered in this benchmarking.
4.3 Predictive performance on the PBC2 data
The RCV estimates of the C index, tdAUC and Brier score for the PBC2 dataset are presented in Table 3 and Figures 5 and 6.
Cross-validated C index estimates (and standard deviation) for the PBC2 dataset. For each landmark, the top two methods are highlighted in bold.
| Landmark | |||
|---|---|---|---|
| Method | 2.5 | 3 | 3.5 |
| Static Cox | 0.813 (0.007) | 0.791 (0.007) | 0.788 (0.010) |
| Landmarking | 0.798 (0.009) | 0.782 (0.011) | 0.796 (0.016) |
| MFPCCox | 0.753 (0.022) | 0.786 (0.041) | 0.781 (0.038) |
| PRC | 0.826 (0.011) | 0.809 (0.010) | 0.821 (0.009) |
| FunRSF | 0.760 (0.017) | 0.732 (0.05) | 0.749 (0.035) |
| DynForest | 0.796 (0.012) | 0.766 (0.009) | 0.782 (0.015) |

Cross-validation estimates of the time-dependent AUC for the prediction of time to death in the PBC2 dataset. The corresponding numeric values can be found in Supplementary Table 10.

Cross-validated Brier score estimates for the prediction of time to death in the PBC2 dataset. The corresponding numeric values can be found in Supplementary Table 11.
C index (Table 3). At all landmark times, PRC achieves the highest value of the C index. Interestingly, the static Cox model is the second best performing method at landmarks 2.5 and 3 and third best at landmark 3.5, suggesting that the added predictive value of the longitudinal measurements may be more limited for the PBC2 dataset. LOCF landmarking is the third bist method at landmarks 2.5 and 3, and second best at landmark 3.5. It is followed by DynForest, MFPCCox and FunRSF. By comparing the standard errors in Tables 1–3, we can see that the standard errors around the estimated C index are larger on PBC2 than on ADNI and ROSMAP. This result can be ascribed to the substantially smaller sample size of this dataset.
Time-dependent AUC (Figure 5). PRC achieves the highest tdAUC values at all landmarks. At landmark 2, it is followed by the Static Cox model, DynForest and LOCF landmarking; FunRSF and MFPCCox have considerably lower predictive accuracy. At landmark 3, DynForest, LOCF landmarking and MFPCCox achieve similar tdAUC values. At landmark 3.5, the trends are somewhat harder to extrapolate. Lastly, at all landmarks the performance of the static Cox model deteriorates quickly as the prediction horizon increases.
Brier score (Figure 6). At landmark 2.5, the estimated Brier scores for landmarking, static Cox, PRC and DynForest are very similar, especially at the earlier prediction horizons. The two methods that rely on MFPCA are once again outperformed by all other methods. At landmark 3, the static Cox model has lower Brier score for predictions after 1 and 2 years from the landmark, and PRC for predictions from 3 years from the landmark onwards. At landmark 3.5, the differences between methods tend to be small for horizons at 4.5 and 5.5 years, after which the lowest Brier scores are achieved by landmarking and PRC, and the highest by FunRSF and the static Cox model.
Summary. The estimates of predictive performance for the PBC2 study are generally less accurate than for the ADNI and ROSMAP datasets due to the smaller sample size. Despite this, the relative performance of the different methods is to a wide extent in line with the patterns observed on the ADNI and ROSMAP datasets.
4.4 Computing time
We now turn our attention to computing time. Tables 4–6 report the average computing time required to estimate each model in k − 1 folds and to obtain predictions for the remaining fold (averaged over all RCV iterations). Computing time is generally higher for the ROSMAP dataset, where we have the largest sample size and number of predictors, and lowest for the PBC2 dataset where n, P and Q are small. By comparing computing time across landmarks, we observe that the computing time typically decreases as the landmark increases; this is simply an artifact of the fact that at later landmarks less individuals are still at risk, reducing the sample size used for model training and prediction.
Average computing time per CV fold (in minutes) for the ADNI dataset.
| Landmark | ||||
|---|---|---|---|---|
| Method | 2 | 3 | 4 | Average |
| Static Cox | 0.009 | 0.007 | 0.006 | 0.007 |
| Landmarking | 0.010 | 0.007 | 0.006 | 0.008 |
| MFPCCox | 0.080 | 0.046 | 0.046 | 0.057 |
| PRC | 0.776 | 0.482 | 0.453 | 0.571 |
| FunRSF | 0.240 | 0.122 | 0.125 | 0.163 |
| DynForest | 12.501 | 9.077 | 8.099 | 9.892 |
Average computing time per CV fold (in minutes) for the ROSMAP dataset.
| Landmark | ||||||
|---|---|---|---|---|---|---|
| Method | 2 | 3 | 4 | 5 | 6 | Average |
| Static Cox | 0.022 | 0.019 | 0.016 | 0.014 | 0.011 | 0.017 |
| Landmarking | 0.023 | 0.020 | 0.017 | 0.014 | 0.011 | 0.017 |
| MFPCCox | 0.194 | 0.135 | 0.066 | – | 0.064 | 0.115 |
| PRC | 3.755 | 1.151 | 1.138 | 1.115 | 1.124 | 1.657 |
| FunRSF | 0.604 | 0.367 | 0.115 | – | 0.124 | 0.302 |
| DynForest | – | 11.504 | 17.657 | – | 34.147 | 21.102 |
Average computing time per CV fold (in seconds) for the PBC2 dataset.
| Landmark | ||||
|---|---|---|---|---|
| Method | 2.5 | 3 | 3.5 | Average |
| Static Cox | 0.253 | 0.246 | 0.177 | 0.226 |
| Landmarking | 0.260 | 0.240 | 0.188 | 0.229 |
| MFPCCox | 0.902 | 0.522 | 0.556 | 0.660 |
| PRC | 7.832 | 7.616 | 7.762 | 7.737 |
| FunRSF | 2.776 | 1.393 | 1.526 | 1.898 |
| DynForest | 186.136 | 136.747 | 128.914 | 150.599 |
Estimation of models that do not model the longitudinal data, i.e. the static Cox and LOCF landmarking models, is almost immediate: this is because these approaches only require to estimate a Cox model through R’s function coxph(). The multi-step methods are more computationally intensive. Among them, MFPCCox is the fastest method, followed by FunRSF. This shows that multi-step methods that use MFPCA are faster than those that use LMMs. The faster computing time of MFPCCox is likely due to the use of a Cox model, whose estimation is faster than that of RSF (used by FunRSF). Lastly, we can observe that among the LMM-based methods, PRC is substantially faster (between 12 and 20 times, depending on the dataset) than DynForest. This last result is mainly due to the fact that in DynForest, the LMMs are re-estimated at each node split within each random survival tree, making this method considerably slower than all other alternatives.
5 Discussion
This study aimed to benchmark four recently-proposed multi-step dynamic prediction methods (MFPCCox, PRC, FunRSF, and DynForest) that are capable of using numerous longitudinal predictors to predict survival. Each method uses a different combination of MFPCA or LMMs to model the longitudinal predictors, and of Cox models or RSF to predict the survival outcome: therefore, by comparing these methods we can also get a feeling of the extent to which the choices of using MFPCA or LMMs, and a Cox model or an RSF, can affect predictive performance. Hereafter we summarize the findings and limitations of this study, focusing our attention on the modelling flexibility, predictive performance, and ease of use of the different methods.
5.1 Modelling flexibility
The four multi-step methods reviewed in this article represent an important and valuable addition to the dynamic prediction toolbox. Differently from traditional landmarking methods, multi-step dynamic prediction methods employ models that enable efficient modelling of the longitudinal data, and to account for the presence of measurement errors. These methods are particularly suitable for datasets where tens, hundreds or (potentially) thousands of longitudinal predictors may be available, where the estimation of JMs is not only very computationally intensive, but usually also unfeasible.
At the same time, when modelling real-world longitudinal and survival data it may sometimes necessary to deal with competing risks, interval censoring, and missing data. Each of these features can lead to complications with some of the multi-step methods considered in this work, at least in their present state of development:
whereas DynForest has been developed within a competing risks framework, MFPCCox, FunRSF and PRC don’t allow to model competing risks directly. While extending them to competing risks is possible, currently users would need to do that by themselves;
none of the four methods currently allows interval censored survival outcomes, which are common in longitudinal studies. Therefore, users need to simplify the survival outcome to right-censored;
PRC and DynForest are currently limited to LMMs (and multivariate LMMs in the case of PRC). Extending them to GLMMs would enable a more appropriate modelling of skewed as well as discrete longitudinal covariates, although it can be anticipated that such an extension would be computationally more complex, potentially leading to more frequent problems during the estimation of the GLMMs and the computation of the predicted random effects;
as seen on the ROSMAP dataset, a high percentage of missingness in the longitudinal data can considerably complicate the estimation of MFPCCox, FunRSF and DynForest. For MFPCCox and FunRSF, this makes it harder to estimate the covariance matrices during the univariate FPCA steps. For DynForest, this creates problems with the estimation of LMMs in children nodes with a small number of subjects.
5.2 Predictive performance
The results presented in Section 4 are mostly consistent across datasets and performance measures. The most interesting patterns are:
multi-step methods that use LMMs (PRC and DynForest) clearly outperform methods that use MFPCA (MFPCCox and FunRSF) to model the trajectories of the longitudinal covariates. PRC is frequently the best performing method with respect to the different metrics considered; it is usually followed by DynForest and MFPCCox, while FunRSF almost always exhibits the worst predictive performance;
conditionally on the same method being used to summarize the longitudinal covariates, the methods that use a Cox model (PRC and MFPCCox) tend to outperform the corresponding method that uses RSF (DynForest and FunRSF, respectively). However, the difference is substantially smaller than the one between LMM-based and MFPCA-based methods, and it may be dataset specific, as the performance of RSF may improve on datasets with larger sample sizes;
interestingly, LOCF landmarking often showed a good predictive performance, often outperforming some of the multi-step methods. This suggests that this simple and intuitive prediction method can often deliver good predictions of survival. Thus, we recommend that practitioners interested in the use of multi-step methodsir always compare the predictive performance to that of LOCF landmarking, because the latter approach may be preferable in case of similar predictive performance;
the extent to which longitudinal data may improve the predictive accuracy of a dynamic RPM over a static RPM varies across datasets. To quantify the extent of this improvement for a specific dataset, we suggest comparing the predictive performance of the developed dynamic RPM to that of a simpler static RPM that only uses baseline information.
5.3 Software and computing time
Software availability and computing time are two matters of practical importance when attempting to estimate and implement a RPM, affecting the ease with which a practitioner may be able to estimate a RPM and employ it to compute predictions.
PRC and DynForest come with fairly advanced software implementations and documentation [16], 17] that make it easier for users to implement those methods. On the contrary, MFPCCox and FunRSF lack software implementations and documentation: this forces users to reuse the code presented in the corresponding publications, and often requires them to adapt it to deal with missing data and irregular measurement grids.
MFPCCox and FunRSF are computationally inexpensive: their estimation takes less than a minute on all three datasets considered in our benchmarking. PRC is a bit more intensive, taking less than a minute on PBC2 and ADNI, and approximately 1.66 min on ROSMAP. Finally, DynForest is by far the slowest method: its estimation takes about 2.5 min on PBC2, 10 on ADNI, and 21 on ROSMAP.
Supplementary Material
Supplementary tables and figures are available in the “Supplementary Material” pdf file.
Acknowledgements
ROSMAP. Data from the ROSMAP study were obtained from the Rush Alzheimer’s Disease Center (RADC) of the RUSH University (https://www.radc.rush.edu). ROSMAP is supported by P30AG10161, P30AG72975, R01AG15819, R01AG17917, U01AG46152, and U01AG61356. ADNI. Data from the ADNI study were obtained from the ADNI database (http://adni.loni.usc.edu). As such, the investigators within the ADNI contributed to the design and implementation of ADNI and/or provided data but did not participate in analysis or writing of this paper. A complete listing of ADNI investigators can be found at https://adni.loni.usc.edu/wp-content/uploads/how_to_apply/ADNI_Acknowledgement_List.pdf.
-
Research ethics: Not applicable.
-
Informed consent: Not applicable.
-
Author contributions: MS and SR conceived and designed the study. SR implemented the analyses under the supervision of MS. MS and SR interpreted the results of the study. MS wrote the manuscript. All authors have accepted responsibility for the entire content of this manuscript and approved its submission.
-
Use of Large Language Models, AI and Machine Learning Tools: None declared.
-
Conflict of interest: The authors state no conflict of interest.
-
Research funding: None declared.
-
Data availability: The R code used to perform the analyses described in this paper is available at https://github.com/mirkosignorelli/comparisonDynamicPred. A copy of the PBC2 data the same URL. The data from the ROSMAP study are available upon motivated request from the Rush Alzheimer’s Disease Center (RADC) of the RUSH University. ROSMAP resources can be requested at https://www.radc.rush.edu and www.synpase.org. The ADNI data are available upon motivated request from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (http://adni.loni.usc.edu).
References
1. Steyerberg, EW. Clinical prediction models. Springer; 2009.10.1007/978-0-387-77244-8Search in Google Scholar
2. Van Houwelingen, HC. Dynamic prediction by landmarking in event history analysis. Scand J Stat 2007;34:70–85. https://doi.org/10.1111/j.1467-9469.2006.00529.x.Search in Google Scholar
3. Hickey, GL, Philipson, P, Jorgensen, A, Kolamunnage-Dona, R. Joint modelling of time-to-event and multivariate longitudinal outcomes: recent developments and issues. BMC Med Res Methodol 2016;16:1–15. https://doi.org/10.1186/s12874-016-0212-5.Search in Google Scholar PubMed PubMed Central
4. Fournier, M-C, Foucher, Y, Blanche, P, Legendre, C, Girerd, S, Ladrière, M, et al.. Dynamic predictions of long-term kidney graft failure: an information tool promoting patient-centred care. Nephrol Dial Transplant 2019;34:1961–9. https://doi.org/10.1093/ndt/gfz027.Search in Google Scholar PubMed
5. Spreafico, M, Ieva, F. Dynamic monitoring of the effects of adherence to medication on survival in heart failure patients: a joint modeling approach exploiting time-varying covariates. Biom J 2021;63:305–22. https://doi.org/10.1002/bimj.201900365.Search in Google Scholar PubMed
6. Patel, AV, Jacobs, EJ, Dudas, DM, Briggs, PJ, Lichtman, CJ, Bain, EB, et al.. The American Cancer Society’s Cancer Prevention Study 3 (CPS-3): recruitment, study design, and baseline characteristics. Cancer 2017;123:2014–24. https://doi.org/10.1002/cncr.30561.Search in Google Scholar PubMed
7. Bennett, DA, Buchman, AS, Boyle, PA, Barnes, LL, Wilson, RS, Schneider, JA. Religious orders study and rush memory and aging project. J Alzheim Dis 2018;64:S161–S1 https://doi.org/10.3233/jad-179939.Search in Google Scholar
8. Weiner, MW, Aisen, PS, Jack Jr, CR, Jagust, WJ, Trojanowski, JQ, Shaw, L, et al.. The Alzheimer’s disease neuroimaging initiative: progress report and future plans. Alzheimer’s Dementia 2010;6:202–11. https://doi.org/10.1016/j.jalz.2010.03.007.Search in Google Scholar PubMed PubMed Central
9. Signorelli, M, Ayoglu, B, Johansson, C, Lochmüller, H, Straub, V, Muntoni, F, et al.. Longitudinal serum biomarker screening identifies malate dehydrogenase 2 as candidate prognostic biomarker for Duchenne muscular dystrophy. J Cachexia, Sarcopenia Mscle 2020;11:505–17. https://doi.org/10.1002/jcsm.12517.Search in Google Scholar PubMed PubMed Central
10. Filbin, MR, Mehta, A, Schneider, AM, Kays, KR, Guess, JR, Gentili, M, et al.. Longitudinal proteomic analysis of severe Covid-19 reveals survival-associated signatures, tissue-specific cell death, and cell-cell interactions. Cell Rep Med 2021;2. https://doi.org/10.1016/j.xcrm.2021.100287.Search in Google Scholar PubMed PubMed Central
11. Li, K, Luo, S. Dynamic prediction of Alzheimer’s disease progression using features of multiple longitudinal outcomes and time-to-event data. Stat Med 2019;38:4804–18. https://doi.org/10.1002/sim.8334.Search in Google Scholar PubMed PubMed Central
12. Signorelli, M, Spitali, P, Szigyarto, CA-K, Consortium, M-M, Tsonaka, R. Penalized regression calibration: a method for the prediction of survival outcomes using complex longitudinal and high-dimensional data. Stat Med 2021;40:6178–96. https://doi.org/10.1002/sim.9178.Search in Google Scholar PubMed PubMed Central
13. Lin, J, Li, K, Luo, S. Functional survival forests for multivariate longitudinal outcomes: dynamic prediction of Alzheimer’s disease progression. Stat Methods Med Res 2021;30:99–111. https://doi.org/10.1177/0962280220941532.Search in Google Scholar PubMed PubMed Central
14. Devaux, A, Helmer, C, Genuer, R, Proust-Lima, C. Random survival forests with multivariate longitudinal endogenous covariates. Stat Methods Med Res 2023;32:2331–46. https://doi.org/10.1177/09622802231206477.Search in Google Scholar PubMed
15. Gomon, D, Putter, H, Fiocco, M, Signorelli, M. Dynamic prediction of survival using multivariate functional principal component analysis: a strict landmarking approach. Stat Methods Med Res 2024;33:256–72. https://doi.org/10.1177/09622802231224631.Search in Google Scholar PubMed PubMed Central
16. Signorelli, M. Pencal: an R package for the dynamic prediction of survival with many longitudinal predictors. arXiv preprint arXiv:2309.15600 2024.10.32614/RJ-2024-014Search in Google Scholar
17. Devaux, A, Proust-Lima, C, Genuer, R. Random Forests for time-fixed and time-dependent predictors: the DynForest R package. arXiv preprint arXiv:2302.02670 2024.10.32614/RJ-2025-002Search in Google Scholar
18. Cox, DR. Regression models and life-tables. J Roy Stat Soc: Ser B (Methodological) 1972;34:187–202. https://doi.org/10.1111/j.2517-6161.1972.tb00899.x.Search in Google Scholar
19. Happ, C, Greven, S. Multivariate functional principal component analysis for data observed on different (dimensional) domains. J Am Stat Assoc 2018;113:649–59. https://doi.org/10.1080/01621459.2016.1273115.Search in Google Scholar
20. Yao, F, Müller, H-G, Wang, J-L. Functional data analysis for sparse longitudinal data. J Am Stat Assoc 2005;100:577–90. https://doi.org/10.1198/016214504000001745.Search in Google Scholar
21. McCulloch, CE, Searle, SR. Generalized, linear, and mixed models. John Wiley & Sons; 2004.10.1002/0470011815.b2a10021Search in Google Scholar
22. Iswaran, H, Kogalur, U, Blackstone, E, Lauer, M. Random survival forests. Ann Appl Stat 2008;2:841–60. https://doi.org/10.1214/08-aoas169.Search in Google Scholar
23. Therneau, TM, Grambsch, PM. Modeling survival data: extending the Cox model. Springer; 2000.10.1007/978-1-4757-3294-8Search in Google Scholar
24. Murtaugh, PA, Dickson, ER, Van Dam, GM, Malinchoc, M, Grambsch, PM, Langworthy, AL, et al.. Primary biliary cirrhosis: prediction of short-term survival based on repeated patient visits. Hepatology 1994;20:126–34. https://doi.org/10.1002/hep.1840200120.Search in Google Scholar
25. Graf, E, Schmoor, C, Sauerbrei, W, Schumacher, M. Assessment and comparison of prognostic classification schemes for survival data. Stat Med 1999;18:2529–45. https://doi.org/10.1002/(sici)1097-0258(19990915/30)18:17/18<2529::aid-sim274>3.0.co;2-5.10.1002/(SICI)1097-0258(19990915/30)18:17/18<2529::AID-SIM274>3.3.CO;2-XSearch in Google Scholar
26. Heagerty, PJ, Lumley, T, Pepe, MS. Time-dependent ROC curves for censored survival data and a diagnostic marker. Biometrics 2000;56:337–44. https://doi.org/10.1111/j.0006-341x.2000.00337.x.Search in Google Scholar
27. Pencina, MJ, D’Agostino, RB. Overall C as a measure of discrimination in survival analysis: model specific population value and confidence interval estimation. Stat Med 2004;23:2109–23. https://doi.org/10.1002/sim.1802.Search in Google Scholar
Supplementary Material
This article contains supplementary material (https://doi.org/10.1515/ijb-2025-0049).
© 2025 the author(s), published by De Gruyter, Berlin/Boston
This work is licensed under the Creative Commons Attribution 4.0 International License.