Abstract
Generalized additive models (GAMs) based on the binomial and Poisson distributions can be used to provide flexible semi-parametric modelling of binary and count outcomes. When used with the canonical link function, these GAMs provide semi-parametrically adjusted odds ratios and rate ratios. For adjustment of other effect measures, including rate differences, risk differences and relative risks, non-canonical link functions must be used together with a constrained parameter space. However, the algorithms used to fit these models typically rely on a form of the iteratively reweighted least squares algorithm, which can be numerically unstable when a constrained non-canonical model is used. We describe an application of a combinatorial EM algorithm to fit identity link Poisson, identity link binomial and log link binomial GAMs in order to estimate semi-parametrically adjusted rate differences, risk differences and relative risks. Using smooth regression functions based on B-splines, the method provides stable convergence to the maximum likelihood estimates, and it ensures that the estimates always remain within the parameter space. It is also straightforward to apply a monotonicity constraint to the smooth regression functions. We illustrate the method using data from a clinical trial in heart attack patients.
1 Introduction
Binary and count response data are commonly encountered in biostatistical settings, and generalized linear models (GLMs) are used to estimate the adjusted effects of several covariates on the risk or rate of these outcomes. The link function in the GLM determines the scale on which the effect measure is expressed. Using the canonical link with a binomial model for binary data gives adjusted odds ratios, and with a Poisson model for count data gives adjusted rate ratios [1].
In biostatistics, other effect measures such as adjusted rate differences, risk differences or relative risks are often of interest, meaning that a non-canonical link function must be used. Under such models, constraints on the parameter space are required to ensure that fitted rates are non-negative and fitted risks lie within
Marschner [3], Donoghoe and Marschner [4] and Marschner and Gillett [5] have described stable methods for finding the maximum likelihood estimate (MLE) of Poisson GLMs with an identity link, and binomial GLMs with identity and log links. The methods allow estimation of adjusted rate differences, risk differences and relative risks respectively, avoiding convergence problems. All of these are applications of the combinatorial expectation–maximization (CEM) algorithm presented by Marschner [6].
With these models, however, the functional form of any continuous covariates must be specified. Generalized additive models (GAMs) are an extension of GLMs that allow for extra flexibility through the inclusion of semi-parametric terms [7,8]. This can potentially lead to a better model fit, or help to identify a more parsimonious model for the outcome. Model-fitting with GAMs often uses similar algorithms as those used for GLMs, and hence is subject to similar convergence issues, particularly with non-standard link functions. In fact, in some GAM packages such as PROC GAM in SAS [9], only the canonical link is permitted. In other packages, such as the R packages discussed in Section 6, non-canonical links are permitted but can be numerically unstable.
In this paper, we extend the existing CEM algorithms for GLMs to GAMs, by the addition of smooth semi-parametric functions based on B-splines [10,11]. We begin by defining the general GAM in Section 2, and in Section 3 we explore the properties of the B-spline basis functions. In Section 4 we explain the method for finding the MLE of each of these models, as well as how to apply an optional monotonicity constraint to the smooth functions. In Section 5, we demonstrate our methods by applying them to data from a clinical trial in heart attack patients. In Section 6, we summarize other popular methods for fitting GAMs and their performance in this dataset.
2 Model specification
Consider a sample of independent random variables
so that
In a GAM, the standardized mean
Here,
In order to impose a finite dimensional structure on the problem, the MLE
a restriction we denote by
One possible choice of basis functions is a sequence of indicator functions that serve as the increments between successive unique observed values of the covariate. That is, if
The resulting
Under the step function model, a large number of degrees of freedom are sacrificed in order to ensure that the semi-parametric function estimate fits closely to the observed data. Here we will instead focus on a smooth semi-parametric regression technique, in which the number of parameters that need to be estimated is reduced and
This could be achieved with any of a wide range of basis functions [8]. We will use the polynomial B-splines, as they are highly flexible and their properties allow us to easily integrate them into the existing CEM algorithms, as discussed in the next section.
3 B-splines
3.1 Definition and properties
The B-splines are a series of basis functions for polynomial splines, defined by a grid of
We will restrict our focus to the case of B-splines of order 3, where each basis function is made up of a series of quadratic curves between each pair of turning points, with adjacent curves constrained to have equal gradient at their boundaries. The sequence of
Given a knot sequence, the
The B-splines are normalized such that
for all w, meaning that a constraint must be applied to the parameters for each c to ensure that they are identifiable. We do this by setting
Each basis function
As an illustration we present a graphical representation in Figure 1. The four B-spline basis functions of order 3 with a single internal turning point, and hence 7 knots, are shown in Figure 1(A). Figure 1(C) shows the curve
which can be equivalently expressed as
by using eq. (3). Figure 1(B) and (D) illustrates the use of B-splines to restrict

(A) B-spline basis functions and (C) example resulting curve with one internal knot, and (B) the equivalent monotonic B-spline basis functions and (D) resulting curve based on the same coefficients.
3.2 Model constraints
The models considered in this paper have constraints on their parameter spaces due to the restricted range of the response variable: rates must be non-negative and risks must lie in
The CEM algorithm is ideal for fitting these models, as it applies the parameter space constraints while guaranteeing stable convergence to the MLE. A crucial step in the definition of a CEM algorithm is that the parameter space is partitioned into a sequence of subspaces, each of which corresponds to a particular set of constraints on individual parameters. The properties of the B-splines as discussed in Section 3.1 allow us to extend the existing methods to include these semi-parametric functions.
As an example, we restrict our attention to
The function space
it is easy to see that
For a particular identifiability constraint
One characterization of the constrained function space is that any curve
An analogous result can be achieved by fixing
A useful property of the existing CEM algorithms is that they can easily accommodate non-negativity or non-positivity constraints on individual coefficients. By applying such constraints, we can find the MLE of the GAM under the restriction that each
3.3 Monotonic B-splines
In some contexts, it may be sensible to restrict
Assuming non-negative coefficients, the identifiability constraint
for
This can be achieved in the same manner as the unrestricted case by re-expressing eq. (2) as
Here,
Continuing with the illustration provided in Section 3.1, the monotonic B-spline basis functions associated with the B-spline bases in Figure 1(A) are shown in Figure 1(B). Because the B-spline coefficients in eq. (4) are monotonically non-decreasing, the resulting curve in Figure 1(C) can be expressed in terms of the monotonic B-splines
and this is demonstrated in Figure 1(D), where the intercept term is shown as a horizontal line.
For the alternate case in which we wish for the curve
and their coefficients
3.4 Knot selection
The number and placement of the turning points will influence the shape of the resulting function by determining the space in which our estimated
In some situations, such as OLS, the positioning of the turning points is crucially important [15], and a large number of knot selection methods have been derived (e.g. [10], pp. 174–196, [16]). Many of these methods can be integrated into the approach we describe, but they often lead to a large increase in the computational burden. In a general setting, Ramsay [12] noted that the shape of the resulting estimate is not particularly sensitive to knot placement.
We apply an adaptation of cardinal splines discussed by Hastie and Tibshirani ([7], p. 24), placing the
Of more importance is the choice of the number of turning points. With too few, we may fail to detect important features of the relationship, but with too many, we are at risk of over-fitting. One common way to resolve this trade-off is to use a sufficiently large number of turning points to broaden the function space for
However, with a penalty term, the CEM algorithm central to these stable methods cannot be used directly. We discuss this further, and propose a possible solution in Section 7. In general, we will use the Akaike information criterion
As a criterion for choosing the optimal smoothing parameter in a wide range of scenarios, the AIC and
Figure 2 shows the results from a simulation study in which we examined the performance of the AIC value in selecting the optimal number of knots. We simulated 500 datasets of 750 binomial observations, each with true risk functions that were based on B-splines with 1, 2 or 3 internal knots. For each dataset, we found the MLE corresponding to a binomial model in which the continuous variable was included as a linear term or as a B-spline with between 0 and 5 internal knots. Of these, the model with the smallest AIC was selected as the optimal model. This approach selected the correct number of knots in the vast majority of samples, and the mean number of knots selected was close to the true value. As has been observed in other contexts, the BIC was biased towards a low number of knots (oversmoothing), and the GCV criterion tended to undersmooth by choosing a higher number of knots more often (data not shown).

Results from 500 simulations of the performance of the AIC for selecting the optimal number of knots. The panels on the left-hand side show the true risk function
The computational effort required by full cross-validation renders it infeasible in this situation. Nevertheless, the focus of this paper is on providing a method of estimation for a single model, which may be applied within any scheme for determining the optimal number of knots.
4 Method
4.1 CEM algorithm
Each of the methods that we use to fit the models in the subsequent sections is an application of CEM algorithms [6]. A CEM algorithm is a general approach in which we consider a finite family of complete data models, indexed by
The complete data models are defined such that an expectation–maximization (EM) algorithm [24] can be used to find the constrained maximum of the likelihood
When the model of interest is a GLM, the complete data models typically impose some constraint on the individual model parameters, such as non-negativity. In the case of GAMs as described below, we use augmented complete data models, where the effect of such constraints on the spline coefficients is to constrain the shape of the smooth curves
4.2 Adjusted rate differences
Adjusted rate differences can be estimated by fitting an identity link Poisson GAM. Specifically, if
With link function
The Poisson means must be non-negative, and so the parameter space within which the MLE
where
For
For a particular choice of
By considering all possible reference vectors
For
The complete data model is augmented with
If the basis function values
As discussed in Section 3.2, the non-negativity constraints on the
where
Let
We have thus defined a CEM algorithm for finding the overall MLE
This approach can be implemented directly using the existing CEM algorithm for identity link Poisson GLMs. Using this algorithm, it is straightforward to apply non-negativity constraints to the coefficients of some continuous covariates by considering only one of the two possible reference levels for that covariate.
For a particular choice of
Repeating this for all
The process can be halted if one such application converges to a point in the interior of the constrained parameter space, as we can be sure that this is the overall MLE.
The process of cycling through the parameter space partition is illustrated in Figure 3 for a Poisson model with
where the usual dependence on c has been suppressed since
which has

Illustration of the effect of parameter constraints on the smooth curve in a simulated Poisson dataset. The red dotted line represents the true underlying rate and the grey dotted lines denote the knot locations. The black solid line in each panel is the MLE under the identifiability constraint shown in the heading, with all other parameters constrained to be non-negative. The MLE is obtained with the constraint
One point to note is that the non-negativity of the coefficients
4.3 Adjusted relative risks
With binomial data we have
With
The fitted risks must lie within
where
With
That is,
The union of these constrained parameter spaces over all possible choices of
Extension to include smooth terms follows an analogous approach to that used in Section 4.2. For a particular
By considering each possible
Again, the sequence may stop early if a maximum in the interior of the parameter space is identified. Furthermore, the same caution also applies here as it did for the identity link Poisson model: non-positivity of the coefficients is only a sufficient condition for non-positivity of the linear predictor, so we are not searching the entire parameter space where
4.4 Adjusted risk differences
Adjusted risk differences can be estimated using a binomial GAM with identity link, that is,
Again we require that the probabilities
With
However, the inclusion of the semi-parametric terms cannot proceed in exactly the same way as in Sections 4.2 and 4.3. The covariate space considered for continuous covariates entered into the identity link binomial CEM algorithm is the Cartesian product of the observed ranges of those covariates, that is,
Instead we must use a slightly different approach, based on the ordering of the B-spline coefficients. We begin by choosing
Recall that if we define new basis functions
and non-negativity constraints on the associated coefficients will impose an order restriction on the original coefficients, that is,
We denote by
There are
As with the other models, we may stop early if we find a constrained MLE in the interior of the parameter space, and there are various strategies to identify the parameterizations that are more likely to contain this MLE in order to potentially reduce computing time [6].
4.5 Monotonic smooth regression
Imposing a monotonicity restriction on one or more of the smooth curves is straightforward for all of the models examined so far. As discussed in Section 3.3, a sufficient condition for
For the identity link binomial model in Section 4.4, applying this constraint is trivial. In iterating through the possible choices of
For the other models, we employ the methods discussed in Section 3.3. In an identity link Poisson model, we replace the B-spline function values
For the log link binomial model we proceed similarly, replacing
It is important to note that the monotonicity of the coefficients is only a sufficient condition for the monotonicity of the resulting smooth function. Hence the function space over which we search for the MLE
5 Application
The ASSENT-2 study [26] was a randomized clinical trial designed to assess the safety and efficacy of tenecteplase versus alteplase in 16,949 patients treated within 6 hours of an acute myocardial infarction (MI). The primary outcome was 30-day mortality after randomization, and the primary analysis of this outcome showed that the two treatments were equivalent.
To demonstrate our method, we undertake a risk factor analysis, in which the risk of death is modelled in terms of a semi-parametric age effect and three categorical covariates: MI severity (Killip class I, II or III/IV), treatment delay (

AIC of the fitted binomial model in the ASSENT-2 study with identity (black, solid line) and log (red, dotted line) links, for a varying number of parameters associated with the age covariate.
Figure 4 shows that the inclusion of a linear age term (1 parameter) improves the AIC considerably for both the log link and identity link models. In the risk difference model, allowing the age term to be quadratic (2 parameters) and semi-parametric with one internal knot (3 parameters) further reduces the AIC substantially, but beyond this, the penalty of additional parameters overrides the small improvements in fit. In the relative risk model, additional flexibility in the age term does not give vastly superior AIC values when compared to the linear age model, although the best in terms of AIC is a model with two internal knots, and hence 4 parameters for the semi-parametric age term. The fitted risks by age for all 27 groups (

Fitted risk of 30-day mortality by age in the ASSENT-2 study for different combinations of MI severity (thin: I, medium: II, thick: III/IV), treatment delay (solid:
Although a monotonic dependence on age is natural for mortality, in this particular analysis it has little effect on the fitted model. Figure 6 displays the fitted age-specific regression functions for both models, showing that the fitted function is virtually identical for the identity link model with and without monotonicity, and is identical for the log link model.

(A) Adjusted risk difference and (B) adjusted relative risk associated with age (versus 40 years), with pointwise 95% confidence intervals, estimated using the information matrix (shaded) and bootstrap resampling (dashed lines). The dotted red line in panel (A) shows the estimated risk difference when a monotonicity constraint is applied.
Also shown in Figure 6 is a comparison of two approaches for confidence interval estimation. The first approach, shown by the shaded regions, uses asymptotic normality and the information matrix evaluated at the MLE. The information matrix is obtained as the expected second derivative matrix corresponding to the binomial log-likelihood function with probabilities specified by eqs (1) and (2), with either the identity or log link function. The second approach, shown by the dashed lines, uses bootstrapping with pointwise confidence intervals determined using the percentile method. We used 1,000 bootstrap resamples with replacement, and due to the stability of our fitting methods, convergence to the MLE was achieved in every resample. It can be seen that the two methods of confidence interval estimation show very close agreement, which provides some level of support for their use in this analysis.
Adjusted rate differences can also be estimated for this data by using our method for an identity link Poisson model, with semi-parametric adjustment for age. Because in this case all of the patients were observed for the same period of time, the parameter estimates are very similar to those from the identity link binomial model; however, they have a different interpretation – they are the absolute change in the rate of death per patient-month. A comparison of the parameter estimates from the two models is shown in Table 1.
Parameter estimates from identity link binomial (risk difference) and identity link Poisson (rate difference) models on the ASSENT-2 data.
Covariate | Risk difference | Rate difference |
Severity (vs. I) | ||
II | −0.0607 | −0.0605 |
III/IV | −0.2676 | −0.2692 |
Treatment delay (vs. <2 hours) | ||
2–4 hours | −0.0024 | −0.0025 |
>4 hours | −0.0014 | −0.0010 |
Region (vs. Western) | ||
Latin America | −0.0047 | −0.0050 |
Eastern Europe | −0.0373 | −0.0359 |
Age (vs. 40 years) | ||
50 years | −0.0045 | −0.0045 |
60 years | −0.0179 | −0.0178 |
70 years | −0.0553 | −0.0553 |
80 years | −0.1534 | −0.1541 |
6 Other methods
The methods described in this paper are fully implemented in R packages addreg [27] and logbin [28], available from the Comprehensive R Archive Network (CRAN). There are three other notable R packages that provide methods for fitting GAMs: gam, gamlss and mgcv. Each of these allows the models with non-canonical link functions discussed in this paper; however, all employ iterative algorithms involving variants of Fisher scoring or Newton–Raphson, making them subject to instability.
The gam function in the gam package in R [29] fits GAMs using cubic smoothing splines by employing a local scoring algorithm. This consists of a backfitting (Gauss–Seidel) algorithm for fitting the non-parametric parts of the model within a Newton–Raphson step for updating the parametric parts [30]. The inner loop can be shown to always converge, but the outer loop is only guaranteed to do so if some form of step-size optimization is performed ([7], p. 151). However, the implementation in R does not include any option for step-size modification, and additionally there is no check for the validity of the fitted means. This means that convergence is not guaranteed, and when the method does converge, it may be to a value outside the parameter space. For the 1,000 bootstrap samples used to produce the confidence intervals in Figure 6, the algorithm failed to converge to a valid solution in nearly half (49.1%) of the samples for the identity link model. For the log link model, the algorithm converged in all 1,000 samples, but some of the fitted risks exceeded 1 in every case.
The gamlss package [31] provides a method for non-parametric modelling of various parameters of the distribution, including the mean. Similarly to gam, its fitting algorithm uses backfitting iterations for the non-parametric parts within Newton–Raphson steps that update the parameter estimates. Unlike gam, the user is able to specify the step length for updating parameter estimates, but the function terminates with an error if the update produces invalid fitted values. For the relative risk model fitted on the bootstrap samples from Section 5, gamlss converged in only 57 samples when the default step size was used. Convergence in all 1,000 samples was achieved when the step size was made sufficiently small. For the risk difference model, however, gamlss did not converge in any of the bootstrap samples regardless of the step size chosen, due to the error caused by violation of the parameter space constraint.
The mgcv package [32] provides a flexible gam function, allowing for a wide variety of smoothers, as well as methods for automatic selection of the level of smoothing. The fitting method uses a penalized iteratively reweighted least squares algorithm [32], which is not guaranteed to converge, but step-halving is invoked if the penalized deviance increases markedly between iterations, or the estimates move outside the parameter space. Additionally, the user can specify a ridge regression penalty to assist with convergence issues caused by unidentifiable estimates. In order to directly compare the performance of mgcv’s gam function to our method, we fitted identical unpenalized B-spline models to the ASSENT-2 bootstrap data. We found that gam achieved stable convergence whenever the MLE was in the interior of the parameter space, but convergence problems were possible when the MLE was on the parameter space boundary, particularly for the identity link model. The nature of these convergence problems was dependent on the version of mgcv that was used. In particular, when using version 1.7 we found that convergence could occur to a sub-optimal boundary point, while in version 1.8 we found that the algorithm could fail to declare convergence when the estimates reached the MLE. This behaviour occurred in a large proportion of our bootstrap replications and persisted even when a stricter convergence criterion and a greater number of iterations were used.
None of these methods support monotonicity constraints on the smooth curves. The GMBBoost [13] and GMonBoost [14] methods employ likelihood-based boosting techniques to fit GAMs with monotonicity constraints, but in the current implementation only canonical link models are allowed, so we cannot compare them with our approach.
Overall this discussion illustrates that although there are other approaches that could potentially be used for semi-parametric modelling of rate differences, risk differences and relative risks, numerical instability is often an issue. Furthermore, monotonicity constraints for non-canonical models are not available in existing software. The stability and flexibility of our method therefore means that it is a useful addition to existing GAM methodology.
7 Discussion
We have presented a method for smooth semi-parametric adjustment of rate differences, relative risks and risk differences. In general, this can be achieved by using GAMs; however, these effect measures require non-standard link functions and the usual fitting algorithms can fail to converge to the MLE. Our method avoids this by employing variants of existing stable CEM algorithms for fully parametric versions of these models, using B-spline basis functions for the smooth components.
The method is itself a CEM algorithm, and relies on the fact that the EM algorithm will always converge to the MLE within a constrained parameter space. Each constrained parameter space is defined by placing a restriction on the shape of the smooth curve, and by applying the algorithm for each constrained parameter space we are guaranteed to find the overall MLE.
We applied our method to data from the ASSENT-2 clinical trial, showing that semi-parametric adjustment for age provided a better fit than entering age as a linear term in both risk difference and relative risk models. Furthermore, the stability of our fitting algorithm allowed us to use bootstrap resampling to estimate confidence intervals for the semi-parametric relationship. Adjusted rate differences can also be estimated using an identity link Poisson model.
The calculations required at each iteration of the EM algorithms for each method presented here are very simple, although the EM algorithm may take a large number of iterations to converge. The overall computational time required by these methods depends on the number of parameters that must be estimated in a particular model. The models presented in Figure 5 required approximately 3 and 2 minutes, respectively, to find the MLE on a 3.4 GHz processor. One potential method for reducing the computational time is to exploit the fact that the EM algorithms for each parameterization are independent, and could be conducted in parallel on a multi-core processor. Other techniques for speeding up convergence of CEM algorithms have been discussed by Marschner [6].
Further adaptations of our approach are possible. For example, Marschner, Gillett and O’Connell [33] have presented an extension of the CEM algorithm for the identity link Poisson model, in which additional categorical stratification factors have a multiplicative effect on the Poisson means. The algorithm is similar to that for the additive model, but each constrained MLE is found by using an expectation–conditional maximization (ECM) algorithm. It is straightforward to impose non-negativity constraints on the additive parameters, so we can extend this method to allow smooth semi-parametric components by using the same approach described in Section 4.2. An alternative approach to incorporating both additive and multiplicative effects in the same model was provided by the LEXPIT model of Kovalchik, Varadhan, Fetterman, Poitras, Wacholder and Katki [34]. We anticipate that our approach may be useful in extending this model semi-parametrically, although we have not yet investigated this.
Most applications of GAMs use penalized likelihood to allow for flexibility while lessening the tendency to overfit. However, any reasonable penalty term will cause the M-step of the EM algorithm to lose its parameter separation and become a multi-dimensional maximization problem. Marschner and Gillett [5] proposed a solution to this for log binomial models by employing the one-step-late algorithm of Green [35]. Here, the M-step is modified such that parameters associated with the penalty term are replaced by their current estimates. However, this is no longer an EM algorithm, and does not guarantee that the parameter estimates will remain in the parameter space, or even that the likelihood will increase at each step. This can be remedied by a process similar to step-halving, whereby if the new estimate has lower likelihood or is outside the parameter space, we replace it with an estimate between the unpenalized and penalized updates, moving closer to the unpenalized estimate until the conditions are met. Nonetheless, while a penalized likelihood version of our method would be possible, the simple spline-based model used here is likely to provide sufficient flexibility in practice.
Rate differences, relative risks and risk differences are useful in biostatistical settings to provide effect size measures in randomized trials and epidemiological studies. However, the GLMs and GAMs used to estimate these effects are also used in other areas. For example, the identity link Poisson model has been recently used in an ecological study [36], the log link binomial model in a study of socioeconomic status [37] and the identity link binomial model in a study of international politics [38]. This suggests that our method may also have wide applicability outside biostatistics.
Funding statement: Funding: This research was supported by the Australian Research Council DP110101254.
Acknowledgements
The authors thank the ASSENT-2 investigators for providing data to the National Health and Medical Research Council Clinical Trials Centre, and the anonymous reviewers for their constructive comments to help improve this paper.
References
1. McCullaghP, NelderJA. Generalized linear models, 2nd ed. London: Chapman & Hall, 1989.10.1007/978-1-4899-3242-6_2Suche in Google Scholar
2. MarschnerIC. glm2: fitting generalized linear models with convergence problems . R J2011;3:12–15.10.32614/RJ-2011-012Suche in Google Scholar
3. MarschnerIC. Stable computation of maximum likelihood estimates in identity link Poisson regression . J Comput Graph Stat2010;19:666–83.10.1198/jcgs.2010.09127Suche in Google Scholar
4. DonoghoeMW, MarschnerIC. Stable computational methods for additive binomial models with application to adjusted risk differences . Comput Stat Data Anal2014;80:184–96.10.1016/j.csda.2014.06.019Suche in Google Scholar
5. MarschnerIC, GillettAC. Relative risk regression: reliable and flexible methods for log-binomial models . Biostatistics2012;13:179–92.10.1093/biostatistics/kxr030Suche in Google Scholar PubMed
6. MarschnerIC. Combinatorial EM algorithms . Stat Comput2014;24:921–40.10.1007/s11222-013-9411-7Suche in Google Scholar
7. HastieT, TibshiraniR. Generalized additive models, 1st ed. London: Chapman & Hall, 1990.Suche in Google Scholar
8. WoodSN. Generalized additive models: an introduction with R. Boca Raton, FL: Chapman & Hall, 2006.Suche in Google Scholar
9. SAS Institute Inc. SAS/STAT Software, Version 9.4, Cary, NC, 2013. Available at: http://www.sas.com/Suche in Google Scholar
10. de BoorC. A practical guide to splines (Applied mathematical sciences). New York: Springer-Verlag, 1978.10.1007/978-1-4612-6333-3Suche in Google Scholar
11. EilersPHC, MarxBD. Flexible smoothing with B-splines and penalties . Stat Sci1996;11:89–121.10.1214/ss/1038425655Suche in Google Scholar
12. RamsayJO. Monotone regression splines in action . Stat Sci1988;3:425–41.10.1214/ss/1177012761Suche in Google Scholar
13. LeitenstorferF, TutzG. Generalized monotonic regression based on B-splines with an application to air pollution data . Biostatistics2007;8:654–73.10.1093/biostatistics/kxl036Suche in Google Scholar PubMed
14. TutzG, LeitenstorferF. Generalized smooth monotonic regression in additive modeling . J Comput Graph Stat2007;16:165–88.10.1198/106186007X180949Suche in Google Scholar
15. RuppertD, WandMP, CarrollRJ. Semiparametric regression (Cambridge series in statistical and probabilistic mathematics). New York: Cambridge University Press, 2003:131.Suche in Google Scholar
16. FriedmanJH, SilvermanBW. Flexible parsimonious smoothing and additive modeling (with discussion) . Technometrics1989;31:3–39.10.2307/1270362Suche in Google Scholar
17. RuppertD. Selecting the number of knots for penalized splines . J Comput Graph Stat2002;11:735–57.10.1198/106186002853Suche in Google Scholar
18. YaoF, LeeTCM. On knot placement for penalized spline regression . J Korean Stat Soc2008;37:259–67.10.1016/j.jkss.2008.01.003Suche in Google Scholar
19. GreenP, SilvermanB. Nonparametric regression and generalized linear models: a roughness penalty approach (Monographs on statistics and applied probability), 1st ed. London: Chapman & Hall, 1994.10.1007/978-1-4899-4473-3Suche in Google Scholar
20. AkaikeH. A new look at the statistical model identification . IEEE Trans Autom Control1974;19:716–23.10.1109/TAC.1974.1100705Suche in Google Scholar
21. BurnhamKP, AndersonDR. Model selection and multimodel inference: a practical information-theoretic approach, 2nd ed. New York: Springer-Verlag, 2002.Suche in Google Scholar
22. HurvichCM, SimonoffJS, TsaiC-L. Smoothing parameter selection in nonparametric regression using an improved Akaike information criterion . J R Stat Soc Ser B1998;60:271–93.10.1111/1467-9868.00125Suche in Google Scholar
23. MalloyEJ, SpiegelmanD, EisenEA. Comparing measures of model selection for penalized splines in Cox models . Comput Stat Data Anal2009;53:2605–16.10.1016/j.csda.2008.12.008Suche in Google Scholar
24. DempsterAP, LairdNM, RubinDB. Maximum likelihood from incomplete data via the EM algorithm . J R Stat Soc Ser B1977;39:1–38.Suche in Google Scholar
25. BakerSG. The multinomial–Poisson transformation . Statistician1994;43:495–504.10.2307/2348134Suche in Google Scholar
26. ASSENT-2 Investigators. Single-bolus tenecteplase compared with front-loaded alteplase in acute myocardial infarction: the ASSENT-2 double-blind randomised trial . Lancet1999;354:716–22.10.1016/S0140-6736(99)07403-6Suche in Google Scholar
27. DonoghoeM. addreg: additive regression for discrete data. 2014a. Available at: http://CRAN.R-project.org/package=addreg, R package version 1.2Suche in Google Scholar
28. DonoghoeM. logbin: relative risk regression using the log binomial model. 2014b. Available at: http://CRAN.R-project.org/package=logbin, R package version 1.0Suche in Google Scholar
29. HastieTJ. gam: generalized additive models. 2013. Available at: http://CRAN.R-project.org/package=gam, R package version 1.09Suche in Google Scholar
30. HastieTJ. Generalized additive models (Chap. 7). In: ChambersJM and HastieTJ, editors. Statistical models in S. Pacific Grove, CA: Wadsworth & Brooks/Cole, 1992;249–307.Suche in Google Scholar
31. RigbyRA, StasinopoulosDM. Generalized additive models for location, scale and shape (with discussion) . Appl Stat2005;54:507–54.10.1111/j.1467-9876.2005.00510.xSuche in Google Scholar
32. WoodSN. Fast stable restricted maximum likelihood and marginal likelihood estimation of semiparametric generalized linear models . J R Stat Soc Ser B2011;73:3–36.10.1111/j.1467-9868.2010.00749.xSuche in Google Scholar
33. MarschnerIC, GillettAC, O‘ConnellRL. Stratified additive Poisson models: computational methods and applications in clinical epidemiology . Comput Stat Data Anal2012;56:1115–30.10.1016/j.csda.2011.08.002Suche in Google Scholar
34. KovalchikSA, VaradhanR, FettermanB, PoitrasNE, WacholderS, KatkiHA. A general binomial regression model to estimate standardized risk differences from binary response data . Stat Med2013;32:808–21.10.1002/sim.5553Suche in Google Scholar PubMed PubMed Central
35. GreenPJ. On use of the EM algorithm for penalized likelihood estimation . J R Stat Soc Ser B1990;52:443–52.10.1111/j.2517-6161.1990.tb01798.xSuche in Google Scholar
36. StjernmanM, GreenM, LindströmÅ, OlssonO, OttvallR, SmithHG. Habitat-specific bird trends and their effect on the farmland bird index . Ecol Indic2013;24:382–91.10.1016/j.ecolind.2012.07.016Suche in Google Scholar
37. HiyoshiA, FukudaY, ShipleyMJ, BartleyM, BrunnerEJ. A new theory-based social classification in Japan and its validation using historically collected information . Soc Sci Med2013;87:84–92.10.1016/j.socscimed.2013.03.021Suche in Google Scholar PubMed
38. BergerD, CorvalanA, EasterlyW, SatyanathS. Do superpower interventions have short and long term consequences for democracy? J Comp Econ2013;41:22–34.10.1016/j.jce.2013.01.004Suche in Google Scholar
© 2015 Walter de Gruyter GmbH, Berlin/Boston
Artikel in diesem Heft
- Frontmatter
- Research Articles
- Within-Subject Mediation Analysis in AB/BA Crossover Designs
- Conditional Transformation Models for Survivor Function Estimation
- A Universal Approximate Cross-Validation Criterion for Regular Risk Functions
- Double Bias: Estimation of Causal Effects from Length-Biased Samples in the Presence of Confounding
- Flexible Regression Models for Rate Differences, Risk Differences and Relative Risks
- Nearest-Neighbor Estimation for ROC Analysis under Verification Bias
- Quantifying an Agreement Study
- Robust Bayesian Sensitivity Analysis for Case–Control Studies with Uncertain Exposure Misclassification Probabilities
- A Semi-stationary Copula Model Approach for Bivariate Survival Data with Interval Sampling
- Comparison of Splitting Methods on Survival Tree
Artikel in diesem Heft
- Frontmatter
- Research Articles
- Within-Subject Mediation Analysis in AB/BA Crossover Designs
- Conditional Transformation Models for Survivor Function Estimation
- A Universal Approximate Cross-Validation Criterion for Regular Risk Functions
- Double Bias: Estimation of Causal Effects from Length-Biased Samples in the Presence of Confounding
- Flexible Regression Models for Rate Differences, Risk Differences and Relative Risks
- Nearest-Neighbor Estimation for ROC Analysis under Verification Bias
- Quantifying an Agreement Study
- Robust Bayesian Sensitivity Analysis for Case–Control Studies with Uncertain Exposure Misclassification Probabilities
- A Semi-stationary Copula Model Approach for Bivariate Survival Data with Interval Sampling
- Comparison of Splitting Methods on Survival Tree