Home Robust estimation for partial functional linear regression models based on FPCA and weighted composite quantile regression
Article Open Access

Robust estimation for partial functional linear regression models based on FPCA and weighted composite quantile regression

  • Peng Cao EMAIL logo and Jun Sun
Published/Copyright: December 31, 2021

Abstract

In this paper, we consider a novel estimation for partial functional linear regression models. The functional principal component analysis method is employed to estimate the slope function and the functional predictive variable, respectively. An efficient estimation based on principal component basis function approximation is used for minimizing the proposed weighted composite quantile regression (WCQR) objective function. Since the proposed WCQR involves a vector of weights, we develop a computational strategy for data-driven selection of the optimal weights. Under some mild conditions, the theoretical properties of the proposed WCQR method are obtained. The simulation study and a real data analysis are provided to illustrate the numerical performance of the resulting estimators.

MSC 2010: 62J05; 62M10

1 Introduction

In the literature of functional data analysis, functional linear regression models (FLRMs) provide an efficient method to analyze the relationship between a functional predictor and a scalar response. Therefore, much effort has been devoted to studying its estimation and other relevant inference problems, see [1,2, 3,4,5, 6,7,8, 9,10,11, 12,13]. However, in practice, we often see that a scalar response is related not only to functional covariates but also to scalar covariates. Thus, Shin considered partial functional linear regression models (PFLRMs) to balance the flexibility of FLRMs and interpretation of classical linear regression models [14]. Generally, PFLRM has the following form:

(1) Y = Z T θ + T X ( t ) β ( t ) d t + ε ,

where Y is a real-valued response variable defined on a probability space ( Ω , , P ) , Z is a p -dimensional random vector and θ = ( θ 1 , , θ p ) T is an unknown parameter vector, { X ( t ) ; t T } is a zero mean, second-order stochastic process defined on ( Ω , , P ) with sample paths in H = L 2 ( T ) , the Hilbert space containing square integrable functions on T with inner product x , y = T x ( t ) y ( t ) d t , x , y H and norm x = x , x 1 / 2 , β ( t ) is an unknown slope function belonging to H , ε is the random error independent of Z and X , E ( ε Z , X ) = 0 . T denotes the transport operation throughout this paper. Without loss of generality, we further assume T = [ 0 , 1 ] .

The PFLRMs have been studied by many authors. For example, Shin [14] proposed the estimation method based on functional principal component analysis (FPCA) and investigated the asymptotic properties of the estimators for model (1). Zhou et al. [15] considered the polynomial spline method to estimate model (1) and established the asymptotic normality for the parameter vector and the global convergence rate for the slope function. Furthermore, Kong et al. [16] studied variable selection in model (1) with multiple functional and ultrahigh-dimensional scalar predictors. Ling et al. [17] developed the k-nearest-neighbors estimation for PFLRMs. Kong et al. [18] proposed the partial functional linear cox regression model for censored outcomes. Cao et al. [19] developed a two-step estimation procedure via maximizing the quasi likelihood function in the framework of generalized PFLRMs. Yuan and Zhang [20] investigated the hypothesis test of the parametric components in PFLRMs based on B-spline. Zhu et al. [21] considered estimation and testing problems for PFLRMs when the covariates of the non-functional linear component are measured with additive error. Li et al. [22] proposed two test statistics for series correlation in PFLRMs and derived their asymptotic distributions under the null hypothesis.

Quantile regression (QR) is an effective approach for functional data analysis. This is due to the fact that QR models compared to the mean regression models [23] offer more robust merits especially the observations contain some outliers. For relevant literature, see [24,25,26]. Hence, QR method may result in an arbitrarily small relative efficiency compared with the mean regression. To conquer this drawback of QR method, Zou and Yuan proposed the composite quantile regression (CQR) method for linear models and showed that the relative efficiency of the CQR estimator is greater than 70% regardless of the error distribution compared with the least square (LS) estimator [27]. In many cases, CQR provides more efficient estimation compared with a single QR or the LS estimator for non-normal error distributions. This method has been deeply investigated in the literature and widely applied in many statistical models. For further references about the CQR method on nonparametric/semi-parametric models see [28,29,30, 31,32]. Even though CQR has been well developed, theory and methodology of PFLRMs are rare. Du et al. [33] used the penalized CQR method to study variable selection for parametric part in the PFLRMs, Yu et al. [34] considered composite quantile estimation for the PFLRMs with errors from short-range dependent and strictly stationary linear processes. However, the aforementioned CQR method for PFLRMs is a sum of different QRs with equal weights. Intuitively, using equal weights is not optimal in general, though CQR enjoys great advantages in terms of estimation efficiency. Therefore, Jiang et al. [35] addressed the issue of statistical inference for nonlinear models with large-dimensional covariates by using the weighted composite quantile regression (WCQR) method. Moreover, they pointed out that their method was more robust than some existing methods. Guo et al. [36] investigated group SCAD penalized WCQR estimation for varying coefficient models with diverging number of parameters. Jiang et al. [37] developed a data-driven WCQR for heteroscedastic partially linear varying coefficient models. Similar conclusions of WCQR method have been further confirmed in [6,38,39]. Until now, the WCQR method has not been used in functional regression analysis. Consequently, motivated by this fact, we devote to extending the WCQR method to PFLRMs in this paper. We show that the resulting estimators for both parametric components and slope function are more efficient in the case of asymmetrical error distribution, and as asymptotically efficient as the corresponding LS estimators when the error distribution is normal.

It is well known that many smoothing methods have been proposed for functional regressions, and two popular approaches for estimating the slope function β ( ) include the FPCA approach based on spectral decompositions of both the covariance of X ( t ) and its estimator (see [3,7,40]) and the basis function expansion method with both β ( ) and X ( t ) approximated by basis functions such as polynomial spline functions in [15]. In this paper, we propose a WCQR estimation method for PFLRMs through weighted composite quantile loss function with principal component basis function approximations. The main contributions of this paper are threefold:

  1. This is the first attempt to provide robust estimate via WCQR and establish the asymptotic properties of the proposed estimators for model (1).

  2. We develop a data-driven weighting strategy that maximizes the efficiency of the WCQR estimators. Practically, the proposed method is more efficient than the CQR method in [33] and rank regression method proposed under asymmetrical error distributed data set in [38].

  3. Theoretically, the proposed WCQR method generalizes the results of the composite quantile estimators in [8]. Particularly, our method reduces to that of Du et al. when the weights are all equal.

The rest of this paper is organized as follows. In Section 2, we describe the implementation details of the proposed algorithm and establish the theoretical properties of the estimators. In Section 3, we discuss the selection of weights and tuning parameters. Simulation study and a real data analysis are conducted in Sections 4 and 5 to evaluate the performance of our proposed approach, and then followed by a short conclusion in Section 6. All technical proofs are collected in the Appendix.

2 Method and main results

In partial functional linear QR model, for a given quantile level τ ( 0 , 1 ) , ( Z i , X i ( ) , Y i ) , i = 1 , , n is an independent and identically distributed (i.i.d.) sample from model (1), that is

(2) Q τ ( Y Z , X ( t ) ) = Z T θ τ + T X ( t ) β τ ( t ) d t + ε τ ,

where Q τ ( Y Z , X ( t ) ) is the τ th conditional quantile of Y given ( Z , X ( t ) ) , Y is a real-valued random variable defined on a probability space ( Ω , , P ) , Z is a p -dimensional vector of random variables with finite second moments, θ τ is a p × 1 coefficient vector to be estimated, β τ ( t ) is an unknown square integrable function on [ 0 , 1 ], and ε τ is a random error whose τ th quantile conditional on ( Z , X ( t ) ) being zero. For the convenience of presentation, we will omit τ from θ τ , β τ ( t ) and ε τ in model (2) wherever clear from the context, but we should keep in mind that those quantities are τ -specific.

Denote the covariance function of the process X ( ) and its empirical version, respectively, as

K ( s , t ) = Cov ( X ( s ) , X ( t ) ) , K ˆ ( s , t ) = 1 n i = 1 n X i ( s ) X i ( t ) .

The covariance function K defines a linear operator which maps a function f to K f given by ( K f ) ( u ) = K ( u , v ) f ( v ) d v . We assume that the linear operator with kernel K is positive definite. By Mercer’s theorem, then

K ( s , t ) = j = 1 λ j ϕ j ( s ) ϕ j ( t ) , K ˆ ( s , t ) = j = 1 λ ˆ j ϕ ˆ j ( s ) ϕ ˆ j ( t ) ,

where λ 1 > λ 2 > > 0 and λ ˆ 1 λ ˆ 2 λ ˆ n + 1 = = 0 are, respectively, the ordered eigenvalue sequences of the linear operators with kernels K and K ˆ , and { ϕ j } and { ϕ ˆ j } are the corresponding orthonormal eigenfunction sequences. Obviously, the vectors { ϕ j } and { ϕ ˆ j } each forms an orthonormal basis in H . According to the Karhunen-Loèvere presentation, then

(3) X ( t ) = i = 1 ξ i ϕ i ( t ) , β ( t ) = i = 1 γ i ϕ i ( t ) ,

where ξ i and γ i are defined by

ξ i = 0 1 X ( t ) ϕ i ( t ) d t , γ i = 0 1 β ( t ) ϕ i ( t ) d t ,

where ξ i is referred to as the i th functional principal component score of X ( t ) . It follows that ξ i are uncorrelated random variables with mean 0 and variance E ( ξ i 2 ) = λ i . For more details see [41]. Substituting (3) into model (2), the model (2) can be written as

(4) Y = Z T θ + j = 1 γ j ϕ j , X + ε .

Therefore, the regression model in (4) can be well approximated by

(5) Y Z T θ + j = 1 m γ j ϕ j , X + ε ,

where m n is the truncation level that trades off approximation error against variability and typically diverges with n . Replace ϕ j by ϕ ˆ j for j = 1 , , m , model (5) can be rewritten as

(6) Y Z T θ + U T γ + ε ,

where U = { X , ϕ ˆ j } j = 1 , , m , γ = ( γ 1 , , γ m ) T . Thus, the estimation problem of the parameters and the slope function are transformed into the estimation problem of the parameter vectors θ and γ .

Note that the CQR method used the same weight for different QR models. Intuitively, it will be more effective if different weights are used. According to Jiang et al. [35], the WCQR procedure estimates θ and γ via minimizing the following loss function:

(7) n ( θ , γ , c ; ω ) = k = 1 q ω k i = 1 n ρ τ k ( Y i c τ k Z i T θ U i T γ ) ,

where ρ τ k ( r ) = τ k r r I ( r < 0 ) , k = 1 , , q , q is the number of quantiles, τ k ( 0 , 1 ) . Typically, we use equally spaced quantile position: τ k = k / ( 1 + q ) , ω = ( ω 1 , , ω q ) T is a vector of weight satisfying ω k 0 . c τ k is 100 τ k % quantile of ε , c = ( c τ 1 , , c τ q ) T . Let the coefficients γ ^ = ( γ ^ 1 , , γ ^ m ) T and θ ^ = ( θ ^ 1 , , θ ^ p ) T be the solution of equation (6), then the resulting estimator of the slope function is given by

β ^ ( t ) = i = 1 m γ ^ i ϕ ^ i ( t ) .

3 Assumptions and asymptotic results

To establish the asymptotic properties of the proposed estimators, first, we introduce some notations. Let θ 0 and β 0 ( t ) be the true values of θ and β ( t ) , respectively, the notation is the L 2 norm for a function or the Euclidean norm for a vector. Let C denote a generic constant that might assume different values at different places, a n b n means that a n / b n is bounded away from 0 and infinity as n . The following technical assumptions are imposed.

  1. The random function X ( ) satisfies E X ( ) 4 < .

  2. For each j , E [ U j 4 ] C λ j 2 . For the eigenvalues λ j and Fourier coefficients γ j , we require that λ j λ j + 1 C 1 j a 1 and γ j C j b for j > 1 , a > 1 and b > a / 2 + 1 .

  3. The tuning parameter m satisfies m n 1 / ( a + 2 b ) .

  4. The random vector Z has bounded fourth moment, E Z 4 < .

  5. Z = η + g , X , where η = ( η 1 , , η p ) T is zero-mean random variable, g = ( g 1 , , g p ) T with g j L 2 ( [ 0 , 1 ] ) , j = 1 , , p .

  6. E [ η ] = 0 and E [ η η T ] = Σ . Furthermore, we need that Σ is a positive definite matrix.

  7. ε has the distribution function F ( ) with density f ( ) . The density function f ( ) of ε is positive and continuous at the τ k th quantiles c τ k .

Remark 1

Assumptions A1–A3 are required in the classical functional linear regression (see [5,7,24]). Specifically, Assumption A1 ensures the consistency of K ˆ ( s , t ) . Assumption A2 makes the slope function sufficiently smooth to be approximated [14]. Assumption A3 gives the order of the truncation parameter m to obtain the convergence rate of the slope function. Assumptions A4–A5 are quite parallel to [24,33]. Assumption A6 is used to establish the asymptotic normality of the parameters in model. Assumption A7 is very common in WCQR, see [35].

Theorem 3.1

Suppose that Assumptions A1–A7 are satisfied, then

  1. β ^ ( t ) β 0 ( t ) 2 = O p ( n ( 2 b 1 ) / ( a + 2 b ) ) ,

  2. n ( θ ^ θ 0 ) N ( 0 , σ 2 ( ω ) Σ 1 ) ,

where

σ 2 ( ω ) = k = 1 q ω k f ( c τ k ) 2 k , k = 1 q ω k ω k min ( τ k , τ k ) ( 1 max ( τ k , τ k ) ) .

Remark 2

The result of Theorem 3.1(i) indicates that the estimator β ^ ( t ) has the same rate of convergence as for the estimators of [7,20], which is optimal in minimax sense; (ii) indicates that if the tuning parameter m and weight ω are properly chosen, the WCQR estimator of parameter vector is n -consistent.

Remark 3

When q = 1 and τ 1 = τ , denote the τ th QR estimate of θ and β ( t ) by θ ^ QR and β ^ QR ( t ) , respectively. When all ω k are equal, denote the CQR estimate of θ and β ( t ) by θ ^ CQR and β ^ CQR ( t ) , respectively. Similarly, we denote the LS estimators of θ and β ( t ) by θ ^ LS and β ^ LS ( t ) , respectively. After some simple calculations, we can draw the following conclusions (1)–(3) and generalize the results of the LS estimators in [14], and the single-level quantile estimators in [24] and the composite quantile estimators in [33], respectively.

  1. Under the assumptions of Theorem 3.1, then

    1. β ^ LS ( t ) β 0 ( t ) 2 = O p ( n ( 2 b 1 ) / ( a + 2 b ) ) ,

    2. n ( θ ^ LS θ 0 ) N ( 0 , Σ LS ) ;

  2. Under the assumptions of Theorem 3.1, then

    1. β ^ QR ( t ) β 0 ( t ) 2 = O p ( n ( 2 b 1 ) / ( a + 2 b ) ) ,

    2. n ( θ ^ QR θ 0 ) N ( 0 , Σ QR ) ;

  3. Under the assumptions of Theorem 3.1, then

    1. β ^ CQR ( t ) β 0 ( t ) 2 = O p ( n ( 2 b 1 ) / ( a + 2 b ) ) ,

    2. n ( θ ^ CQR θ 0 ) N ( 0 , Σ CQR ) ,

    where Σ LS = σ 2 Σ 1 with σ 2 = E ( ε 2 ) , Σ QR = τ ( 1 τ ) f 2 ( c τ ) Σ 1 and Σ CQR = k , k = 1 q min ( τ k , τ k ) ( 1 max ( τ k , τ k ) ) k = 1 q f ( c τ k ) 2 Σ 1 .

4 Computational issues

From Theorem 3.1, we find that the asymptotic variance of θ ^ depends on ω only through σ 2 ( ω ) . Thus, the optimal choice of weights for maximizing efficiency of the estimator θ ^ is ω opt = arg min ω σ 2 ( ω ) . Based on [36], the optimal choice of weights is to solve the following quadratic optimization problem:

(8) ω opt = arg min ω ( ω T Ω ω ) subject to ω T f = 1 , ω 0 ,

where f = ( f ( c τ 1 ) , , f ( c τ q ) ) T and Ω is a q × q matrix with the ( k , k ) element Ω k , k = min ( τ k , τ k ) τ k τ k .

It can be seen from (8) that the optimal weight vector is rather complicated and involves the density of the errors c k = F 1 ( τ k ) and f ( c k ) , k = 1 , , q . In practice, the error density f ( ) is generally unknown. Following [36], we propose an estimation procedure as follows.

Step 1. We give the initial estimators θ ˜ and β ˜ ( t ) by minimizing the LS objective function and estimate σ 2 ( ) by

σ ˜ 2 = 1 n i = 1 n Y i Z i T θ ˜ 0 1 X i ( t ) β ˜ ( t ) d t 2 .

Step 2. Compute ε ˜ i = Y i Z i T θ ˜ 0 1 X i ( t ) β ˜ ( t ) d t σ ˜ and then we can use kernel density estimation to estimate f ( ) as follows:

f ˜ ( ) = 1 n h i = 1 n K ε ˜ i h .

Similar to Silverman [42], the bandwidth h is taken as

h = 0.9 × min std ( ε ˜ 1 , , ε ˜ n ) , IQR ( ε ˜ 1 , , ε ˜ n ) 1.34 × n 1 / 5 .

K ( ) is always chosen as the Gaussian kernel. Here std( ) and IQR( ) denote the sample standard deviation and sample interquartile, respectively.

Step 3. Estimate f { F 1 ( τ k ) } by f ˜ { F ˜ 1 ( τ k ) } , where F ˜ 1 ( τ k ) is the sample τ k quantile of { ε ˜ i , i = 1 , , n } .

The proposed estimators involve the tuning parameter m , which needs to be selected by minimizing some selection criteria such as cross validation (CV), generalized cross validation (GCV), AIC information criterion, BIC information criterion, etc. We use a BIC-type criterion to select m [36], that is

BIC ( m ) = log k = 1 q ω k i = 1 n ρ τ k ( Y i c ˆ τ k ( m ) Z i T θ ˆ ( m ) U i T γ ˆ ( m ) ) + log n 2 n ( m + p + q ) ,

where p is the number of components in parameter vector. c ˆ τ k ( m ) , θ ˆ ( m ) , and γ ˆ ( m ) are obtained by minimizing equation (7) with the first m principal components.

5 Simulation study

In this section, we conduct a Monte Carlo simulation example to assess the finite sample performance of the proposed procedures. Moreover, we include four competitors in our comparisons: (1) the LS method [14], (2) the QR method (QR) with τ = 0.5 [24], (3) the rank regression (RR) method [43], (4) the composite quantile regression (CQR) method [33]. Following [30], the performance of CQR estimation does not depend sensitively on the choice of the number of quantiles q . For the sake of simplicity, we only consider q = 9 for CQR and WCQR method in our simulation. The data sets are generated from the following PFLRM:

Y i = Z i 1 θ 1 + Z i 2 θ 2 + 0 1 X i ( t ) β ( t ) d t + ε i ,

where both Z i 1 and Z i 2 are standard normal with correlation coefficient 0.5, θ 1 = 0.8 and θ 2 = 2 . For the functional linear component, we take the same form as [40]. Specifically, the slope function β ( t ) = 2 sin ( π t / 2 ) + 3 2 sin ( 3 π t / 2 ) and X ( t ) = j = 1 100 ξ j v j ( t ) , where the ξ j are independently distributed as the normal with mean 0 and variances given by λ j = 10 ( ( j 0.5 ) π ) 2 , v j ( t ) = 2 sin ( ( j 0.5 ) π t ) .

In order to show the robustness of our estimators, the following six different error distributions are considered: standard normal distribution N(0, 1), t (3) distribution that is used to produce heavy-tailed distribution, F distribution F(4, 6), exponential distribution Exp(1), Log normal distribution LN(0, 1), and Weibull distribution Wb(1, 1.5) are used to produce asymmetric error distributions.

To assess the performance of the estimated parameters, the following criteria are considered:

(1) the average value of the corresponding coefficients (MEAN);

(2) the sample standard deviation (SD);

(3) the median of absolute bias (ABISE). To assess the performance of the slope function, we consider the following square root of average square errors (RASE):

RASE = 1 n grid k = 1 n grid ( β ˆ ( t k ) β ( t k ) ) 2 1 / 2 ,

where { t k , k = 1 , , n grid } are the grid points at which the slope function β ( ) is evaluated. In our simulation experiments, we set n grid = 200 , 500 repetitions are carried out with sample size n = 100 , 200, 400. In addition, the range of the selected m is from 1 to 8. The simulation results are presented in Tables 1, 2, 3, 4, 5, 6. From Tables 16, we have the following findings: (1) As the sample size n increases, the ABISE, SD, and RASE values of all estimators decrease. As expected, increasing the sample size leads to a better performance of all estimators. (2) The standard errors become smaller and the estimators of MEANs for the parametric components are very close to the true values θ as n increases, due to the fact of root- n consistency of the parameter estimators. This result confirms the asymptotic property that the parameter estimators are asymptotically unbiased as given in Theorem 3.1. (3) When the error follows N(0, 1), LS is the best one among the five estimators and WCQR performs nearly as well as CQR and RR. Although the QR method with quantile being 0.5 have similar performances, the WCQR seems to be slightly better than QR in most scenarios. (4) When the error distribution is symmetrical, CQR, RR, and WCQR perform similarly and satisfactorily. For all asymmetric errors: F(4, 6), Exp(1), LN(0, 1), and Wb(1, 1.5), WCQR performs better than others.

Table 1

Simulation results for θ with N(0, 1) random error

n Method θ 1 (MEAN) θ 1 (SD) θ 1 (ABISE) θ 2 (MEAN) θ 2 (SD) θ 2 (ABISE) β ( t ) (RASE)
100 LS 0.8010 0.1003 0.0778 1.9969 0.0989 0.0784 0.4293
QR 0.5162 0.4613 0.5243 0.4427 0.5346 0.5016 0.5162
RR 0.4331 0.4403 0.3859 0.3696 0.3931 0.4097 0.4331
CQR 0.4371 0.4408 0.3882 0.3717 0.3942 0.4072 0.4371
WCQR 0.4385 0.4445 0.3654 0.3492 0.3625 0.3721 0.4385
200 LS 0.2894 0.3419 0.3935 0.3033 0.3840 0.3712 0.2894
QR 0.3405 0.3302 0.3791 0.2966 0.3807 0.3419 0.3405
RR 0.2930 0.3141 0.2762 0.2504 0.2774 0.2909 0.2903
CQR 0.2940 0.3138 0.2764 0.2500 0.2778 0.2908 0.2940
WCQR 0.2941 0.3153 0.2587 0.2350 0.2602 0.2645 0.2941
400 LS 0.2007 0.2316 0.2794 0.2133 0.2828 0.2483 0.2007
QR 0.2273 0.2275 0.2737 0.2078 0.2701 0.2336 0.2273
RR 0.2025 0.2160 0.1987 0.1776 0.1899 0.1957 0.2025
CQR 0.2037 0.2162 0.1990 0.1783 0.1910 0.1959 0.2037
WCQR 0.2045 0.2160 0.1877 0.1676 0.1788 0.1800 0.2045
Table 2

Simulation results for θ with t (3) random error

n Method θ 1 (MEAN) θ 1 (SD) θ 1 (ABISE) θ 2 (MEAN) θ 2 (SD) θ 2 (ABISE) β ( t ) (RASE)
100 LS 0.7997 0.1735 0.1323 2.0071 0.1671 0.1311 0.4742
QR 0.7993 0.1409 0.1119 2.0033 0.1457 0.1147 0.4613
RR 0.7959 0.1256 0.1007 1.9999 0.1310 0.1017 0.4403
CQR 0.7961 0.1268 0.1015 1.9994 0.1319 0.1022 0.4408
WCQR 0.7957 0.1287 0.1015 1.9986 0.1352 0.1052 0.4445
200 LS 0.8124 0.1229 0.0969 1.9922 0.1142 0.0911 0.3419
QR 0.8090 0.1019 0.0813 2.0029 0.0951 0.0757 0.3302
RR 0.8102 0.0941 0.0749 1.9984 0.0862 0.0689 0.3141
CQR 0.8096 0.0935 0.0743 1.9988 0.0860 0.0689 0.3138
WCQR 0.8096 0.0937 0.0741 2.0004 0.0888 0.0713 0.3153
400 LS 0.7984 0.0831 0.0660 1.9967 0.0836 0.0665 0.2316
QR 0.8029 0.0696 0.0561 2.0016 0.0696 0.0555 0.2275
RR 0.8009 0.0623 0.0504 2.0006 0.0642 0.0505 0.2160
CQR 0.8007 0.0623 0.0502 2.0002 0.0637 0.0501 0.2162
WCQR 0.8012 0.0623 0.0498 2.0006 0.0634 0.0495 0.2160
Table 3

Simulation results for θ with F(4, 6) random error

n Method θ 1 (MEAN) θ 1 (SD) θ 1 (ABISE) θ 2 (MEAN) θ 2 (SD) θ 2 (ABISE) β ( t ) (RASE)
100 LS 0.7799 0.2804 0.2112 1.9781 0.2627 0.1940 0.5477
QR 0.7965 0.2059 0.1661 1.9935 0.1992 0.1591 0.5243
RR 0.7970 0.0870 0.0684 1.9940 0.0921 0.0731 0.3859
CQR 0.7956 0.0887 0.0693 1.9931 0.0929 0.0738 0.3882
WCQR 0.7990 0.0552 0.0431 1.9979 0.0584 0.0459 0.3654
200 LS 0.7998 0.1780 0.1379 1.9956 0.1754 0.1346 0.3935
QR 0.8030 0.1429 0.1161 1.9984 0.1391 0.1144 0.3791
RR 0.7998 0.0570 0.0446 1.9966 0.0564 0.0452 0.2762
CQR 0.7999 0.0567 0.0448 1.9972 0.0566 0.0455 0.2764
WCQR 0.8007 0.0330 0.0254 1.9985 0.0343 0.0273 0.2587
400 LS 0.8054 0.1368 0.1065 2.0018 0.1212 0.0966 0.2794
QR 0.8046 0.1078 0.0872 2.0088 0.1154 0.0948 0.2737
RR 0.7999 0.0426 0.0332 2.0009 0.0371 0.0301 0.1987
CQR 0.8004 0.0430 0.0337 2.0010 0.0379 0.0305 0.1990
WCQR 0.8003 0.0213 0.0168 2.0012 0.0211 0.0166 0.1877
Table 4

Simulation results for θ with Exp(1) random error

n Method θ 1 (MEAN) θ 1 (SD) θ 1 (ABISE) θ 2 (MEAN) θ 2 (SD) θ 2 (ABISE) β ( t ) (RASE)
100 LS 0.8006 0.1473 0.1171 2.0058 0.1478 0.1165 0.4633
QR 0.8001 0.1255 0.0956 2.0021 0.1279 0.0993 0.4427
RR 0.7976 0.0721 0.0565 2.0019 0.0742 0.0578 0.3696
CQR 0.7975 0.0717 0.0558 2.0017 0.0731 0.0567 0.3717
WCQR 0.7970 0.0484 0.0371 2.0025 0.0478 0.0370 0.3492
200 LS 0.8014 0.1012 0.0793 1.9995 0.1045 0.0819 0.3033
QR 0.8026 0.0780 0.0615 1.9981 0.0831 0.0651 0.2966
RR 0.7983 0.0486 0.0388 2.0022 0.0461 0.0358 0.2504
CQR 0.7984 0.0481 0.0384 2.0020 0.0460 0.0362 0.2500
WCQR 0.7997 0.0270 0.0214 2.0012 0.0260 0.0201 0.2350
400 LS 0.7996 0.0730 0.0581 1.9989 0.0671 0.0536 0.2133
QR 0.8010 0.0565 0.0445 2.0002 0.0512 0.0416 0.2078
RR 0.8000 0.0308 0.0246 1.9991 0.0336 0.0267 0.1776
CQR 0.7999 0.0307 0.0246 1.9992 0.0329 0.0264 0.1783
WCQR 0.8005 0.0157 0.0123 2.0001 0.0157 0.0123 0.1676
Table 5

Simulation results for θ with LN(0, 1) random error

n Method θ 1 (MEAN) θ 1 (SD) θ 1 (ABISE) θ 2 (MEAN) θ 2 (SD) θ 2 (ABISE) β ( t ) (RASE)
100 LS 0.8032 0.2684 0.2074 1.9916 0.2784 0.2112 0.5724
QR 0.8091 0.2067 0.1646 1.9828 0.2042 0.1671 0.5346
RR 0.8015 0.0937 0.0725 1.9982 0.0906 0.0711 0.3931
CQR 0.8017 0.0943 0.0736 1.9961 0.0910 0.0718 0.3942
WCQR 0.8009 0.0597 0.0456 1.9991 0.0574 0.0445 0.3625
200 LS 0.7921 0.1878 0.1469 1.9992 0.1898 0.1473 0.3840
QR 0.7934 0.1530 0.1265 1.9964 0.1561 0.1271 0.3807
RR 0.7953 0.0581 0.0460 1.9998 0.0606 0.0481 0.2774
CQR 0.7949 0.0589 0.0471 1.9998 0.0608 0.0484 0.2778
WCQR 0.7989 0.0327 0.0261 2.0005 0.0332 0.0259 0.2602
400 LS 0.8036 0.1418 0.1135 1.9964 0.1380 0.1076 0.2828
QR 0.8014 0.1186 0.0971 1.9953 0.1158 0.0928 0.2701
RR 0.7992 0.0415 0.0328 1.9975 0.0422 0.0337 0.1899
CQR 0.7994 0.0425 0.0337 1.9971 0.0430 0.0345 0.1910
WCQR 0.7997 0.0222 0.0175 2.0003 0.0207 0.0162 0.1788
Table 6

Simulation results for θ with Wb(1, 1.5) random error

n Method θ 1 (MEAN) θ 1 (SD) θ 1 (ABISE) θ 2 (MEAN) θ 2 (SD) θ 2 (ABISE) β ( t ) (RASE)
100 LS 0.7913 0.2093 0.1658 2.0096 0.2212 0.1768 0.5117
QR 0.7972 0.1829 0.1423 2.0069 0.1928 0.1527 0.5016
RR 0.7916 0.1024 0.0806 2.0000 0.1071 0.0843 0.4097
CQR 0.7924 0.1011 0.0799 2.0014 0.1039 0.0821 0.4072
WCQR 0.7958 0.0642 0.0500 2.0030 0.0667 0.0522 0.3721
200 LS 0.7888 0.1497 0.1199 1.9919 0.1450 0.1155 0.3712
QR 0.7881 0.1246 0.0984 1.9944 0.1175 0.0912 0.3419
RR 0.7991 0.0660 0.0515 1.9997 0.0718 0.0557 0.2909
CQR 0.7983 0.0650 0.0508 1.9994 0.0706 0.0550 0.2908
WCQR 0.7999 0.0381 0.0292 1.9992 0.0402 0.0306 0.2645
400 LS 0.7932 0.1092 0.0878 1.9959 0.1113 0.0869 0.2483
QR 0.7988 0.0852 0.0661 1.9961 0.0829 0.0637 0.2336
RR 0.7986 0.0473 0.0375 2.0016 0.0502 0.0403 0.1957
CQR 0.7984 0.0471 0.0374 2.0012 0.0502 0.0402 0.1959
WCQR 0.8007 0.0226 0.0177 1.9993 0.0233 0.0181 0.1800

6 Application to Tecator data

We illustrate the proposed approach by analyzing a real estate data set which is available at http://lib.stat.cmu.edu/datasets/tecator. The data have been widely used to predict fat content on samples of finely chopped meat [44,45,46]. For each food sample, the functional data consist of a 100-channel spectrum of absorbances recorded on a Tecator Infratec Food and Feed Analysis working in the wavelength range 850–1,050 nm by the near infrared transmission principle. More details on the data can be found in Ferraty and Vieu [47]. In this section, we construct the following PFLRM model:

Y i = α 1 Z i + α 2 U i + 850 1050 X i ( t ) β ( t ) d t + e i ,

We denote the scale of fat content as Y i , the protein content as Z i , the moisture content as U i , and the absorbances as X i ( t ) .

To evaluate the performance of the method, the sample is randomly divided into two subsamples: the training sample, I 1 = { ( X i , Y i , Z i , U i ) , I 1 = 165 } , where I 1 denotes the cardinality of I 1 and the remaining is test sample, I 2 = { ( X i , Y i , Z i , U i ) , I 2 = 50 } . To verify our robust estimation procedures, we reanalyzed this data set by including some outliers in the response variable. The case is considered through controlling the number of outliers and the values of outliers, we randomly generate κ outliers by increasing the value Y to Y + t ( κ = 20 , t = 5 ) . The training samples are used to estimate the parameters and the test samples are employed to verify the quality of predictions. For this, we compute the mean square error of prediction MSEP [44], which is defined by

MSEP = 1 50 i I 2 ( Y i Y ˆ i ) 2 / Var I 2 ( Y i ) ,

where Y ˆ i is the predicted value based on the training sample and Var I 2 is the variance of response variables from test sample. We use different estimation methods to predict the fat content of a meat sample based on its protein and/or moisture contents and/or its NIT absorbance spectrum, and the corresponding average MSEPs based on 200 times randomly splits are shown in Table 7. Compared with other methods, the MSEP of our proposed method (WCQR) is relatively smaller. Meanwhile, based on random simulations, the average estimated weight is ω opt = ( 0.4287 , 0.07228 , 0.0119 , 0.0114 , 0.0624 , 0.0631 , 0.0244 , 0.0022 , 0 ) and the corresponding standard deviation is (0.1608, 0.0895, 0.0481, 0.0257, 0.0275, 0.0403, 0.0690, 0.1261, 0). From Figure 1, we can see that the estimated slope functions β ( t ) ˆ by our WCQR method and CQR method, respectively, and it demonstrates that the two estimation approaches perform similarly.

Table 7

MSEPs for different methods

Method MSEP
LS 0.5684
QR 0.0522
CQR 0.0503
WCQR 0.0495
Figure 1 
               Curves of slope function 
                     
                        
                        
                           
                              
                                 β
                                 
                                    (
                                    
                                       t
                                    
                                    )
                                 
                              
                              
                                 ˆ
                              
                           
                        
                        \hat{\beta \left(t)}
                     
                   with CQR (solid line) and WCQR (dotted line).
Figure 1

Curves of slope function β ( t ) ˆ with CQR (solid line) and WCQR (dotted line).

7 Conclusion

In this paper, a novel and robust procedure based on WCQR and principal component basis function approximations is developed for PFLRMs. Theoretical properties of the estimators of both slope function and linear parameters are derived under some mild assumptions. We show that the estimators do not require any specification for error distribution, and thus more robust than those based on the LS, QR and CQR in the case of asymmetric errors. Furthermore, we present an efficient algorithm for the selection of optimal weights and discuss the selection of tuning parameters. Numerical studies and the real data analysis illustrate that our proposed method performs very well for moderate sample size.

In addition, we could extend the proposed procedure to the functional regression models with diverse dimension of covariates. These deserve to be studied further.

  1. Conflict of interest: Authors state no conflict of interest.

Appendix Proof of Theorem 3.1(i)

Let δ n = n 2 b 1 2 ( a + 2 b ) , u n = δ n 1 ( θ ˆ θ 0 ) , s n = δ n 1 ( γ ˆ γ 0 ) , v n k = δ n 1 ( c ˆ τ k c τ k ) , v n = ( v n 1 , , v n q ) T , R n i = 0 1 X i ( t ) β 0 ( t ) d t U i T γ 0 , Γ n = Z i Z i T , Λ n = diag ( λ ˆ 1 , , λ ˆ m ) , N = ( X 1 ( t ) , , X n ( t ) , Z 1 ( 1 ) , , Z n ( 1 ) , , Z 1 ( p ) , , Z n ( p ) ) T , and n = { ( u n , s n , v n ) : ( u n T , s n T , v n T ) T = C } , where C is a large enough constant, our aim is to show that for any given ϱ > 0 , there is a large constant C such that, for large n , we have

(A.1) P { inf ( u n , s n , v n ) n n ( θ 0 + δ n u n , γ 0 + δ n s n , c + δ n v n , ω ) > n ( θ 0 , γ 0 , c , ω ) } 1 ϱ .

This implies that, with probability tending to one, there is local minimum θ ^ and γ ^ in the ball { ( θ 0 + δ n u n , γ 0 + δ n s n , c + δ n v n ) : ( u n T , s n T , v n T ) T C } such that θ ^ θ 0 = O p ( δ n ) and γ ^ γ 0 = O p ( δ n ) .

First, by ϕ j ϕ ˆ j 2 = O p ( n 1 j 2 ) (see Crambes et al. [48]), one has

R n i 2 = 0 1 X i ( t ) β 0 ( t ) d t U i T γ 0 2 2 j = 1 m X i , ϕ ˆ j ϕ j γ 0 j 2 + 2 j = m + 1 X i , ϕ j γ 0 j 2 2 A 1 + 2 A 2 .

For A 1 , by Assumptions A1–A2 and the Hölder inequality, it is obtained

A 1 = j = 1 m X i , ϕ j ϕ ˆ j γ 0 j 2 c m j = 1 m ϕ j ϕ ˆ j 2 γ 0 j 2 c m j = 1 m O p ( n 1 j 2 2 b ) = O p ( n a + 4 b 4 a + 2 b ) = o p ( δ n 2 ) .

As for A 2 , due to

E j = m + 1 X i , ϕ j γ 0 j = 0 , Var j = m + 1 X i , ϕ j γ 0 j = j = m + 1 λ j γ 0 j 2 c j = m + 1 j ( a + 2 b ) = O n a + 2 b 1 a + 2 b ,

one has A 2 = O p n a + 2 b 1 a + 2 b = o p ( δ n 2 ) . Taking these together, then R n i 2 = O p n a + 2 b 1 a + 2 b = o p ( δ n 2 ) .

S n = n ( θ 0 + δ n u n , γ 0 + δ n s n , c + δ n v n , ω ) n ( θ 0 , γ 0 , c , ω ) ,

By the study of Knight [49],

z y z = y sgn ( z ) + 2 ( y z ) { I ( 0 < z < y ) I ( y < z < 0 ) } ,

then

ρ τ ( r s ) ρ τ ( r ) = s ( I ( r < 0 ) τ ) + 0 s [ I ( r t ) I ( r 0 ) ] d t ,

we can rewrite S n as

(A.2) S n = k = 1 q ω k i = 1 n δ n ( Z i T u n + U i T s n + v n k ) [ I ( ε i < R n i + c τ k ) τ k ] + k = 1 q ω k i = 1 n 0 δ n ( Z i T u n + U i T s n + v n k ) [ I ( ε i x + R n i + c τ k ) I ( ε i R n i + c τ k ) ] d x = n δ n ( A n T u n + C n T s n + D n T v n ) + k = 1 q ω k B n ( k ) ,

where

B n ( k ) = i = 1 n 0 δ n ( Z i T u n + U i T s n + v n k ) [ I ( ε i x + R n i + c τ k ) I ( ε i R n i + c τ k ) ] d x , A n = n 1 / 2 i = 1 n Z i k = 1 q ω k [ I ( ε i < R n i + c τ k ) τ k ] , C n = n 1 / 2 i = 1 n U i k = 1 q ω k [ I ( ε i < R n i + c τ k ) τ k ] , D n , k = n 1 / 2 i = 1 n ω k [ I ( ε i < R n i + c τ k ) τ k ] , D n = ( D n , 1 , , D n , q ) T .

Note that, by Assumptions A2 and A4, respectively, we have Γ n = O ( 1 ) , Λ n = O ( 1 ) , since ε i is independent of Z i and X i ( t ) , it have that E ( A n T u n ) = 0 , E { ( A n T u n ) 2 } = u n T E ( A n A n T ) u n = O ( u n 2 ) . Then we have, A n T u n = O ( u n ) . Similarly, we get C n T s n = O ( s n ) . This combined with (A.2) leads to

(A.3) S n = k = 1 q ω k B n ( k ) + o p ( n δ n 2 ) u n + o p ( n δ n 2 ) s n .

Invoking Assumption A7, a simple calculation yields

(A.4) E ( B n ( k ) N ) = E i = 1 n 0 δ n ( Z i T u n + U i T s n + v n k ) [ I ( ε i x + R n i + c τ k ) I ( ε i R n i + c τ k ) ] d x N = i = 1 n 0 δ n ( Z i T u n + U i T s n + v n k ) [ F ( x + R n i + c τ k ) F ( R n i + c τ k ) ] d x = i = 1 n 0 δ n ( Z i T u n + U i T s n + v n k ) [ f ( R n i + c τ k ) x ( 1 + o p ( 1 ) ) ] d x 3 f ( c τ k ) 2 n δ n 2 ( v n k 2 + u n T Γ n u n + s n T Λ n s n ) ( 1 + o p ( 1 ) ) .

Similarly,

(A.5) Var ( B n ( k ) N ) = Var i = 1 n 0 δ n ( Z i T u n + U i T s n + v n k ) [ I ( ε i x + R n i + c τ k ) I ( ε i R n i + c τ k ) ] d x N i = 1 n E 0 δ n ( Z i T u n + U i T s n + v n k ) [ I ( ε i x + R n i + c τ k ) I ( ε i R n i + c τ k ) ] d x 2 N

i = 1 n 0 δ n ( Z i T u n + U i T s n + v n k ) 0 δ n ( Z i T u n + U i T s n + v n k ) × [ F ( δ n ( Z i T u n + U i T s n + v n k ) + R n i + c τ k ) F ( R n i + c τ k ) ] d x 1 d x 2 o i = 1 n δ n ( Z i T u n + U i T s n + v n k ) 2 = o p ( n δ n 2 ) ( u n 2 + s n 2 + v n 2 ) .

Hence,

B n ( k ) 3 f ( c τ k ) 2 n δ n 2 ( v n k 2 + u n T Γ n u n + s n T Λ n s n ) ( 1 + o p ( 1 ) ) .

Then, we can obtain

(A.6) S n 3 2 n δ n 2 k = 1 q ω k f ( c τ k ) ( v n k 2 + u n T Γ n u n + s n T Λ n s n ) ( 1 + o p ( 1 ) ) + o p ( n δ n 2 ) ( u n 2 + s n 2 + v n 2 ) .

It follows from (A.2) to (A.6), we can obtain that S n is dominated by the positive quadratic term n δ n 2 k = 1 q ω k f ( c τ k ) ( v n k 2 + u n T Γ n u n + s n T Λ n s n ) as long as C is large enough and there exists local minimizer θ ^ and γ ^ such that

(A.7) θ ^ θ 0 = O p ( δ n ) , γ ^ γ 0 = O p ( δ n ) .

Observe that

β ˆ ( t ) β 0 ( t ) 2 = j = 1 m γ ˆ j ϕ ˆ j j = 1 γ 0 j ϕ j 2 2 j = 1 m γ ˆ j ϕ ˆ j j = 1 m γ 0 j ϕ j 2 + 2 j = m + 1 γ 0 j ϕ j 2 4 j = 1 m ( γ ˆ j γ 0 j ) ϕ ˆ j 2 + 4 j = 1 m γ 0 j ( ϕ ˆ j ϕ j ) 2 + 2 j = m + 1 γ 0 j 2 4 J 1 + 4 J 2 + 2 J 3 .

By Assumption A2, the orthogonality of { ϕ ˆ j } and ϕ j ϕ ˆ j 2 = O p ( n 1 j 2 ) , one has

(A.8) J 1 = j = 1 m ( γ ˆ j γ 0 j ) ϕ ˆ j 2 j = 1 m γ ˆ j γ 0 j 2 = γ ˆ γ 0 2 = O p ( δ n 2 ) ,

(A.9) J 2 = j = 1 m γ 0 j ( ϕ ˆ j ϕ j ) 2 m j = 1 m ϕ ˆ j ϕ j 2 γ 0 j 2 m n O p j = 1 m j 2 γ 0 j 2 = O p n 1 m j = 1 m j 2 2 b = O p ( n 1 m ) = o p ( n 2 b 1 a + 2 b ) = o p ( δ n 2 ) ,

and

(A.10) J 3 = j = m + 1 γ 0 j 2 C j = m + 1 j 2 b = O ( n 2 b 1 a + 2 b ) = O ( δ n 2 ) .

Then, combining (A.8)–(A.10), we can complete the proof of Theorem 3.1(i).

B Proof of Theorem 3.1(ii)

According to Theorem 3.1(i), we know that, as n , with probability tending to 1, n ( θ , γ , c , ω ) attains the minimal value at ( θ ˆ , γ ˆ , c ˆ , ω ) . Then, we have the following score equations:

(A.11) n 1 k = 1 q ω k i = 1 n Z i T ψ τ k ( Y i c τ k ˆ Z i T θ ˆ U i T γ ˆ ) = 0 ,

(A.12) n 1 k = 1 q ω k i = 1 n U i T ψ τ k ( Y i c τ k ˆ Z i T θ ˆ U i T γ ˆ ) = 0 ,

where ψ τ k ( u ) = ρ τ k ( u ) = τ k I ( u < 0 ) . By equations (A.11) and (A.12), we have

(A.13) n 1 k = 1 q ω k i = 1 n Z i T ψ τ k ( ε i c τ k R n i Z i T ( θ ˆ θ 0 ) U i T ( γ ˆ γ 0 ) ) = 0 ,

(A.14) n 1 k = 1 q ω k i = 1 n U i T ψ τ k ( ε i c τ k R n i Z i T ( θ ˆ θ 0 ) U i T ( γ ˆ γ 0 ) ) = 0 .

Moreover, we can write equation (A.13) as

( A.13 ) = i = 1 n H n + k = 1 q ω k B n 1 ( k ) + k = 1 q ω k B n 2 ( k ) ,

where

H n = 1 n i = 1 n Z i k = 1 q ω k [ I ( ε i < R n i + c τ k ) τ k ] , B n 1 ( k ) = 1 n i = 1 n Z i [ F ( R n i + c τ k ) F ( c τ k + R n i + Z i T ( θ ˆ θ 0 ) + U i T ( γ ˆ γ 0 ) ) ] , B n 2 ( k ) = 1 n i = 1 n Z i { [ I ( ε i < R n i + c τ k ) I ( ε i < c τ k + R n i + Z i T ( θ ˆ θ 0 ) + U i T ( γ ˆ γ 0 ) ) ] [ F ( R n i + c τ k ) F ( c τ k + R n i + Z i T ( θ ˆ θ 0 ) + U i T ( γ ˆ γ 0 ) ) ] } .

Invoking Taylor expansion, a simple calculation yields

B n 1 ( k ) = 1 n i = 1 n f ( c τ k ) [ Z i Z i T ( θ ˆ θ 0 ) + Z i U i T ( γ ˆ γ 0 ) ] ( 1 + o ( 1 ) ) .

By direct calculation of the mean and variance, we can show, as in the study by Jiang et al. (2012), that B n 2 ( k ) = o p ( δ n ) . Then, we have

(A.15) 1 n i = 1 n Z i k = 1 q ω k [ I ( ε i < R n i + c τ k ) τ k ] = 1 n i = 1 n k = 1 q ω k f ( c τ k ) [ Z i Z i T ( θ ˆ θ 0 ) + Z i U i T ( γ ˆ γ 0 ) ] ( 1 + o ( 1 ) ) .

Similarly, we have

(A.16) 1 n i = 1 n U i k = 1 q ω k [ I ( ε i < R n i + c τ k ) τ k ] = 1 n i = 1 n k = 1 q ω k f ( c τ k ) [ U i Z i T ( θ ˆ θ 0 ) + U i U i T ( γ ˆ γ 0 ) ] ( 1 + o ( 1 ) ) .

Let ϒ n = 1 n i = 1 n U i [ τ k I ( ε i < R n i + c τ k ) ] , Φ n = 1 n i = 1 n U i U i T , Ψ n = 1 n i = 1 n U i Z i T . By equation (A.16), we have

(A.17) γ ˆ γ 0 = ( Φ n + o p ( 1 ) ) 1 [ ϒ n + Ψ n ( θ 0 θ ˆ ) ] .

Substituting equation (A.17) into equation (A.15), we can obtain that

(A.18) 1 n i = 1 n k = 1 q ω k f ( c τ k ) Z i [ Z i Ψ n T Φ n 1 U i ] T ( θ ˆ θ 0 ) + o p ( θ ˆ θ 0 ) = 1 n i = 1 n k = 1 q ω k Z i ( [ τ k I ( ε i < R n i + c τ k ) ] f ( c τ k ) U i T Φ n 1 ϒ n ) .

Note that

(A.19) 1 n i = 1 n k = 1 q ω k f ( c τ k ) Ψ n T Φ n 1 U i [ Z i Ψ n T Φ n 1 U i ] T = 0 ,

(A.20) 1 n i = 1 n k = 1 q ω k f ( c τ k ) Ψ n T Φ n 1 U i { [ τ k I ( ε i < R n i + c τ k ) ] U i T Φ n 1 ϒ n } = 0 ,

According to equation (A.18)–(A.20), it is easy to show that

( θ ˆ θ 0 ) = i = 1 n k = 1 q ω k f ( c τ k ) Z ˜ i Z ˜ i T 1 i = 1 n k = 1 q ω k [ I ( ε i < R n i + c τ k ) τ k ] Z ˜ i ,

where Z ˜ i = Z i Ψ n T Φ n 1 U i . According to Lemma 1 in the study by Yu et al. (2016) and Assumptions A5–A6, as n , we have

1 n i = 1 n Z ˜ i Z ˜ i T p Σ .

Note

i = 1 n k = 1 q ω k f ( c τ k ) 1 i = 1 n k = 1 q ω k [ I ( ε i < R n i + c τ k ) τ k ] Z ˜ i

is normal with mean 0 and covariance matrix σ 2 ( ω ) Σ , then we have

n ( θ ^ θ 0 ) N ( 0 , σ 2 ( ω ) Σ 1 ) .

We complete the proof of Theorem 3.1(ii).

References

[1] H. Cardot , F. Ferraty , and P. Sarda , Functional linear model, Statist. Probab. Lett. 45 (1999), no. 1, 11–22, https://doi.org/10.1016/S0167-7152(99)00036-X . 10.1016/S0167-7152(99)00036-XSearch in Google Scholar

[2] H. Cardot , F. Ferraty , and P. Sarda , Spline estimators for the functional linear model, Statist. Sinica 13 (2003), no. 3, 571–591, http://www.jstor.org/stable/24307112. Search in Google Scholar

[3] F. Yao , H. G. Müller , and J. L. Wang , Functional linear regression analysis for longitudinal data, Ann. Statist. 33 (2005), no. 6, 2873–2903, https://doi.org/10.1214/009053605000000660. Search in Google Scholar

[4] J. Ramsay and B. Silverman , Functional Data Analysis, 2nd edn., Springer, New York, 2005. 10.1007/b98888Search in Google Scholar

[5] T. Cai and P. Hall , Prediction in functional linear regression, Ann. Statist. 34 (2006), no. 5, 2159–2179, https://doi.org/10.1214/009053606000000830 . 10.1214/009053606000000830Search in Google Scholar

[6] H. Liu and H. Yang , Estimation and variable selection in single-index composite quantile regression, Comm. Statist. Simulation Comput. 46 (2016), no. 9, 7022–7039. 10.1080/03610918.2016.1222424Search in Google Scholar

[7] P. Hall and J. Horowitz , Methodology and convergence rates for functional linear regression, Ann. Statist. 35 (2007), no. 1, 70–91, https://doi.org/10.1214/009053606000000957. Search in Google Scholar

[8] C. Crambes , A. Kneip , and P. Sarda , Smoothing splines estimators for functional linear regression, Ann. Statist. 37 (2009), no. 1, 35–72, https://doi.org/10.1214/07-AOS563. Search in Google Scholar

[9] P. Hall and Y. Yang , Ordering and selecting components in multivariate or functional data linear prediction, J. R. Stat. Soc. Ser. B. Stat. Methodol. 72 (2010), no. 1, 93–110. 10.1111/j.1467-9868.2009.00727.xSearch in Google Scholar

[10] T. Cai and M. Yuan , Minimax and adaptive prediction for functional linear regression, J. Amer. Statist. Assoc. 107 (2012), no. 499, 1201–1216, https://www.jstor.org/stable/23427425. 10.1080/01621459.2012.716337Search in Google Scholar

[11] A. Delaigle and P. Hall , Methodology and theory for partial least squares applied to functional data, Ann. Statist. 40 (2012), no. 1, 322–352, https://doi.org/10.1214/11-AOS958. Search in Google Scholar

[12] E. R. Lee and B. U. Park , Sparse estimation in functional linear regression, J. Multivariate Anal. 105 (2012), no. 1, 1–17, https://doi.org/10.1016/j.jmva.2011.08.005. Search in Google Scholar

[13] L. Huang , H. Wang , and A. Zheng , The M-estimator for functional linear regression model, Statist. Probab. Lett. 88 (2014), no. C, 165–173, https://doi.org/10.1016/j.spl.2014.01.016. Search in Google Scholar

[14] H. J. Shin , Partial functional linear regression, J. Statist. Plann. Inference 139 (2009), no. 10, 3405–3418, https://doi.org/10.1016/j.jspi.2009.03.001 . 10.1016/j.jspi.2009.03.001Search in Google Scholar

[15] J. Zhou , Z. Chen , and Q. Peng , Polynomial spline estimation for partial functional linear regression models, Comput. Statist. 31 (2016), no. 3, 1107–1129, https://doi.org/10.1007/s00180-015-0636-0. Search in Google Scholar

[16] D. Kong , K. Xue , F. Yao , and H. Zhang , Partially functional linear regression in high dimensions, Biometrika 103 (2016), no. 1, 147–159, https://doi.org/10.1093/biomet/asv062. Search in Google Scholar

[17] N. Ling , G. Aneiros , and P. Vieu , kNN estimation in functional partial linear modeling, Statist. Papers 61 (2020), no. 1, 423–444, https://doi.org/10.1007/s00362-017-0946-0. Search in Google Scholar

[18] D. Kong , J. Ibrahim , E. Lee , and H. Zhu , FLCRM: functional linear cox regression model, Biometrics 74 (2018), no. 1, 109–117. 10.1111/biom.12748Search in Google Scholar PubMed PubMed Central

[19] R. Cao , J. Du , J. Zhou , and T. Xie , FPCA-based estimation for generalized functional partially linear models, Statist. Papers 61 (2020), no. 6, 2715–2735, https://doi.org/10.1007/s00362-018-01066-8. Search in Google Scholar

[20] M. Yuan and Y. Zhang , Test for the parametric part in partial functional linear regression based on B-spline, Comm. Statist. Simulation Comput. 50 (2021), no. 1, 1–15, https://doi.org/10.1080/03610918.2019.1667391. Search in Google Scholar

[21] H. Zhu , R. Zhang , Z. Yu , H. Lian , and Y. Liu , Estimation and testing for partially functional linear errors-in-variables models, J. Multivariate Anal. 170 (2019), 296–314, https://doi.org/10.1016/j.jmva.2018.11.005. Search in Google Scholar

[22] Q. Li , X. Tan , and L. Wang , Testing for error correlation in partially functional linear regression models, Comm. Statist. Theory Methods 50 (2021), no. 3, 747–761, https://doi.org/10.1080/03610926.2019.1642492 . 10.1080/03610926.2019.1642492Search in Google Scholar

[23] R. Koenker and G. Basset , Regression quantiles, Econometrica 46 (1978), no. 1, 33–50, https://doi.org/10.2307/1913643 . 10.2307/1913643Search in Google Scholar

[24] Y. Lu , J. Du , Z. Sun , Functional partially linear quantile regression model, Metrika 77 (2014), no. 2, 317–332, https://doi.org/10.1007/s00184-013-0439-7. Search in Google Scholar

[25] H. Ma , T. Li , H. Zhu , and Z. Zhu , Quantile regression for functional partially linear model in ultra-high dimensions, Comput. Statist. Data Anal. 129 (2019), no. 1, 135–147, https://doi.org/10.1016/j.csda.2018.06.005. Search in Google Scholar

[26] R. Koenker , Quantile Regression, Cambridge University Press, Cambridge, 2005. 10.1017/CBO9780511754098Search in Google Scholar

[27] H. Zou and M. Yuan , Composite quantile regression and the Oracle model selection theory, Ann. Statist. 36 (2008), no. 3, 1108–1126, https://www.jstor.org/stable/25464661. 10.1214/07-AOS507Search in Google Scholar

[28] B. Kai , R. Li , and H. Zou , Local composite quantile regression smoothing: an efficient and safe alternative to local polynomial regression, J. R. Stat. Soc. Ser. B Stat. Methodol. 72 (2010), no. 1, 49–69, https://doi.org/10.1111/j.1467-9868.2009.00725.x. Search in Google Scholar PubMed PubMed Central

[29] B. Kai , R. Li , and H. Zou , New efficient estimation and variable selection methods for semiparametric varying-coefficient partially linear models, Ann. Statist. 39 (2011), no. 1, 305–332, https://doi.org/10.1214/10-AOS842. Search in Google Scholar PubMed PubMed Central

[30] J. Guo , M. Tang , M. Tian , and K. Zhu , Variable selection in high-dimensional partially linear additive models for composite quantile regression, Comput. Statist. Data Anal. 65 (2013), 56–67, https://doi.org/10.1016/j.csda.2013.03.017. Search in Google Scholar

[31] R. Jiang , Z. Zhou , W. Qian , and Y. Chen , Two step composite quantile regression for single-index models, Comput. Statist. Data Anal. 64 (2013), 180–191, https://doi.org/10.1016/j.csda.2013.03.014. Search in Google Scholar

[32] R. Jiang , W. Qian , and Z. Zhou , Test for single-index composite quantile regression, Hacet. J. Math. Stat. 43 (2014), no. 5, 861–871. Search in Google Scholar

[33] J. Du , D. Xu , and R. Cao , Estimation and variable selection for partially functional linear models, J. Korean Statist. Soc. 47 (2018), no. 4, 436–449, https://doi.org/10.1016/j.jkss.2018.05.002. Search in Google Scholar

[34] P. Yu , T. Li , Z. Zhu , and Z. Zhang , Composite quantile estimation in partial functional linear regression model with dependent errors, Metrika 82 (2019), no. 6, 633–656, https://doi.org/10.1007/s00184-018-0699-3. Search in Google Scholar

[35] X. Jiang , J. Jiang , and X. Song , Oracle model selection for nonlinear models based on weighted composite quantile regression, Statist. Sinica 22 (2012), no. 4, 1479–1506, https://doi.org/10.5705/ss.2010.203. Search in Google Scholar

[36] C. Guo , H. Yang , and J. Lv , Robust variable selection in high-dimensional varying coefficient models based on weighted composite quantile regression, Statist. Papers 58 (2017), no. 4, 1009–1033, https://doi.org/10.1007/s00362-015-0736-5. Search in Google Scholar

[37] R. Jiang , W. Qian , and Z. Zhou , Weighted composite quantile regression for partially linear varying coefficient models, Comm. Statist. Theory Methods 47 (2018), no. 16, 3987–4005, https://doi.org/10.1080/03610926.2017.1366522. Search in Google Scholar

[38] J. Sun , Y. Gai , and L. Lin , Weighted local linear composite quantile estimation for the case of general error distributions, J. Statist. Plann. Inference 143 (2013), no. 6, 1049–1063, https://doi.org/10.1016/j.jspi.2013.01.002. Search in Google Scholar

[39] R. Jiang , W. Qian , and Z. Zhou , Weighted composite quantile regression for single-index models, J. Multivariate Anal. 148 (2016), 34–48, https://doi.org/10.1016/j.jmva.2016.02.015. Search in Google Scholar

[40] P. Yu , Z. Zhang , and J. Du , A test of linearity in partial functional linear regression, Metrika 79 (2016), no. 8, 953–969, https://doi.org/10.1007/s00184-016-0584-x. Search in Google Scholar

[41] L. Horváth and P. Kokoszka , Inference for Functional Data with Applications, Springer, New York, 2012. 10.1007/978-1-4614-3655-3Search in Google Scholar

[42] B. Silverman , Density Estimation, Chapman and Hall, London, 1986. Search in Google Scholar

[43] J. Sun and L. Lin , Local rank estimation and related test for varying-coefficient partially linear models, J. Nonparametr. Stat. 26 (2014), no. 1, 187–206, https://doi.org/10.1080/10485252.2013.841910. Search in Google Scholar

[44] G. Aneiros-Pérez and P. Vieu , Semi-functional partial linear regression, Statist. Probab. Lett. 76 (2006), no. 11, 1102–1110, https://doi.org/10.1016/j.spl.2005.12.007. Search in Google Scholar

[45] G. Aneiros-Pérez and P. Vieu , Partial linear modelling with multi-functional covariates, Comput. Statist. 30 (2015), no. 3, 647–671, https://doi.org/10.1007/s00180-015-0568-8. Search in Google Scholar

[46] P. Yu , J. Du , and Z. Zhan , Single-index partially functional linear regression model, Statist. Papers 61 (2020), no. 3, 1107–1123, https://doi.org/10.1007/s00362-018-0980-6. Search in Google Scholar

[47] F. Ferraty and P. Vieu , Nonparametric Functional Data Analysis: Theory and Practice, Springer, New York, 2006. Search in Google Scholar

[48] C. Crambes , A. Kneip , and P. Sarda , Smoothing splines estimators for functional linear regression, Ann. Statist. 37 (2009), no. 1, 35–72, https://doi.org/10.1214/07-AOS563. Search in Google Scholar

[49] K. Knight , Limiting distributions for L1 regression estimators under general conditions, Ann. Statist. 26 (1998), no. 2, 755–770, https://www.jstor.org/stable/120050.10.1214/aos/1028144858Search in Google Scholar

Received: 2020-09-24
Revised: 2021-08-17
Accepted: 2021-08-19
Published Online: 2021-12-31

© 2021 Peng Cao and Jun Sun, published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Articles in the same Issue

  1. Regular Articles
  2. Sharp conditions for the convergence of greedy expansions with prescribed coefficients
  3. Range-kernel weak orthogonality of some elementary operators
  4. Stability analysis for Selkov-Schnakenberg reaction-diffusion system
  5. On non-normal cyclic subgroups of prime order or order 4 of finite groups
  6. Some results on semigroups of transformations with restricted range
  7. Quasi-ideal Ehresmann transversals: The spined product structure
  8. On the regulator problem for linear systems over rings and algebras
  9. Solvability of the abstract evolution equations in Ls-spaces with critical temporal weights
  10. Resolving resolution dimensions in triangulated categories
  11. Entire functions that share two pairs of small functions
  12. On stochastic inverse problem of construction of stable program motion
  13. Pentagonal quasigroups, their translatability and parastrophes
  14. Counting certain quadratic partitions of zero modulo a prime number
  15. Global attractors for a class of semilinear degenerate parabolic equations
  16. A new implicit symmetric method of sixth algebraic order with vanished phase-lag and its first derivative for solving Schrödinger's equation
  17. On sub-class sizes of mutually permutable products
  18. Asymptotic solution of the Cauchy problem for the singularly perturbed partial integro-differential equation with rapidly oscillating coefficients and with rapidly oscillating heterogeneity
  19. Existence and asymptotical behavior of solutions for a quasilinear Choquard equation with singularity
  20. On kernels by rainbow paths in arc-coloured digraphs
  21. Fully degenerate Bell polynomials associated with degenerate Poisson random variables
  22. Multiple solutions and ground state solutions for a class of generalized Kadomtsev-Petviashvili equation
  23. A note on maximal operators related to Laplace-Bessel differential operators on variable exponent Lebesgue spaces
  24. Weak and strong estimates for linear and multilinear fractional Hausdorff operators on the Heisenberg group
  25. Partial sums and inclusion relations for analytic functions involving (p, q)-differential operator
  26. Hodge-Deligne polynomials of character varieties of free abelian groups
  27. Diophantine approximation with one prime, two squares of primes and one kth power of a prime
  28. The equivalent parameter conditions for constructing multiple integral half-discrete Hilbert-type inequalities with a class of nonhomogeneous kernels and their applications
  29. Boundedness of vector-valued sublinear operators on weighted Herz-Morrey spaces with variable exponents
  30. On some new quantum midpoint-type inequalities for twice quantum differentiable convex functions
  31. Quantum Ostrowski-type inequalities for twice quantum differentiable functions in quantum calculus
  32. Asymptotic measure-expansiveness for generic diffeomorphisms
  33. Infinitesimals via Cauchy sequences: Refining the classical equivalence
  34. The (1, 2)-step competition graph of a hypertournament
  35. Properties of multiplication operators on the space of functions of bounded φ-variation
  36. Disproving a conjecture of Thornton on Bohemian matrices
  37. Some estimates for the commutators of multilinear maximal function on Morrey-type space
  38. Inviscid, zero Froude number limit of the viscous shallow water system
  39. Inequalities between height and deviation of polynomials
  40. New criteria-based ℋ-tensors for identifying the positive definiteness of multivariate homogeneous forms
  41. Determinantal inequalities of Hua-Marcus-Zhang type for quaternion matrices
  42. On a new generalization of some Hilbert-type inequalities
  43. On split quaternion equivalents for Quaternaccis, shortly Split Quaternaccis
  44. On split regular BiHom-Poisson color algebras
  45. Asymptotic stability of the time-changed stochastic delay differential equations with Markovian switching
  46. The mixed metric dimension of flower snarks and wheels
  47. Oscillatory bifurcation problems for ODEs with logarithmic nonlinearity
  48. The B-topology on S-doubly quasicontinuous posets
  49. Hyers-Ulam stability of isometries on bounded domains
  50. Inhomogeneous conformable abstract Cauchy problem
  51. Path homology theory of edge-colored graphs
  52. Refinements of quantum Hermite-Hadamard-type inequalities
  53. Symmetric graphs of valency seven and their basic normal quotient graphs
  54. Mean oscillation and boundedness of multilinear operator related to multiplier operator
  55. Numerical methods for time-fractional convection-diffusion problems with high-order accuracy
  56. Several explicit formulas for (degenerate) Narumi and Cauchy polynomials and numbers
  57. Finite groups whose intersection power graphs are toroidal and projective-planar
  58. On primitive solutions of the Diophantine equation x2 + y2 = M
  59. A note on polyexponential and unipoly Bernoulli polynomials of the second kind
  60. On the type 2 poly-Bernoulli polynomials associated with umbral calculus
  61. Some estimates for commutators of Littlewood-Paley g-functions
  62. Construction of a family of non-stationary combined ternary subdivision schemes reproducing exponential polynomials
  63. On the evolutionary bifurcation curves for the one-dimensional prescribed mean curvature equation with logistic type
  64. On intersections of two non-incident subgroups of finite p-groups
  65. Global existence and boundedness in a two-species chemotaxis system with nonlinear diffusion
  66. Finite groups with 4p2q elements of maximal order
  67. Positive solutions of a discrete nonlinear third-order three-point eigenvalue problem with sign-changing Green's function
  68. Power moments of automorphic L-functions related to Maass forms for SL3(ℤ)
  69. Entire solutions for several general quadratic trinomial differential difference equations
  70. Strong consistency of regression function estimator with martingale difference errors
  71. Fractional Hermite-Hadamard-type inequalities for interval-valued co-ordinated convex functions
  72. Montgomery identity and Ostrowski-type inequalities via quantum calculus
  73. Universal inequalities of the poly-drifting Laplacian on smooth metric measure spaces
  74. On reducible non-Weierstrass semigroups
  75. so-metrizable spaces and images of metric spaces
  76. Some new parameterized inequalities for co-ordinated convex functions involving generalized fractional integrals
  77. The concept of cone b-Banach space and fixed point theorems
  78. Complete consistency for the estimator of nonparametric regression model based on m-END errors
  79. A posteriori error estimates based on superconvergence of FEM for fractional evolution equations
  80. Solution of integral equations via coupled fixed point theorems in 𝔉-complete metric spaces
  81. Symmetric pairs and pseudosymmetry of Θ-Yetter-Drinfeld categories for Hom-Hopf algebras
  82. A new characterization of the automorphism groups of Mathieu groups
  83. The role of w-tilting modules in relative Gorenstein (co)homology
  84. Primitive and decomposable elements in homology of ΩΣℂP
  85. The G-sequence shadowing property and G-equicontinuity of the inverse limit spaces under group action
  86. Classification of f-biharmonic submanifolds in Lorentz space forms
  87. Some new results on the weaving of K-g-frames in Hilbert spaces
  88. Matrix representation of a cross product and related curl-based differential operators in all space dimensions
  89. Global optimization and applications to a variational inequality problem
  90. Functional equations related to higher derivations in semiprime rings
  91. A partial order on transformation semigroups with restricted range that preserve double direction equivalence
  92. On multi-step methods for singular fractional q-integro-differential equations
  93. Compact perturbations of operators with property (t)
  94. Entire solutions for several complex partial differential-difference equations of Fermat type in ℂ2
  95. Random attractors for stochastic plate equations with memory in unbounded domains
  96. On the convergence of two-step modulus-based matrix splitting iteration method
  97. On the separation method in stochastic reconstruction problem
  98. Robust estimation for partial functional linear regression models based on FPCA and weighted composite quantile regression
  99. Structure of coincidence isometry groups
  100. Sharp function estimates and boundedness for Toeplitz-type operators associated with general fractional integral operators
  101. Oscillatory hyper-Hilbert transform on Wiener amalgam spaces
  102. Euler-type sums involving multiple harmonic sums and binomial coefficients
  103. Poly-falling factorial sequences and poly-rising factorial sequences
  104. Geometric approximations to transition densities of Jump-type Markov processes
  105. Multiple solutions for a quasilinear Choquard equation with critical nonlinearity
  106. Bifurcations and exact traveling wave solutions for the regularized Schamel equation
  107. Almost factorizable weakly type B semigroups
  108. The finite spectrum of Sturm-Liouville problems with n transmission conditions and quadratic eigenparameter-dependent boundary conditions
  109. Ground state sign-changing solutions for a class of quasilinear Schrödinger equations
  110. Epi-quasi normality
  111. Derivative and higher-order Cauchy integral formula of matrix functions
  112. Commutators of multilinear strongly singular integrals on nonhomogeneous metric measure spaces
  113. Solutions to a multi-phase model of sea ice growth
  114. Existence and simulation of positive solutions for m-point fractional differential equations with derivative terms
  115. Bernstein-Walsh type inequalities for derivatives of algebraic polynomials in quasidisks
  116. Review Article
  117. Semiprimeness of semigroup algebras
  118. Special Issue on Problems, Methods and Applications of Nonlinear Analysis (Part II)
  119. Third-order differential equations with three-point boundary conditions
  120. Fractional calculus, zeta functions and Shannon entropy
  121. Uniqueness of positive solutions for boundary value problems associated with indefinite ϕ-Laplacian-type equations
  122. Synchronization of Caputo fractional neural networks with bounded time variable delays
  123. On quasilinear elliptic problems with finite or infinite potential wells
  124. Deterministic and random approximation by the combination of algebraic polynomials and trigonometric polynomials
  125. On a fractional Schrödinger-Poisson system with strong singularity
  126. Parabolic inequalities in Orlicz spaces with data in L1
  127. Special Issue on Evolution Equations, Theory and Applications (Part II)
  128. Impulsive Caputo-Fabrizio fractional differential equations in b-metric spaces
  129. Existence of a solution of Hilfer fractional hybrid problems via new Krasnoselskii-type fixed point theorems
  130. On a nonlinear system of Riemann-Liouville fractional differential equations with semi-coupled integro-multipoint boundary conditions
  131. Blow-up results of the positive solution for a class of degenerate parabolic equations
  132. Long time decay for 3D Navier-Stokes equations in Fourier-Lei-Lin spaces
  133. On the extinction problem for a p-Laplacian equation with a nonlinear gradient source
  134. General decay rate for a viscoelastic wave equation with distributed delay and Balakrishnan-Taylor damping
  135. On hyponormality on a weighted annulus
  136. Exponential stability of Timoshenko system in thermoelasticity of second sound with a memory and distributed delay term
  137. Convergence results on Picard-Krasnoselskii hybrid iterative process in CAT(0) spaces
  138. Special Issue on Boundary Value Problems and their Applications on Biosciences and Engineering (Part I)
  139. Marangoni convection in layers of water-based nanofluids under the effect of rotation
  140. A transient analysis to the M(τ)/M(τ)/k queue with time-dependent parameters
  141. Existence of random attractors and the upper semicontinuity for small random perturbations of 2D Navier-Stokes equations with linear damping
  142. Degenerate binomial and Poisson random variables associated with degenerate Lah-Bell polynomials
  143. Special Issue on Fractional Problems with Variable-Order or Variable Exponents (Part I)
  144. On the mixed fractional quantum and Hadamard derivatives for impulsive boundary value problems
  145. The Lp dual Minkowski problem about 0 < p < 1 and q > 0
Downloaded on 12.9.2025 from https://www.degruyterbrill.com/document/doi/10.1515/math-2021-0095/html?lang=en
Scroll to top button