Home Variance reduction estimation for return models with jumps using gamma asymmetric kernels
Article Publicly Available

Variance reduction estimation for return models with jumps using gamma asymmetric kernels

  • Yuping Song ORCID logo EMAIL logo , Weijie Hou and Shengyi Zhou
Published/Copyright: April 9, 2019

Abstract

This paper discusses Nadaraya-Watson estimators for the unknown coefficients in second-order diffusion model with jumps constructed with Gamma asymmetric kernels. Compared with existing nonparametric estimators constructed with Gaussian symmetric kernels, local constant smoothing using Gamma asymmetric kernels possesses some extra advantages such as boundary bias correction, variance reduction and resistance to sparse design points, which is validated through theoretical details and finite sample simulation study. Under the regular conditions, the weak consistency and the asymptotic normality of these estimators are presented. Finally, the statistical advantages of the nonparametric estimators are depicted through 5-minute high-frequency data from Shenzhen Stock Exchange in China.

1 Introduction

In financial modeling and econometrics, diffusion models are widely used to describe uncertainty of changes in native economic variables and asset prices, one of the most famous models is the Black & Scholes asset pricing model. Moreover, as an extension of continuous trajectory process, diffusion model with jumps has a wide range of applications in the field of finance and economics. Jump-diffusion process Xt is represented by the following stochastic differential equation:

(1)dXt=μ(Xt)dt+σ(Xt)dWt+Ec(Xt,z)r(ω,dt,dz),

where Wt is a standard Brownian motion, μ(⋅) and σ(⋅) are the infinitesimal conditional drift and variance respectively, E=R{0}, r(ω,dt,dz)=(pq)(dt,dz),p(dt,dz) is a time-homogeneous Poisson random measure on R+×R independent of Wt, and q(dt, dz) is its intensity measure, that is, E[p(dt,dz)]=q(dt,dz)=f(z)dzdt, f(z) is its Lévy density. Moreover, jump factor in model (1) is introduced statistically to capture the fatter tail behavior of underlying asset and economically to accommodate impact of sudden and large shocks to financial markets, such as macroeconomic announcements or a dramatic interest rate cut by the Federal Reserve, more details discussed in Johannes (2004). However, on the one hand, because the Brownian motion is not differentiated at any point in probability 1, the mentioned model (1) above can not represent the integrated and differentiated process for integrated economic phenomena, that is, the observation of dynamic process at the discrete time is sometimes not easy to obtain, while the cumulation of the process can be directly observed; on the other hand, model (1) and its subsequent derivative models were mainly involved with the price of asset, while less models characterized returns of asset which is a complete and scale-free summary of the investment opportunity for average investors. Fortunately, the second-order diffusion process with jumps (2) provides a valuable model for the solution of these problems.

Here we extend the continuous autoregression of order two in Nicolau (2007) to a discontinuous and realistic case with jumps for return series based on model (1), which is called as the second-order jump-diffusion model (2),

(2){dYt=Xtdt,dXt=μ(Xt)dt+σ(Xt)dWt+Ec(Xt,z)r(ω,dt,dz),

where Xt represents the continuously compounded return of underlying assets and Yt denotes the asset price by means of the cumulation of the returns plus initial asset value. It also can model the integrated econometric phenomena, for example, continuously compounded return series Xt of underlying assets can not be directly observed, while its realization Yt such as the prices of assets can be directly observed.

Despite of the above two advantages, model (2) can be used commonly in empirical financial for another three advantages. Firstly, model (2) can accommodate nonstationary original process and be made stationary by differencing, which is a more widely used technique in empirical financial, which is verified through the Augmented Dickey-Fuller test statistic in the empirical analysis part. Secondly, compared with the model in Nicolau (2007), model (2) accommodate the impact of sudden and large shocks in combination with jump component, which is also testified through the test statistic proposed in Barndorff-Nielsen and Shephard (2006) for empirical financial data. Thirdly, model (2) can also be employed in empirical finance for integrated volatility. In empirical financial, Nicolau (2008) employed the second-order diffusion process to study the return series of three companies and three stock indices in USA and etc. Based on the generalized likelihood ratio statistics, Yan and Mei (2016) validated the parameter forms of the volatility function for returns of stock indices in Europe, America, Asia (Hong Kong, Japan, Korea) and four companies in USA stock market.

For the return model characterized by second-order diffusion process, many scholars considered parametric and nonparametric estimation for the unknown coefficients in model (2). For the case with c(x,z)0, Gloter (2001) obtained the parameter estimators for the second-order Ornstein-Uhlenbeck process, and Gloter (2006) gave the minimum deviation parameter estimation for the drift and diffusion coefficients. Not assuming the specific form of the coefficients, Nicolau (2007) systematically studied Nadaraya-Watson estimators, Wang and Lin (2011) presented the local linear estimations using symmetric kernel for bias correction, Hanif (2015) discussed the Nadaraya-Watson estimators using Gamma asymmetric kernel which didn’t coincide with the asymmetric kernel-based method introduced in Chen (2000). For model (2) with c(x,z)0, Song (2017) provided the nonparametric estimators for the infinitesimal coefficients μ(x) and σ2(x)+Ec2(x,z)f(z)dz in high frequency data based on Gaussian symmetric kernel, Chen and Zhang (2015) discussed the local linear estimators for them based on symmetric kernels, Song, Lin, and Wang (2013) proposed a re-weighted Nadaraya-Watson estimator for the infinitesimal conditional expectation, Funke and Schmisser (2018) constructed adaptive nonparametric estimator for the drift coefficient based on penalized least squares method.

However, the current research on the bias corrections for the unknown coefficients in the second-order jump-diffusion process is mainly focused on the simple bias representation (mainly using local linear or re-weighted nonparametric method). According to the asymptotic distribution theorem of the estimators, we can easily find that the mean square error of the estimators is mainly from two parts, one is the mean error of the estimator and the other is the perturbation caused by the asymptotic normal variance. The previous research work does not take variance reduction into full account for these unknown quantities. Based on Gamma asymmetric kernel fully discussed in Chen (2000), Kristensen (2010) considered a kernel-based approach for the realized integrated volatility estimation, Gospodinovy and Hirukawa (2012) constructed nonparametric estimators for scalar diffusion models of spot interest rates. Nadaraya-Watson estimators using Gamma asymmetric kernel have some benefits such as variable bandwidth, variance reduction and resistance to sparse design. Compared with Nadaraya-Watson estimators constructed with Gaussian symmetric kernels, firstly, the curve shapes of Gamma asymmetric kernel vary with the smoothing parameter and the location of the design point similar as the variable bandwidth methods. Secondly, the finite variance of the curve estimation decreases, especially for sparse design points, due to the fact that the smaller asymptotic convergence rate and their variances vary along with the location of the design point x, which is shown in Theorem 2 and Theorem 3. Thirdly, Gamma asymmetric kernels are free from boundary bias (allowing a larger bandwidth to pool more data) and achieve the optimal rate of convergence in mean square error within the class of nonnegative kernel estimators with order two. Fourthly, the estimators constructed with Gamma kernels possesses another advantage: nonnegativity for volatility function compared with local linear estimators. In conclusion, asymmetric kernels is a combination of a boundary correction, variance reduction device and a “variable bandwidth” method.

In this paper, we will propose Nadaraya-Watson estimators for μ(x) and σ2(x)+Ec2(x,z)f(z)dz using Gamma asymmetric kernels in model (2) for variance reduction. It is organized as follows. In Section 2, we propose Nadaraya-Watson estimators with asymmetric kernels and some ordinary assumptions for model (2). We present the asymptotic results in Section 3. Section 4 presents the finite sample performance for two models through Monte Carlo simulation study. Estimators are depicted through some empirical financial data in Section 5. Section 6 concludes. Some technical lemmas and the main proofs are explicitly shown in Appendix A.2.

2 Nadaraya-Watson estimators with asymmetric kernels and assumptions

According to the asymmetric kernels used in Chen (2000), the Gamma kernel function is defined as

(3)KG(x/h+1,h)(u)=ux/hexp(u/h)hx/h+1Γ(x/h+1)0u,

where Γ(m)=0ym1exp(y)dy, m > 0 is the Gamma function and h is the smoothing parameter. Since the density of Gamma distribution has support [0, ∞), the Gamma kernel function does not generate boundary bias for nonnegative variables or nonnegative part of underlying variables. Here we use modified Gamma kernel function KG(x/h+1,h)(u) instead of KG(x/h,h)(u) due to the fact KG(x/h,h)(u) is unbounded near at x = 0. The Gamma function has shapes varying with the design point x, which is nonnegative and asymmetric, especially at the boundary points. Furthermore, the asymptotic variance of the nonparametric Gamma kernel estimation depends on the design point x, which yields the optimal rate of convergence in mean integrated squared error of nonnegative kernels.

In this paper we consider the Nadaraya-Watson estimators with Gamma asymmetric kernel for the unknown coefficients of the second-order jump-diffusion model. Different from model (1), nonparametric estimations constructed for the coefficients in second-order jump-diffusion model (2) give rise to new challenges for two main reasons.

On the one hand, we usually get observations {YiΔn;i=1,2,} rather than {XiΔn;i=1,2,}. The value of Xti cannot be obtained from Yti=Y0+0tiXsds in a fixed sample intervals. Additionally, nonparametric estimations of the unknown qualities in model (2) could not be constructed on the observations {YiΔn;i=1,2,} due to the unknown conditional distribution. As Nicolau (2007) showed, with observations {YiΔn;i=1,2,} and given that

YiΔnY(i1)Δn=(i1)ΔniΔnXudu,

we can obtain an approximation value of XiΔn by

(4)X~iΔn=YiΔnY(i1)ΔnΔn.

On the other hand, the Markov properties for statistical inference of unknown qualities in model (2) based on the samples {X~iΔn;i=1,2,} should be built, which are infinitesimal conditional expectations characterized by infinitesimal operators. Fortunately, under Lemma 1 in subsection 7.2 we can build the following infinitesimal conditional expectations for model (2)

(5)E[X~(i+1)ΔnX~iΔnΔn|F(i1)Δn]=μ(X(i1)Δn)+Op(Δn),
(6)E[(X~(i+1)ΔnX~iΔn)2Δn|F(i1)Δn]=23σ2(X(i1)Δn)+23Ec2(X(i1)Δn,z)f(z)dz+Op(Δn).

where Ft=σ{Xs,st}. One can refer to Appendix A in Song, Lin, and Wang (2013) for detailed calculations.

We briefly discuss the Gamma Nadaraya-Watson estimators for the coefficients in model (2) based on {X~iΔn;i=0,1,2,}. We firstly construct Gamma Nadaraya-Watson estimators for them based on equations (5) and (6). The Nadaraya-Watson estimators for μ(x) and M(x):=σ2(x)+Ec(x,z)f(z)dz based on the Gamma asymmetric kernel are solutions to the following optimal problem:

(7)argminμ(x)i=1n(X~(i+1)ΔnX~iΔnΔnμ(x))2KG(x/hn+1,hn)(X~(i1)Δn),
(8)argminM(x)i=1n(32(X~(i+1)ΔnX~iΔn)2ΔnM(x))2KG(x/hn+1,hn)(X~(i1)Δn),

where KG(x/hn+1,hn)() is a asymmetric Gamma kernel function and the smoothing parameter hn → 0 as n → ∞.

The solutions to (7) and (8) are

(9)a^n(x)=An(x)p^n(x),b^n(x)=Bn(x)p^n(x),

where p^n(x)=1ni=1nKG(x/hn+1,hn)(X~(i1)Δn), An(x)=1ni=1nKG(x/hn+1,hn)(X~(i1)Δn)X~(i+1)ΔnX~iΔnΔn, Bn(x)=1ni=1nKG(x/hn+1,hn)(X~(i1)Δn)32(X~(i+1)ΔnX~iΔn)2Δn.

According to the property of Gamma kernel function, we can know that the asymptotic variance of Gamma Nadaraya-Watson estimators for the unknown estimators varies with the design point x in the theorems. For convenience of describing the different asymptotic properties, denote by “interior x” and “boundary x” for a design point x that satisfies x/hn and x/hnκ for some κ > 0 as n,T, respectively.

We now present some assumptions used in the paper. In what follows, let D=(l,u) with l and u ≤ ∞ denote the admissible range of the process Xt, K denotes KG(x/hn+1,hn).

Assumption 1

  1. (Local Lipschitz continuity) For eachnN, there exist a constant Lnand a functionζn:ER+withEζn2(z)f(z)dz<such that, for any|x|n,|y|n,zE,

    |μ(x)μ(y)|+|σ(x)σ(y)|Ln|xy|,|c(x,z)c(y,z)|ζn(z)|xy|.
  2. (Linear growthness) For eachnN, there existζnas above and C, such that for allxR,zE,

    |μ(x)|+|σ(x)|C(1+|x|),|c(x,z)|ζn(z)(1+|x|).

Remark 1

Assumption 1 guarantees the existence and uniqueness of a solution to Xt in model (1) on the probability space (Ω,F,P), see Jacod and Shiryaev (2003).

Assumption 2

The process Xt is ergodic and stationary with a finite invariant measure ϕ(x). Furthermore, The process Xt is ρ-mixing with i1ρ(iΔn)=O(1Δnα),n, where α<12.

Remark 2

The finite invariant measure implies that the process Xt is positive Harris recurrent with the stationary probability measure p(x)=ϕ(x)ϕ(D),xD. The hypothesis that Xt is a stationary process is obviously a plausible assumption because for major integrated time series data, a simple differentiation generally assures stationarity. The same condition yielding information on the rate of decay of ρ-mixing coefficients for Xt was mentioned in Assumption 3 of Gugushvili and Spereij (2012).

Assumption 3

For any 2 ≤ i ≤ n, g is a differentiable function on and ξn,i=θX(i1)Δn+(1θ)X~(i1)Δn, 0θ1, the conditions hold:

  1. limhn0E[|hnK(ξn,i)g(X(i1)Δn)|]<,

  2. limhn0hn1/2E[|hn2K2(ξn,i)g(X(i1)Δn)|]<for “interior x”,

  3. limhn0hnE[|hn2K2(ξn,i)g(X(i1)Δn)|]<for“boundary x”.

Remark 3

According to the procedure for Assumption 3 in Appendix (7.1), we can easily deduce the following results:

  1. limhn0E[|K(X(i1)Δn)g(X(i1)Δn)|]<;

  2. limhn0hn1/2E[|K2(X(i1)Δn)g(X(i1)Δn)|]< for “interior x”;

  3. limhn0hnE[|K2(X(i1)Δn)g(X(i1)Δn)|]< for “boundary x”.

Assumption 4

For all p ≥ 1, supt0E[|Xt|p]<, and E|z|pf(z)dz<.

Remark 4

This assumption guarantees that Lemma 1 can be used properly throughout the article. If Xt is a Lévy process with bounded jumps (i.e. supt|ΔXt|C< almost surely, where C is a nonrandom constant), then E{|Xt|n}<n, that is, Xt has bounded moments of all orders, see Protter (2004). This condition is widely used in the estimation of an ergodic diffusion or jump-diffusion from discrete observations, see Florens-Zmirou (1989), Kessler (1997), and Shimizu and Yoshida (2006).

Assumption 5

Δn0, hn0, nΔnhnΔnlog(1Δn)0, hnnΔn1+α, as n.

Remark 5

The relationship between hn and Δn is similar as the stationary case in Hanif (2016), (b1) , (b2) of A8 in Nicolau (2007) and Assumption 7 in Song (2017). Wang and Zhou (2017) presented the optimal bandwidth of symmetric kernel nonparametric threshold estimator of diffusion function in jump-diffusion models. We will select the optimal smoothing parameter hn for Gamma asymmetric kernel estimation of second-order jump-diffusion models by means of minimizing the mean square error (MSE) in Remark 6.

3 Large sample properties.

Based on the above assumptions and the lemmas in the following proof procedure part, we have the following asymptotic properties. To simplify notations, we define xD to be a

interiorx''ifx/hn''orboundaryx''ifx/hnκ''
Theorem 1

If Assumption 1Assumption 5 hold, then

p^n(x):=1ni=1nKG(x/hn+1,hn)(X~(i1)Δn)pp(x).
Theorem 2

  1. Under theAssumption 1Assumption 5, we have

    a^n(x):=An(x)p^n(x)pμ(x).
  2. Under theAssumption 1Assumption 5, for “interior x”, ifhn=O((nΔn)2/5), then

    nΔnhn1/2(a^n(x)μ(x)hnBa^n(x))dN(0,M(x)2πx1/2p(x)),

    and for “boundary x”, ifhn=O((nΔn)1/3), then

    nΔnhn(a^n(x)μ(x)hnBa^n(x))dN(0,M(x)Γ(2κ+1)22κ+1Γ2(κ+1)p(x)),

    whereBa^n(x)denotes the bias of the estimator ofa^n(x), that is

    Ba^n(x)=μ(x){1+xp(x)p(x)}+x2μ(x).

Theorem 3

  1. Under the AssumptionsAssumption 1Assumption 5, we have

    b^n(x):=Bn(x)p^n(x)pσ2(x)+Ec2(x,z)f(z)dz.
  2. Under theAssumption 1Assumption 5, for “interior x”, ifhn=O((nΔn)2/5), then

    nΔnhn1/2(b^n(x)M(x)hnBb^n(x))dN(0,Ec4(x,z)f(z)dz2πx1/2p(x)),

    for “boundary x”, ifhn=O((nΔn)1/3), then

    nΔnhn(b^n(x)M(x)hnBb^n(x))dN(0,Ec4(x,z)f(z)dzΓ(2κ+1)22κ+1Γ2(κ+1)p(x)),

    whereBb^n(x)denotes the bias of the estimator ofb^n(x), that is

    Bb^n(x)=M(x){1+xp(x)p(x)}+x2M(x).

Remark 6

Theorem 2 and Theorem 3 give the weak consistency and the asymptotic normality of Nadaraya-Watson estimators for the unknown coefficients in model (2) using Gamma asymmetric kernels. In the practical applications, it is very important to consider the choice of the smoothing parameter hn for the nonparametric estimation using asymmetric kernels. Here we will select the optimal smoothing parameter hn based on the mean square error (MSE) according to Theorem 2 and Theorem 3. Take μ(x) for example, for “interior x”, the optimal smoothing parameter hn is

hn,opt=(1nΔnM(x)2πx1/2p(x)1Ba^n(x)2)25=Op(1nΔn)25,

for “boundary x”, the optimal smoothing parameter hn is

hn,opt=(1nΔnM(x)Γ(2κ+1)22κ+1Γ2(κ+1)p(x)1Ba^n(x)2)13=Op(1nΔn)13.

Note that the optimal bandwidth doesn’t coincide with that of the stationary case in Bandi and Nguyen (2003) nor that in Song (2017) in which the optimal smoothing parameters hn are the same for all design points and are of order Op(1nΔn)15.

Remark 7

The pointwise estimators considered here and in Chen and Zhang (2015), Song (2017), and Song, Lin, and Wang (2013) are based on kernel approach, whereas the adaptive estimator studied in Funke and Schmisser (2018) was done based on a model selection approach and focused on L2-risk. Moreover, Funke and Schmisser (2018) only focused on the regression-type estimator for the drift coefficient and could not provide the exact bias term or the central limit theorem. However, we and Chen and Zhang (2015), Song (2017), and Song, Lin, and Wang (2013) investigated nonparametric estimator of the drift function as well as the volatility function and proved the consistency and asymptotically normal distribution for the underlying estimators.

In our case of variance reduction for the nonparametric estimator, we only focus on the kernel-based estimators for the unknown functions such as drift and volatility, which can provide the central limit theorems. Therefore, in the following parts such as Remark 9 or the simulation section, we will focus on the theoretical and finite-sample numerical comparisons between the proposed approach and the existing methods mentioned in Song (2017), Chen and Zhang (2015), and Song, Lin, and Wang (2013), not Funke and Schmisser (2018).

Remark 8

Here we briefly describe the procedure for constructing the confidence intervals based on asymptotic normality for the further theoretical and numerical comparisons. Take μ(x) for example. The asymptotic normality of Nadaraya-Watson estimator using Gamma asymmetric kernel for μ(x) in this paper is different from that in Song (2017), the local linear estimator in Chen and Zhang (2015) and Song, Lin, and Wang (2013) constructed with Gaussian symmetric kernel. For brevity, the estimator proposed in this paper is denoted as AS, the estimator in Song (2017), Chen and Zhang (2015), and Song, Lin, and Wang (2013) as GS. The asymptotic normality for AS is shown in Theorem 2 and the asymptotic normality for GS was

hnnΔn(μ^nGS(x)μ(x))dN(0,M(x)2πp(x)),

where μ^nGS(x) denotes the GS estimator for μ(x).

For “interior x”, if the smoothing parameter hn=O((nΔn)2/5), the normal confidence interval for μ(x) using Gamma asymmetric kernel and Gaussian symmetric kernel at the significance level 100(1 − α)% are constructed as follows,

Iμ,αAS=[a^n(x)hnB^a^n(x)z1α/21nΔnhn1/2b^n(x)2πx1/2p^nAS(x),a^n(x)hnB^a^n(x)+z1α/21nΔnhn1/2b^n(x)2πx1/2p^nAS(x)],Iμ,αGS=[μ^nGS(x)z1α/21nΔnhnM^nGS(x)2πp^nGS(x),μ^nGS(x)+z1α/21nΔnhnM^nGS(x)2πp^nGS(x)],

where a^n(x), b^n(x) , μ^nGS(x), M^nGS(x) denote the estimators of μ(x), M(x) in (9) using Gamma asymmetric kernel or Gaussian symmetric kernel mentioned in Song (2017), Chen and Zhang (2015), and Song, Lin, and Wang (2013), respectively. z1α/2 is the inverse CDF for the standard normal distribution evaluated at 1 − α/2. Moreover, the consistent estimators for p(x) are p^nAS(x)=1ni=1nKGamma(x/hn+1,h)(X~(i1)Δn), p^nGS(x)=1nhni=1nKGaussian(xX~(i1)Δnhn). Furthermore, the estimators for the derivative μ(x), μ(x) and p(x) in B^a^n(x) of Iμ,αAS can be estimated by taking the derivative of the estimators of μ(x) in (9) and p^nAS(x) above using Gamma asymmetric kernel, one can refer to Fan and Gijbels (1996) for more similar details. For “boundary x”, one can construct the confidence intervals based on asymptotic normality similarly as the above procedure.

Similarly, the normal confidence interval for M(x) at a spatial point x using Gamma asymmetric kernel and Gaussian symmetric kernel at the significance level 100(1 − α)% can be constructed. The final interval for M(x) should be taken as the intersection of IM,αAS or IM,αGS with [0,+) to coincide with the nonnegativity of the conditional variance M(x).

Remark 9

There are two main differences: on one hand, the convergence rate of Nadaraya-Watson estimator using Gamma asymmetric kernel is different for the location of the design point x such as “interior x” and “boundary x”; on the other hand, the variance of Nadaraya-Watson estimator using Gamma asymmetric kernel is inversely proportional to the design x, which shows that the variance decreases as the design point x increases.

Here we briefly commented the theoretical comparison for the convergence rate and the variance of Nadaraya-Watson estimator for “interior x” and “boundary x” using Gamma asymmetric kernel or Gaussian symmetric kernel, which dominate the length of the confidence interval. For “interior x”, the convergence rate of Nadaraya-Watson estimator based on Gamma asymmetric kernel is 1nΔnhn1/2, which is much smaller than 1nΔnhn of that based on Gaussian symmetric kernel with a given hn. Additionally, compared with the ones using Gaussian symmetric kernel, the variance of local constant estimator using Gamma asymmetric kernel is inversely proportional to the design x, which shows that the length of the confidence interval decreases as the design point x increases.

For “boundary x”, although the convergence rate of Nadaraya-Watson estimator based on Gamma asymmetric kernel is the same as that based on Gaussian symmetric kernel, the coefficient in their variance differs a little such as Γ(2κ+1)22κ+1Γ2(κ+1) for Gamma asymmetric kernel and 12π for Gaussian symmetric kernel. Under numeral calculations, from Table 1 we can conclude that when κ ≤ 0.7, the variance based on Gamma asymmetric kernel is larger, whereas when κ ≥ 0.75, the variance based on Gamma asymmetric kernel is smaller than that based on Gaussian symmetric kernel. This reveals that the closer to the origin, the shorter length of confidence interval based on Gaussian symmetric kernel, whereas away from the origin such as the sparse design boundary points, the shorter length of confidence interval based on Gamma asymmetric kernel.

Table 1:

The difference between Γ(2κ+1)22κ+1Γ2(κ+1) and 12π in the variance for various κ.

Value of κ0.250.30.350.40.450.50.550.6
Difference0.09930.08380.07010.05770.04640.03620.02690.0183
Value of κ0.650.70.750.80.850.90.951
Difference0.01030.003−0.004−0.01−0.016−0.022−0.027−0.032
Value of κ1.251.51.7522.252.52.753
Difference−0.053−0.07−0.083−0.095−0.104−0.112−0.12−0.126
Value of κ3.253.53.7544.254.54.755
Difference−0.132−0.137−0.141−0.145−0.149−0.153−0.156−0.159
  1. Note that the difference for a κ means the value of Γ(2κ+1)22κ+1Γ2(κ+1) minus 12π .

Remark 10

In contrary to the second-order diffusion model without jumps in Hanif (2015), the second infinitesimal moment estimator for second-order jump-diffusion model has a rate of convergence that is the same as the rate of convergence of the first infinitesimal moment estimator. Apparently, this is due to the presence of discontinuous breaks that have an equal impact on all the functional estimates. As Johannes (2004) pointed out, for the conditional variance of interest rate changes, not only diffusion play a certain role, but also jumps account for more than half at lower interest level rates, almost two-thirds at higher interest level rates, which dominate the conditional volatility of interest rate changes. Thus, it is extremely important to estimate the conditional variance as σ2(x)+Ec2(x,z)f(z)dz which reflects the fluctuation of the underlying asset.

4 Monte carlo simulation study

In this section, compared with the existing kernel methods mentioned in Song (2017), Chen and Zhang (2015), and Song, Lin, and Wang (2013) using Gaussian kernels, the benefits of Nadaraya-Watson estimators for the drift function μ(x) and the volatility function M(x) using the asymmetric kernels are demonstrated through the finite-sample performance for various second-order jump-diffusion models. For brevity, the estimator proposed in this paper is denoted as AS, the estimator in Song (2017) as NW, the estimator in Chen and Zhang (2015) as LL and the estimator in Song, Lin, and Wang (2013) as RNW.

Throughout this section, we employ Gamma kernel KG(x/hn+1,hn)(u)=ux/hnexp(u/hn)hnx/hn+1Γ(x/hn+1) and Gaussian kernel K(x)=12πex22. The common bandwidths for the estimators investigated in Song (2017), Chen and Zhang (2015), and Song, Lin, and Wang (2013) using Gaussian kernels are selected as h=cS^(nΔn)15=cS^T15, where Ŝ denotes the standard deviation of the data and c represents different constants for different estimators with c = 2.8 for μ^n(x) and c = 1.3 for M^n(x), one can refer to part (4.2) in Xu and Phillips (2011) for more details. In addition, the normal confidence level is assumed to be 95%.

Example 1: Gaussian jumps.

In this subsection, a simple simulation experiment is considered for the model (10) aimed at evaluating the better finite-sampling performance of nonparametric estimators (9) constructed with Gamma asymmetric kernels

(10){dYt=Xtdt,dXt=Xtdt+0.01+0.01Xt2dWt+dJt,

where the coefficients of continuous part are similar as the ones used in Nicolau (2007) and Jt is a compound Poisson jump process, that is, Jt=n=1NtZtn with arrival intensity λT=20 and jump size ZnN(0,0.0362) corresponding to Bandi and Nguyen (2003), where tn is the nth jump of the Poisson process Nt. Taking the integral from 0 to t in the second expression of (10), we obtain

(11)Xt=0tXsds+0t0.01+0.01Xs2dWt+n=1NtZtn.

By (11) we have

(12)Yt=0tXsds=(Xt0t0.01+0.01Xs2dWtn=1NtZtn).

Xt can be sampled by the Euler-Maruyama scheme according to (11), which will be detailed in the following algorithm (one can refer to Cont and Tankov (2004)). One sample trajectory of the differentiated process Xt and integrated process Yt with T = 10, n = 1000, X0 = 0 and Y0 = 100 using Algorithm 1 is shown in Figure 1. Through the observation of Figure 1(B), we can find the features of the integrated process Yt: absent mean-reversion, persistent shocks, time-dependent mean and variance, nonnormality, etc.

Figure 1: Sample Paths of Xt and Yt for Model (10).(A) The differentiated process Xt. (B)The integrated process Yt.
Figure 1:

Sample Paths of Xt and Yt for Model (10).

(A) The differentiated process Xt. (B)The integrated process Yt.

Algorithm 1

Algorithm 1 (Simulation for trajectories of second-order jump-diffusion model).

Procedures:

  1. generate a standard normal random variate V and transform it into Di=0.01+0.01Xti12×Δti×V, where Δti=titi1=Tn is the observation time frequency;

  2. generate a Poisson random variate N with intensity λT=20;

  3. generate N random variables τi uniformly distributed in [0, T], which correspond the jump times;

  4. generate N random variables ZτiN(0,0.0362), which correspond the jump sizes;

    One trajectory for Xt is

    Xti=Xti1Xti1×Δti+Di+1{ti1τi<ti}×Zτi.
  5. By substitution of Xti in (12), Yti can be sampled.

Here we will select the optimal smoothing parameter hnAS for Nadaraya-Watson estimation using asymmetric kernels based on the mean square error (MSE) according to Theorem 2 and Theorem 3 as shown in Remark 6. Similarly as the common bandwidths selected for estimators constructed with Gaussian kernels, the data-driven bandwidth is represented as hnAS=cS^T25 for “interior x” or h=cS^T13 for “boundary x”, where c is an unknown parameter to be selected. Curve of MSE(c) for the unknown coefficients in model (10) based on Gamma kernels with λ = 2, T = 10, n = 1000 and jump size ZnN(0,0.0362) is depicted in Figure 2, from which we can select the optimal parameter copt = 1.44 for drift function and copt = 1.11 for volatility function.

Figure 2: Curve of MSE(c) for the unknown coefficients in model (10) based on Gamma kernels with λ = 2, T = 10, n = 1000 and jump size $Z_{n}\sim\mathscr{N}(0,0.036^{2})$Zn∼N(0,0.0362).(A) Curve of MSE(c) for drift function. (B) Curve of MSE(c) for volatility function.
Figure 2:

Curve of MSE(c) for the unknown coefficients in model (10) based on Gamma kernels with λ = 2, T = 10, n = 1000 and jump size ZnN(0,0.0362).

(A) Curve of MSE(c) for drift function. (B) Curve of MSE(c) for volatility function.

The Nadaraya-Watson estimators for drift function μ(x) = −x and volatility function M(x)=0.01+0.01×x2+2×0.0362 in model (10) constructed with Gaussian and Gamma kernels from a sample with λ = 2, T = 10, n = 1000 and jump size ZnN(0,0.0362) are shown in Figure 3. We can observe that AS performs better than the others such as NW, LL and RNW, especially at the boundary points, which is also verified through the biases calculated at various points of the sample Xt, in Table 2. Additionally, Figure 4 represents 95% Monte Carlo confidence intervals for these estimators of the unknown coefficients based on Remark 8, and Table 3 shows the length of confidence bands of various estimators for drift and volatility functions. They show that the lengths of 95% Monte Carlo confidence intervals for μ(x) constructed with Gamma asymmetric kernels are almost shorter than those constructed with Gaussian symmetric kernels and those for M(x) using Gamma asymmetric kernels are shorter than the others with Gaussian symmetric kernels especially at the sparse design boundary point, which coincides with the conclusion in Remark 9 and reveals variance reduction. In addition, at the sparse design boundary point, the true value of μ(x) or M(x) cannot fall in the 95% Monte Carlo confidence bands of some nonparametric estimators using Gaussian symmetric kernels, but the 95% Monte Carlo confidence bands constructed with Nadaraya-Watson estimator with Gamma asymmetric kernel.

Figure 3: Various Estimators for unknown coefficients in model (10) with λ = 2, T = 10, n = 1000 and jump size $Z_{n}\sim\mathscr{N}(0,0.036^{2})$Zn∼N(0,0.0362).(A) Estimators for drift function μ(x) = −x. (B) Estimators for volatility function $M(x)=0.01+0.01\times x^{2}+2\times 0.036^{2}$M(x)=0.01+0.01×x2+2×0.0362.
Figure 3:

Various Estimators for unknown coefficients in model (10) with λ = 2, T = 10, n = 1000 and jump size ZnN(0,0.0362).

(A) Estimators for drift function μ(x) = −x. (B) Estimators for volatility function M(x)=0.01+0.01×x2+2×0.0362.

Figure 4: 95% Monte Carlo confidence intervals for unknown coefficients based on various estimators with λ = 2, T = 10, n = 1000 and jump size $Z_{n}\sim\mathscr{N}(0,0.036^{2})$Zn∼N(0,0.0362).(A) Confidence bands for drift function μ(x) = −x. (B) Confidence bands for volatility function $M(x)=0.01+0.01\times x^{2}+2\times 0.036^{2}$M(x)=0.01+0.01×x2+2×0.0362.
Figure 4:

95% Monte Carlo confidence intervals for unknown coefficients based on various estimators with λ = 2, T = 10, n = 1000 and jump size ZnN(0,0.0362).

(A) Confidence bands for drift function μ(x) = −x. (B) Confidence bands for volatility function M(x)=0.01+0.01×x2+2×0.0362.

Table 2:

The Biases of various Estimators for μ(x) = −x and M(x)=0.01+0.01×x2+2×0.0362 at various points of sample Xt with T = 10, n = 1000, λ = 2 and jump size ZnN(0,0.0362).

Bias(×10−2)Various points of Sample Xt
0.010.020.030.040.050.060.070.080.090.1
μ(x)NW1.18371.92522.66253.39554.12434.84895.56936.28576.99797.7062
LL−0.2409−0.5381−0.9528−1.4934−2.1687−2.9872−3.9576−5.0882−6.3875−7.8635
RNW1.14500.95740.75120.0650−0.6313−1.5302−2.4320−3.3810−3.8591−4.2691
AS0.06350.22630.26110.21180.1113−0.0170−0.1563−0.2954−0.4265−0.5447
M(x)NW−0.0688−0.0617−0.0544−0.0473−0.0403−0.0338−0.0278−0.0224−0.0177−0.0138
LL−0.0655−0.0521−0.0386−0.0256−0.0137−0.00330.00550.01250.01750.0204
RNW−0.0511−0.0295−0.00490.01100.02130.02200.02370.02210.01920.0179
AS0.01990.0066−0.0034−0.0097−0.0126−0.0128−0.0112−0.0087−0.0058−0.0033
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

We will assess the small-sampling performance of Nadaraya-Watson estimators constructed with Gamma asymmetric kernels and those constructed with Gaussian symmetric kernels, for drift function μ(x) and M(x) via the Mean Square Errors (MSE)

Table 3:

The Length of Confidence Bands of various Estimators for μ(x) = −x and M(x)=0.01+0.01×x2+2×0.0362 at various points of sample Xt with T = 10, n = 1000, λ = 2 and jump size ZnN(0,0.0362).

Length of Confidence BandsVarious points of Sample Xt
0.010.020.030.040.050.060.070.080.090.1
μ(x)NW0.11620.11710.11850.12020.12240.12500.12800.13150.13560.1402
LL0.11610.11750.11970.12270.12670.13190.13820.14580.15470.1651
RNW0.11950.11900.12020.11910.11980.12320.12710.12820.13390.1323
AS0.10940.10890.10970.11170.11470.11860.12360.12940.13620.1439
M(x)(×103)NW2.31632.37162.45572.56982.71592.89653.11483.37483.68174.0425
LL2.32892.41082.52492.66992.84493.04943.28383.54903.84774.1838
RNW2.33752.34942.49652.62262.72953.03213.47693.86913.91164.1262
AS5.52455.05404.74514.52184.34984.21084.09433.99333.90343.8216
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

  2. The confidence intervals of various estimators are calculated as those mentioned in Remark 8.

(13)MSE=1mk=0m{μ^(xk)μ(xk)}2

and

(14)MSE=1mk=0m{M^(xk)M(xk)}2,

where μ^(x) or M^(x) are the various estimators of μ(x) or M(x), respectively and {xk}1m are chosen uniformly to cover the range of sample path of Xt.

The mean square error (MSE) of various estimators for the drift function μ(x) or volatility function M(x) are calculated under various lengths of observation time intervals T (= 50, 100, 500) and sample sizes n (= 500, 1000, 5000) with Δn=Tn are listed in Table 4 and Table 7. Table 5 gives the results on these MSE under different jump sizes Zn and sample sizes n (= 500, 1000, 5000). Table 6 and Table 8 show these values of MSE under different jump arrival intensity λ and sample sizes T (= 10, 50, 100) over 500 replicates.

Table 4:

MSE (×10−3) of various estimators for three lengths of time interval (T) and three sample sizes for μ(x) = −x with arrival intensity λ = 2 and jump size ZnN(0,0.0362) over 500 replicates.

Time SpanEstimatorsn = 500n = 1000n = 5000
T = 10NW2.56301.60071.4672
LL1.35610.74640.1899
RNW3.96621.30930.6185
AS0.63280.11360.1121
T = 50NW2.23592.13951.5274
LL1.36820.84600.5652
RNW2.01402.15061.1965
AS0.46830.31370.1577
T = 100NW2.62352.17861.9483
LL1.82121.04490.8200
RNW2.57722.20981.6114
AS0.80700.58410.5215
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

Table 5:

MSE (×10−3) of various estimators for three types of jump size Zn and three sample sizes for μ(x) = −x with arrival intensity λ = 2 and T = 10 over 500 replicates.

Jump SizeEstimatorsn = 500n = 1000n = 5000
ZnN(0,0.0362)NW2.56301.60071.4672
LL1.35610.74640.1899
RNW3.96621.30930.6185
AS0.63280.11360.1121
ZnN(0,1)NW11.11933.53692.3887
LL7.29063.42951.8619
RNW12.31864.12762.0337
AS6.80881.67640.6840
ZnCauchy(0,1)NW36.256731.17467.2090
LL13.926710.81755.4620
RNW10.36665.23468.3132
AS5.39194.80912.8424
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

Table 6:

MSE (×10−3) of various estimators for three types of jump arrival intensity λ and three sample sizes for μ(x) = −x with T = 10 and jump size ZnN(0,0.0362) over 500 replicates.

Jump IntensityEstimatorsn = 500n = 1000n = 5000
λ = 1NW2.21851.41980.9437
LL1.06160.73670.4891
RNW3.61261.42940.5364
AS0.44070.04370.2213
λ = 2NW2.56301.60071.4672
LL1.35610.74640.1899
RNW3.96621.30930.6185
AS0.63280.11360.1121
λ = 5NW4.42354.08841.9067
LL3.90301.80350.6593
RNW5.01941.64320.7351
AS2.57041.66250.3750
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

Table 7:

MSE (×10−5) of various estimators for three lengths of time interval (T) and three sample sizes for M(x)=0.01+0.01×x2+2×0.0362 with arrival intensity λ = 2 and jump size ZnN(0,0.0362) over 500 replicates.

Time SpanEstimatorsn = 500n = 1000n = 5000
T = 10NW0.41900.29850.1958
LL0.36820.26100.1798
RNW1.32950.23120.1792
AS0.26080.22540.1705
T = 50NW0.42760.37540.3099
LL0.37970.32340.2579
RNW1.15750.36540.2874
AS0.35540.29220.1997
T = 100NW0.45440.42310.3303
LL0.44330.38030.3174
RNW1.56750.36010.2911
AS0.40470.36440.2756
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

Table 8:

MSE (×10−5) of various estimators for three types of jump arrival intensity λ and three sample sizes for M(x)=0.01+0.01×x2+λ×0.0362 with T = 10 and jump size ZnN(0,0.0362) over 500 replicates.

Jump IntensityEstimatorsn = 500n = 1000n = 5000
λ = 1NW0.17530.13480.0461
LL0.16270.12230.0323
RNW0.21790.10370.0385
AS0.16180.08890.0134
λ = 2NW0.41900.29850.1958
LL0.36820.26100.1798
RNW1.32950.23120.1792
AS0.26080.22540.1705
λ = 5NW2.28101.72320.8455
LL1.72501.28910.5958
RNW2.03311.09470.6711
AS1.53731.25340.5889
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

From Table 4Table 6 or Table 7Table 8, we can get the following findings.

  • The Nadaraya-Watson estimator for μ(x) or M(x) constructed with Gamma asymmetric kernels performs a little better than that constructed with Gaussian symmetric kernels in terms of MSE under different time spans, different sample sizes, different jump sizes and different arrival intensity. This is mainly due to the fact that the smaller coverage rate for “interior x” and the smaller asymptotic variance for “boundary x”;

  • From Table 4 and Table 7, for the same time interval T, as the sample sizes n tends larger, the performances of the estimators constructed with Gamma asymmetric kernels or Gaussian symmetric kernels for μ(x) or M(x) are improved due to the fact that more information for estimation procedure is sampled as Δn → 0. However, for the same sample sizes n, as the time interval T expands larger, the performance of the estimators for μ(x) or M(x) gets worse due to the fact that more jumps happens in larger time interval T in steps 3 of Algorithm 1;

  • From Table 5, Table 6 or Table 8, for the same jump size or the same jump arrival intensity, as the sample sizes n tends larger, the performances of the estimators for μ(x) or M(x) are also improved due to the fact that more jump information for estimation procedure is collected as Δn → 0. However, for the same sample sizes n, as the amplitude or frequency of jump becomes larger, the MSE of the estimators gradually becomes larger because of larger and more jumps which is more likely to produce outliers;

  • To some extent, the previous remark confirms that the drift and volatility functions cannot be identified in a fixed time span, which corresponds to the results of Theorem 2 and Theorem 3.

Figure 5 and Figure 6 give the QQ plots of Nadaraya-Watson estimators for the drift function μ(x) and conditional variance function M(x) constructed with Gamma asymmetric kernels at boundary and interior points, that is 10%, 50% and 90% quantile points of Xt, with λ = 2, T = 10, n = 1000 and jump size ZnN(0,0.0362). This reveals the normality of the estimators of the drift function μ(x) and conditional variance function M(x) constructed with Gamma asymmetric kernels, which confirms the results in Theorem 2 and Theorem 3.

Figure 5: QQ plots of Nadaraya-Watson estimator for μ(x) = −x using Gamma kernels at boundary and interior points, that is 10%, 50% and 90% quantile points of Xt, with λ = 2, T = 10, n = 1000 and jump size $Z_{n}\sim\mathscr{N}(0,0.036^{2})$Zn∼N(0,0.0362).(A) QQ plot for left point. (B) QQ plot for median point. (C) QQ plot for right point.
Figure 5:

QQ plots of Nadaraya-Watson estimator for μ(x) = −x using Gamma kernels at boundary and interior points, that is 10%, 50% and 90% quantile points of Xt, with λ = 2, T = 10, n = 1000 and jump size ZnN(0,0.0362).

(A) QQ plot for left point. (B) QQ plot for median point. (C) QQ plot for right point.

Figure 6: QQ plots of Nadaraya-Watson estimator for $M(x)=0.01+0.01\times x^{2}+2\times 0.036^{2}$M(x)=0.01+0.01×x2+2×0.0362 using Gamma kernels at boundary and interior points, that is 10%, 50% and 90% quantile points of Xt, with $\lambda=2, \ T=10, \ n=1000$λ=2, T=10, n=1000 and jump size $Z_{n}\sim\mathscr{N}(0,0.036^{2})$Zn∼N(0,0.0362).(A) QQ plot for left point. (B) QQ plot for median point. (C) QQ plot for right point.
Figure 6:

QQ plots of Nadaraya-Watson estimator for M(x)=0.01+0.01×x2+2×0.0362 using Gamma kernels at boundary and interior points, that is 10%, 50% and 90% quantile points of Xt, with λ=2,T=10,n=1000 and jump size ZnN(0,0.0362).

(A) QQ plot for left point. (B) QQ plot for median point. (C) QQ plot for right point.

Example 2: Mixed Gaussian jumps (The Variance Gamma model).

In this subsection, a repeated experiment is conducted for another model (15) aimed at evaluating the better finite-sampling performance of nonparametric estimators (9) constructed with Gamma asymmetric kernels

(15){dYt=Xtdt,dXt=Xtdt+0.01+0.01Xt2dWt+dJt,

where the jump size in Jt is assumed ZnN(0,σJ2V) and VG(ν)=1/Γ(1/b)(1/b)((1/b)ν)1/b1e(1/b)νI{ν0}. Γ(⋅) is the gamma function, b = 0.023 is the variance of V and σJ = 0.02 is the scale parameter. One can refer to Madan and Seneta (1990) and Bandi and Nguyen (2003) for more details about this variance Gamma model.

For this case, based on the sample Xt generated by the model (15), compared with nonparametric kernel estimators constructed with Gaussian symmetric kernels, we can also observe the better finite-sample properties such as smaller bias and shorter length of confidence bands, especially at the sparse design boundary point, for Nadaraya-Watson estimators of drift and volatility coefficients using Gamma asymmetric kernel from Figure 7Figure 11 and Table 9Table 14, which confirms the bias corrections and variance reduction shown in Theorem 2 and Theorem 3.

Figure 7: Sample Paths of Xt and Yt for Model (15).(A) The differentiated process Xt. (B) The integrated process Yt.
Figure 7:

Sample Paths of Xt and Yt for Model (15).

(A) The differentiated process Xt. (B) The integrated process Yt.

Figure 8: Various Estimators for unknown coefficients in model (15) with λ = 2, T = 10, n = 1000 and jump size $Z_{n}\sim\mathscr{N}(0,\sigma_{J}^{2}V)$Zn∼N(0,σJ2V).(A) Estimators for drift function μ(x) = −x. (B) Estimators for volatility function $M(x)=0.01+0.01\times x^{2}+0.02^{2}$M(x)=0.01+0.01×x2+0.022.
Figure 8:

Various Estimators for unknown coefficients in model (15) with λ = 2, T = 10, n = 1000 and jump size ZnN(0,σJ2V).

(A) Estimators for drift function μ(x) = −x. (B) Estimators for volatility function M(x)=0.01+0.01×x2+0.022.

Figure 9: 95% Monte Carlo confidence intervals for unknown functions based on various estimators with λ = 2, T = 10, n = 1000 and jump size $Z_{n}\sim\mathscr{N}(0,\sigma_{J}^{2}V)$Zn∼N(0,σJ2V).(A) Confidence bands for drift function μ(x) = −x. (B) Confidence bands for volatility function $M(x)=0.01+0.01\times x^{2}+0.02^{2}$M(x)=0.01+0.01×x2+0.022.
Figure 9:

95% Monte Carlo confidence intervals for unknown functions based on various estimators with λ = 2, T = 10, n = 1000 and jump size ZnN(0,σJ2V).

(A) Confidence bands for drift function μ(x) = −x. (B) Confidence bands for volatility function M(x)=0.01+0.01×x2+0.022.

Figure 10: QQ plots of Nadaraya-Watson estimator for μ(x) = −x using Gamma kernels at boundary and interior points, that is 10%, 50% and 90% quantile points of Xt, with λ = 2, T = 10, n = 1000 and jump size $Z_{n}\sim\mathscr{N}(0,\sigma_{J}^{2}V)$Zn∼N(0,σJ2V).(A) QQ plot for left point. (B) QQ plot for median point. (C) QQ plot for right point.
Figure 10:

QQ plots of Nadaraya-Watson estimator for μ(x) = −x using Gamma kernels at boundary and interior points, that is 10%, 50% and 90% quantile points of Xt, with λ = 2, T = 10, n = 1000 and jump size ZnN(0,σJ2V).

(A) QQ plot for left point. (B) QQ plot for median point. (C) QQ plot for right point.

Figure 11: QQ plots of Nadaraya-Watson estimator for $M(x)=0.01+0.01\times x^{2}+0.02^{2}$M(x)=0.01+0.01×x2+0.022 using Gamma kernels at boundary and interior points, that is 10%, 50% and 90% quantile points of Xt, with λ = 2, T = 10, n = 1000 and jump size $Z_{n}\sim\mathscr{N}(0,\sigma_{J}^{2}V)$Zn∼N(0,σJ2V).(A) QQ plot for median point. (B) QQ plot for left point. (C) QQ plot for right point.
Figure 11:

QQ plots of Nadaraya-Watson estimator for M(x)=0.01+0.01×x2+0.022 using Gamma kernels at boundary and interior points, that is 10%, 50% and 90% quantile points of Xt, with λ = 2, T = 10, n = 1000 and jump size ZnN(0,σJ2V).

(A) QQ plot for median point. (B) QQ plot for left point. (C) QQ plot for right point.

Table 9:

The Biases of various Estimators for μ(x) = −x and M(x)=0.01+0.01×x2+0.022 at various points of sample Xt with T = 10, n = 1000, λ = 2 and jump size ZnN(0,σJ2V).

Bias(×10−2)Various points of Sample Xt
0.010.020.030.040.050.060.070.080.090.1
μ(x)NW0.21280.91301.61312.31283.01193.71024.40745.10335.79786.4908
LL−0.7333−0.8928−1.0771−1.2929−1.5465−1.8444−2.1929−2.5979−3.0654−3.6009
RNW0.66241.40981.72772.18512.42552.79182.99013.92635.00195.2460
AS−0.0418−0.1094−0.07970.02030.17460.37250.60720.87361.16831.4886
M(x)NW−0.0978−0.0916−0.0852−0.0786−0.0718−0.0649−0.0578−0.0508−0.0438−0.0370
LL−0.0730−0.0587−0.0442−0.0297−0.0154−0.00150.01180.02440.03650.0481
RNW−0.0832−0.0739−0.0631−0.0576−0.0534−0.0503−0.0449−0.0362−0.0254−0.0154
AS−0.0701−0.0600−0.0443−0.0270−0.01070.00310.01380.02150.02660.0294
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

Table 10:

The Length of Confidence Bands of various Estimators for μ(x) = −x and M(x)=0.01+0.01×x2+0.022 at various points of sample Xt with T = 10, n = 1000, λ = 2 and jump size ZnN(0,σJ2V).

Length of Confidence BandsVarious points of Sample Xt
0.010.020.030.040.050.060.070.080.090.1
μ(x)NW0.11420.11490.11590.11740.11920.12130.12380.12670.13000.1336
LL0.11420.11490.11610.11770.11950.12170.12410.12660.12930.1321
RNW0.11400.11400.11490.11500.12030.12050.12380.12530.12840.1324
AS0.11280.11050.10980.11020.11160.11370.11650.11990.12390.1285
M(x)(×103)NW2.07702.14932.24112.35422.49082.65342.84523.06943.33043.6329
LL2.13372.23162.35342.50102.67652.88243.12183.39833.71664.0824
RNW2.27592.34402.34642.32632.48482.55202.66782.77842.89822.9283
AS3.21122.98962.85352.76432.70112.65252.61192.57582.54252.5109
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

Table 11:

MSE (×10−3) of various estimators for three lengths of time interval (T) and three sample sizes for μ(x) = −x with arrival intensity λ = 2 and jump size ZnN(0,σJ2V) over 500 replicates.

Time SpanEstimatorsn = 500n = 1000n = 5000
T = 10NW1.95321.72121.4911
LL0.78930.71460.2343
RNW0.91140.13650.0952
AS0.43110.29880.0793
T = 50NW2.08672.10550.7967
LL0.88190.92520.6314
RNW1.36781.77680.3483
AS0.68150.52610.3247
T = 100NW2.19322.10951.8885
LL1.01320.98740.7237
RNW1.97351.29761.4325
AS0.86970.62060.4619
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

Table 12:

MSE (×10−3) of various estimators for three types of jump arrival intensity λ and three sample sizes for μ(x) = −x with T = 10 and jump size ZnN(0,σJ2V) over 500 replicates.

Jump IntensityEstimatorsn = 500n = 1000n = 5000
λ = 1NW2.75921.46681.6902
LL1.77040.65050.4611
RNW1.75531.03361.4220
AS0.97620.08230.0451
λ = 2NW1.95321.72121.4911
LL0.78930.71460.2343
RNW0.91140.13650.0952
AS0.43110.29880.0793
λ = 5NW4.56593.56922.7208
LL2.83681.07920.6258
RNW4.36182.76491.5333
AS2.01470.53930.3959
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

Table 13:

MSE (×10−6) of various estimators for three lengths of time interval (T) and three sample sizes for M(x)=0.01+0.01×x2+0.022 with arrival intensity λ = 2 and jump size ZnN(0,σJ2V) over 500 replicates.

Time SpanEstimatorsn = 500n = 1000n = 5000
T = 10NW1.45210.49210.4243
LL1.29560.40340.3161
RNW1.40990.32270.1765
AS0.65170.25130.2352
T = 50NW2.11770.79300.4396
LL1.55500.72690.3591
RNW1.43160.58310.5036
AS1.12660.54640.2685
T = 100NW3.10732.26890.5252
LL3.01421.73430.4222
RNW3.53331.65950.4016
AS2.35651.43500.3709
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

Table 14:

MSE (×10−6) of various estimators for three types of jump arrival intensity λ and three sample sizes for M(x)=0.01+0.01×x2+0.022 with T = 10 and jump size ZnN(0,σJ2V) over 500 replicates.

Jump IntensityEstimatorsn = 500n = 1000n = 5000
λ = 1NW0.47170.23110.1088
LL0.30430.15070.0887
RNW0.31860.23220.0536
AS0.24880.10070.0126
λ = 2NW1.45210.49210.4243
LL1.29560.40340.3161
RNW1.40990.32270.1765
AS0.65170.25130.2352
λ = 5NW2.63121.79560.5335
LL2.39271.52270.4344
RNW2.12251.15850.4085
AS2.07450.81510.4297
  1. Note that NW denotes Nadaraya-Watson estimator, LL denotes local linear estimator, RNW denotes reweighted Nadaraya-Watson estimator and AS denotes the estimator (9) proposed in this paper.

5 Empirical analysis

In this section, we apply the second-order jump-diffusion to model the return of stock index between Jul 2014 and Dec 2014 from Shenzhen Stock Exchange in China under 5-minute high-frequency data, and then apply the Nadaraya–Watson estimators to estimate the unknown coefficients in model (2) based on Gamma asymmetric kernels and Gaussian symmetric kernels. One can refer to Supplementary Material for the empirical data and codes.

We assume that

(16){dlogYt=Xtdt,dXt=μ(Xt)dt+σ(Xt)dWt+Ec(Xt,z)r(ω,dt,dz),

where log Yt is the log integrated process for stock index or commodity price and Xt is the latent process for the log-returns. According to (4), we can get the proxy of the latent process

(17)X~iΔn=logYiΔnlogY(i1)ΔnΔn.

The plots of the stock index and its proxy (17) of Shenzhen composite index in 5-minute high frequency data are shown in Figure 12.

Figure 12: Time Series and Proxy of Shenzhen Composite Index (2014) from July 03, 2014 to Dec 31, 2014.(A) Shenzhen Composite Index (2014). (B) Proxy of Shenzhen Composite Index (2014).
Figure 12:

Time Series and Proxy of Shenzhen Composite Index (2014) from July 03, 2014 to Dec 31, 2014.

(A) Shenzhen Composite Index (2014). (B) Proxy of Shenzhen Composite Index (2014).

First, based on the Augmented Dickey-Fuller test statistic, we can easily get that the null hypothesis of non-stationarity is accepted at the 5% significance level for the stock index Yt, but is rejected for the proxy of Xt, which confirms the assumption of stationarity by differencing. Then, we test the existence of jumps for the proxy Xt through the test statistic proposed in Barndorff-Nielsen and Shephard (2006) (denoted by BS Statistic). For the 5-minute high frequency data between Jul 2014 and Dec 2014 from Shenzhen Stock Exchange, the value of BS Statistic is −4.5961, which exceeds [−1.96, 1.96], so there exists jumps in high frequency data at the 5% significance level, which confirms the validity of model (2) for the return of stock index.

Finally, we will construct 95% normal confidence intervals for the unknown coefficients based on Gamma asymmetric kernels and Gaussian symmetric kernels under (17) and Δn=148 for 5-minute data (t = 1 meaning 1 day). The 95% normal confidence bands for the drift and volatility functions are demonstrated in Figure 13. All the quantities are computed at 120 equally spaced nonnegative ordered X~iΔn from 0.01 to 0.601. For a more intuitive comparison of the lengths of normal confidence intervals of the drift and volatility coefficients based on Gaussian kernels and Gamma kernels for Shenzhen Composite Index (2014), here the ratios of the length of confidence bands constructed with Gaussian symmetric kernel (CB-GSK) to that constructed with Gamma asymmetric kernel (CB-GAK) are shown in Figure 14. Note that the blue dotted lines in Figure 8 represent the ratio value of one.

Figure 13: 95% normal confidence intervals of the drift and volatility coefficients for Shenzhen Composite Index (2014) based on Gamma kernels and Gaussian kernels.Note that Nadaraya-Watson is denoted as NW in this figure. (A) Drift Confidence Band. (B) Volatility Confidence Band.
Figure 13:

95% normal confidence intervals of the drift and volatility coefficients for Shenzhen Composite Index (2014) based on Gamma kernels and Gaussian kernels.

Note that Nadaraya-Watson is denoted as NW in this figure. (A) Drift Confidence Band. (B) Volatility Confidence Band.

Figure 14: The ratios between the lengths of normal confidence intervals of the drift and volatility coefficients based on Gaussian kernels and Gamma kernels for Shenzhen Composite Index (2014).(A) Length Ratios for Drift Estimators. (B) Length Ratios for Volatility Estimators.
Figure 14:

The ratios between the lengths of normal confidence intervals of the drift and volatility coefficients based on Gaussian kernels and Gamma kernels for Shenzhen Composite Index (2014).

(A) Length Ratios for Drift Estimators. (B) Length Ratios for Volatility Estimators.

From Figure 13 and Figure 14, we can observe the following findings.

  1. As for the quantities close to zero, the ratios are less than one, which coincides with the discussion in Remark 9 that the closer to the boundary point or the larger bandwidth for fixed “boundary x”, the shorter the length of confidence interval based on Gaussian symmetric kernel.

  2. As the points increase, especially at the sparse points, CB-GSK tends to be longer than CB-GAK and the ratios gradually become larger and greater than 1 (excluding a small amount of points around 0.5 for volatility estimator), which effectively verifies the properties of efficiency gains and resistance to sparse points of Nadaraya-Watson smoothing using Gamma asymmetric kernel through the real high frequency financial data.

6 Conclusion

In this paper, Nadaraya-Watson estimators constructed with Gamma asymmetric kernels for the unknown coefficients in second-order jump-diffusion model are investigated. Compared with estimation constructed with Gaussian symmetric kernels, the local constant smoothing using Gamma asymmetric kernels possesses some attractive advantages such as boundary free bias, variance reduction and resistance to sparse design points, which is validated through Monte Carlo simulation study. Theoretically, under mild regularity conditions, the estimators constructed with Gamma asymmetric kernels possess the consistency and asymptotic normality for large sample and the advantages such as bias reduction and shorter length of confidence bands through simulation experiments for finite sample.

Empirically, the estimators are illustrated through stock composite index from Shenzhen Stock Exchange in China under 5-minute high-frequency data and possess some advantages mentioned above. This means the second-order jump-diffusion model may be an alternative model to describe the dynamic of some financial data, especially to explain integrated economic phenomena, that is the current observation in empirical finance usually behaves as the cumulation of all past perturbations.

Acknowledgments

The authors would like to thank the editor and two anonymous referees for their valuable suggestions, which greatly improved our paper. This research work is supported by Ministry of Education, Humanities and Social Sciences Project (No. 18YJCZH153), National Statistical Science Research Project (No. 2018LZ05), the General Research Fund of Shanghai Normal University (No. SK201720), Funding Programs for Youth Teachers of Shanghai Colleges and Universities (No. A-9103-18-104001), Key Subject of Quantitative Economics and Academic Innovation Team Construction Project in Shanghai Normal University and Shanghai Training Program of Innovation and Entrepreneurship for Undergraduates (NO. 201710270149).

A Proofs

A.1 Procedure for Assumption 3

Notice that the expectation with respect to the distribution of ξn,i depends on the stationary densities of Xn,i and X~n,i because ξn,i is a convex linear combination of Xn,i and X~n,i.

For the case (i): E[|hnK(X(i1)Δn)|]<. For KG(x/hn+1,hn)(u), its first-order derivative has the form of KG(x/hn+1,hn)(u)=(xhn)ux/hn1exp(u/hn)hnx/hn+1Γ(x/hn+1)(1hn)ux/hnexp(u/hn)hnx/hn+1Γ(x/hn+1):=1hnK1(u)+1hnK2(u). Then using the well-known properties of the Γ function, the mean of Gamma distribution and the derivative of the function gp(x):=g(x)p(x) for stationary process Xt, we have

E[|hnK(X(i1)Δn)g(X(i1)Δn)|]=E[|K1(X(i1)Δn)g(X(i1)Δn)|]+E[|K2(X(i1)Δn)g(X(i1)Δn)|]=xhnx/hnΓ(x/hn)hnx/hn+1Γ(x/hn+1)0yx/hn1exp(y/hn)hnx/hnΓ(x/hn)gp(y)dy+0yx/hnexp(y/hn)hnx/hn+1Γ(x/hn+1)gp(y)dy=0yx/hn1exp(y/hn)hnx/hnΓ(x/hn)gp(y)dy+0yx/hnexp(y/hn)hnx/hn+1Γ(x/hn+1)gp(y)dy=E[gp(ξ1)]+E[gp(ξ2)]=E[gp(E(ξ1)+ξ1E(ξ1))]+E[gp(E(ξ2)+ξ2E(ξ2))]=2gp(x)+O(hn)<,

where ξ1=DG(x/hn,hn), ξ2=DG(x/hn+1,hn) and G denotes the Gamma distribution.

For the case (ii):

E[|hn2K2(X(i1)Δn)g(X(i1)Δn)|]2E[|K12(X(i1)Δn)g(X(i1)Δn)|]+2E[|K22(X(i1)Δn)g(X(i1)Δn)|].

Now we only deal with the first part (the second part can be dealt with in the similar way). Note that K1(u)=ux/hn1exp(u/hn)hnx/hnΓ(x/hn) can be considered as a density function for a random variable ξ1=DG(x/hn,hn). By the property of the Γ function, we have with ηx=DG(2x/hn1,hn),

E[|K12(X(i1)Δn)g(X(i1)Δn)|]=Bhn(x)E[gp(ηx)]=Bhn(x)E[gp(E(ξ1)+ξ1E(ξ1))]=Bhn(x)gp(x)gp(x){12πhn1/2x1/2ifx/hn(interiorx'');hn1Γ(2κ1)22κ1Γ2(κ)ifx/hnκ(boundaryx''),

where Bhn(x)=hn1Γ(2x/hn1)22x/b1Γ2(x/hn+1) and the last equation follows from (Chen (2000), P474). hence, the results of limhn0hn1/2E[|hn2K2(ξn,i)g(ξn,i)|]< for “interior x” and limhn0hnE[|hn2K2(ξn,i)g(ξn,i)|]< for “boundary x” hold.

A.2 Some technical lemmas with proofs

We lay out some notations. For x=(x1,,xd), xj:=xj, xj2:=2xj2, xixj2:=2xixj, x:=(x1,,xd), and x2=(xixj2)1i,jd, where ∗ stands for the transpose.

Lemma 1

(Shimizu and Yoshida, 2006) Let Z be a d-dimensional solution-process to the stochastic differential equation

Zt=Z0+0tμ(Zs)ds+0tσ(Zs)dWs+0tEc(Zs,z)r(ω,dt,dz),

where Z0 is a random variable, E=Rd{0}, μ(x),c(x,z) are d-dimensional vectors defined on Rd,Rd×E respectively, σ(x) is a d × d diagnonal matrix defined on Rd, and Wt is a d-dimensional vector of independent Brownian motions.

Let g be a C2(l+1)-class function whose derivatives up to 2(l + 1)th are of polynomial growth. Assume that the coefficients μ(x), σ(x), and c(x, z) are C2l-class function whose derivatives with respective to x up to 2lth are of polynomial growth. Under Assumption 5, the following expansion holds

(18)E[g(Zt)|Fs]=j=0lLjg(Zs)Δnjj!+R,

for t > s and Δn=ts, where R=0Δn0u10ulE[Ll+1g(Zs+ul+1)|Fs]du1dul+1 is a stochastic function of order Δnl+1,Lg(x)=xg(x)μ(x)+12tr[x2g(x)σ(x)σ(x)]+E{g(x+c(x,z))g(x)xg(x)c(x,z)}f(z)dz.

Remark 11

Consider a particularly important model:

{dYt=Xtdt,dXt=μ(Xt)dt+σ(Xt)dWt+Ec(Xt,z)r(w,dt,dz).

As d = 2, we have

(19)Lg(x,y)=x(g/y)+μ(x)(g/x)+12σ2(x)(2g/x2)+E{g(x+c(x,z),y)g(x,y)gxc(x,z)}f(z)dz.

Based on the second-order infinitesimal operator (19), we can calculate many mathematical expectations involving X~iΔn, for instance (5) and (6) which provide the basis for estimators (7) and (8).

Lemma 2

Under Assumption 1, Assumption 2 and Assumption 5, let

p^n0(x)=1ni=1nKG(x/hn+1,hn)(X(i1)Δn),a^n0(x)=An0(x)p^n0(x),b^n0(x)=Bn0(x)p^n0(x),

where

An0(x)=1ni=1nKG(x/hn+1,hn)(X(i1)Δn)XiΔnX(i1)ΔnΔn,Bn0(x)=1ni=1nKG(x/hn+1,hn)(X(i1)Δn)(XiΔnX(i1)Δn)2Δn.

We have

p^n0(x)pp(x),a^n0(x)pμ(x),b^n0(x)pσ2(x)+Ec2(x,z)f(z)dz,

and for “interior x”, if hn=O((nΔn)2/5), then

nΔnhn1/2(a^n0(x)μ(x)hnBa^n0(x))dN(0,M(x)2πx1/2p(x)),nΔnhn1/2(b^n0(x)M(x)hnBb^n0(x))dN(0,Ec4(x,z)f(z)dz2πx1/2p(x)),

for “boundary x”, if hn=O((nΔn)1/3), then

nΔnhn(a^n0(x)μ(x)hnBa^n0(x))dN(0,M(x)Γ(2κ+1)22κ+1Γ2(κ+1)p(x)),nΔnhn(b^n0(x)M(x)hnBb^n0(x))dN(0,Ec4(x,z)f(z)dzΓ(2κ+1)22κ+1Γ2(κ+1)p(x)),

where Ba^n0(x), Bb^n0(x) denotes the bias of the estimators of a^n0(x), b^n0(x), respectively, that is

Ba^n0(x)=μ(x){1+xp(x)p(x)}+x2μ(x),Bb^n0(x)=M(x){1+xp(x)p(x)}+x2M(x),
Remark 12

The main method to obtain the asymptotic properties for estimators of model (2) is to approximate the estimators for (2) by the similar estimator for model (1) in probability such as (21) and (22). Based on the basic idea, we should first know the asymptotic properties for the Nadaraya-Watson estimator for (1) before the Theorem 2 and Theorem 3 presented. Lemma 2 give us the desired properties for Nadaraya-Watson estimator of (1).

Proof.

Corollary 2 in Hanif (2016) is the stationary case for Theorem 2 in Hanif (2016). Here we only mention the procedure for modified proof to Theorem 2 on the asymptotic normality in Hanif (2016). For convenience, we still use the same notations as that in the proof of Theorem 2 in Hanif (2016).

The bias expression hn{M(x){1+xp(x)p(x)}+x2M(x)}+o(hn) can be similarly deduced as equation (B10) in Gospodinovy and Hirukawa (2012) or equation (4.18) in Hanif (2016). Similarly as equations (159–174) in Bandi and Nguyen (2003), the dominant part for the variance of the asymptotic normality is D44 due to the relationship C44=OP(Δn,TD44)=op(D44). We write as Hanif (2016)

D44=i=1nKG(x/hn+1,hn)(XiΔn,T)iΔn,T(i+1)Δn,TY((Xs+c)2Xs22XiΔn,Tc(Xs,y))ν¯(ds,dy)Δn,Ti=1nKG(x/hn+1,hn)(XiΔn,T)=i=1nKG(x/hn+1,hn)(XiΔn,T)iΔn,T(i+1)Δn,TY((Xs+c)2Xs2)ν¯(ds,dy)Δn,Ti=1nKG(x/hn+1,hn)(XiΔn,T)i=1nKG(x/hn+1,hn)(XiΔn,T)iΔn,T(i+1)Δn,TY2XiΔn,Tc(Xs,y)ν¯(ds,dy)Δn,Ti=1nKG(x/hn+1,hn)(XiΔn,T):=an,T+bn,T.

By the Lemma 3 in Hanif (2016), we have

Δn,Ti=1nKG(x/hn+1,hn)(XiΔn,T)a.s.L¯X(T,x).

According to the similar procedures as equations (130), (169) and (172) in Bandi and Nguyen (2003), we can obtain

L¯X(T,x)(an,TNumΔn,Ti=1nKG(x/hn+1,hn)(XiΔn,T))DN(0,Ahn(x)E((x+c(x,z))2x2)2f(z)dz)

and

L¯X(T,x)(bn,TNumΔn,Ti=1nKG(x/hn+1,hn)(XiΔn,T))DN(0,Ahn(x)E4c2(x,z)x2f(z)dz),

where from Chen (2000)

Ahn(x)={hn1/22πx1/2+o(hn1/2)ifx/hn:interiorx'';hn1Γ(2κ+1)22κ+1Γ2(κ+1)+o(hn1)ifx/hnκ:boundaryx''.

Finally, using the similar procedure as equation (173) in Bandi and Nguyen (2003), it can be shown that the limiting covariance between

L¯X(T,x)(an,TNumΔn,Ti=1nKG(x/hn+1,hn)(XiΔn,T)) and L¯X(T,x)(bn,TNumΔn,Ti=1nKG(x/hn+1,hn)(XiΔn,T)) is characterized as

Asicov(L¯X(T,x)an,TNumΔn,Ti=1nKG(x/hn+1,hn)(XiΔn,T),L¯X(T,x)bn,TNumΔn,Ti=1nKG(x/hn+1,hn)(XiΔn,T))a.s.2Ahn(x)E[xc(x,z)((x+c(x,z))2x2)]f(z)dz

Hence,

L¯X(T,x)D44DN(0,Ahn(x)Ec4(x,z)f(z)dz),

that is,

{L¯X(T,x)hn1/2D44DN(0,Ec4(x,z)f(z)dz2πx1/2)ifx/hn:interiorx'';L¯X(T,x)hnD44DN(0,Γ(2κ+1)Ec4(x,z)f(z)dz22κ+1Γ2(κ+1))ifx/hnκ:boundaryx''

due to the form of Ahn(x).

The equation (4.19) in Hanif (2016) shows that

{L¯X(T,x)hn1/2Δn,TC44DN(0,4σ4(x)2πx1/2)ifx/hn:interiorx'';L¯X(T,x)hnΔn,TC44DN(0,Γ(2κ+1)4σ4(x)22κ+1Γ2(κ+1))ifx/hnκ:boundaryx''

from which we can deduce C44=OP(Δn,TD44).   □

Lemma 3

Assumption 1Assumption 5 lead to the following results,

  1. 1ni=1nKG(x/hn+1,hn)(X~(i1)Δn)1ni=1nKG(x/hn+1,hn)(X(i1)Δn)p0,

  2. 1ni=1nKG(x/hn+1,hn)(X~(i1)Δn)X~(i+1)ΔnX~iΔnΔn1ni=1nKG(x/hn+1,hn)(X(i1)Δn)X~(i+1)ΔnX~iΔnΔnp0,
  3. 1ni=1nKG(x/hn+1,hn)(X~(i1)Δn)32(X~(i+1)ΔnX~iΔn)2Δn1ni=1nKG(x/hn+1,hn)(X(i1)Δn)32(X~(i+1)ΔnX~iΔn)2Δnp0,

Remark 13

To some extend, this lemma can simplify the proofs of Theorem 2 and Theorem 3 such as δ1, n and δ2, n. Moreover, we can obtain the limit in probability of minuend in (ii) or (iii) of Lemma 3 by calculating the limit in probability of subtractor in (ii) or (iii) (which is easier to obtain). One can refer to Nicolau (2007) for the same idea.

Proof.

Let ε1,n=p^n(x)p^n0(x)=1ni=1nKG(x/hn+1,hn)(X~(i1)Δn)1ni=1nKG(x/hn+1,hn)(X(i1)Δn).

(20)max1in|X~(i1)ΔnX(i1)Δn|max1in1Δn|(i2)Δn(i1)Δn(XsX(i1)Δn)ds|max1insup(i2)Δns(i1)Δn|XsX(i1)Δn|=Oa.s.(Δnlog(1/Δn)),

the last asymptotic equation for the order of magnitude, one can refer to Bandi and Nguyen (2003) (Equations (94) and (95) on Pages 315–316).

By the mean-value theorem, stationarity, Assumption 3, Assumption 5 and (20), we obtain

E[|ε1,n|]E[1ni=1n|K(ξn,i)(X~(i1)ΔnX(i1)Δn)|]=E[|K(ξn,2)(X~ΔnXΔn)|]Δnlog(1/Δn)E[|K(ξn,2)|]=Δnlog(1/Δn)hnE[|hnK(ξn,2)|]0,

where ξn,2=θXΔn+(1θ)X~Δn0θ1.

Lemma 3 (i) follows from Chebyshev’s inequality.

Now we prove (ii), write

ε2,n=An(x)An0(x)=1ni=1nK(X~(i1)Δn)(X~(i+1)ΔnX~iΔn)Δn1ni=1nK(X(i1)Δn)(X~(i+1)ΔnX~iΔn)Δn.

We have

E[ε2,n]=E[1ni=1n(K(X~(i1)Δn)K(X(i1)Δn))E[X~(i+1)ΔnX~iΔnΔn|F(i1)Δn]]=E[1ni=1n(K(X~(i1)Δn)K(X(i1)Δn))(μ(X(i1)Δn)+Op(Δn))]=E[K(ξn,2)(X~ΔnXΔn)(μ(XΔn)+Op(Δn))]

by (5), the mean-value theorem and stationarity. Hence

|E[ε2,n]|E[hn|K(ξn,2)||X~ΔnXΔn|hn|μ(XΔn)+Op(Δn)|]Δnlog(1/Δn)hnE[hn|K(ξn,2)||μ(XΔn)+Op(Δn)|]0,

by Assumption 3, Assumption 5 and (20). So E[ε2,n]0.

Moreover

Var[ε2,n]=1nΔnhnVar[1ni=1nhnK(ξn,i)(X~(i1)ΔnX(i1)Δn)ΔnX~(i+1)ΔnX~iΔnΔn]=:1nΔnhnVar[1ni=1nfi],

where fi:=hnK(ξn,i)(X~(i1)ΔnX(i1)Δn)ΔnX~(i+1)ΔnX~iΔnΔn.

We find Var[1ni=1nfi]=1ni=1nVar[fi]+εn where εn represents the sum of 2nj=1n1i=j+1n terms involving the autovariances.

Firstly, we prove that E[fi2]<. Now we calculate E[fi2]:

E[fi2]=E[hnK2(ξn,i)(X~(i1)ΔnX(i1)Δn)2E[(X~(i+1)ΔnX~iΔn)2Δn|F(i1)Δn]]=E[hnK2(ξn,i)(X~(i1)ΔnX(i1)Δn)2(23σ2(X(i1)Δn+23Ec2(X(i1)Δn,z)f(z)dz)+Op(Δn))]Δnlog(1/Δn)hnE[hn2K2(ξn,i)(23σ2(X(i1)Δn)+23Ec2(X(i1)Δn,z)f(z)dz+Op(Δn))]=Δnlog(1/Δn)hn{1hn1/2hn1/2E[hn2K2(ξn,i)(23σ2(X(i1)Δn+23Ec2(X(i1)Δn,z)f(z)dz)+Op(Δn))](ifx/hn:interiorx'');1hnhnE[hn2K2(ξn,i)(23σ2(X(i1)Δn+23Ec2(X(i1)Δn,z)f(z)dz)+Op(Δn))](ifx/hnκ:boundaryx'')C{Δnlog(1/Δn)hn3/2ifx/hn(interiorx'');Δnlog(1/Δn)hn2ifx/hnκ(boundaryx'')<

by (6) and Assumption 3 and Assumption 5.

Then, we prove that εn=O(1Δnα).

Similarly as that pointed out in Ditlevsen and Sørensen (2004), we know that {XiΔn,i=1,2,} and {X~iΔn,i=1,2,} are stationary and ρ-mixing with the same size. Moreover, measurable functions of ρ-mixing processes are ρ-mixing with the same size in the space of square integrable functions according to the defination of ρ-mixing.

We have already proved E[fi2]< above, so fi is stationary under Assumption 2 and ρ-mixing with the same size as {XiΔn,i=1,2,} and {X~iΔn,i=1,2,}. Similar argument one can refer to Nicolau (2007). From Lemma 10.1.c with p = q = 2 on page 132 in Lin and Bai (2010), we have

|εn|=2n|j=1n1i=j+1n(EfifjEfiEfj)|2nj=1n1i=j+1n|EfifjEfiEfj|8nj=1n1i=j+1nρ((ij)Δn)(Efi2)12(Efj2)12=8nj=1n1i=j+1nρ((ij)Δn)Ef22

Last equation is due to the stationarity of fi.

We have proved E[f22]< above. Moreover, we have i=j+1nρ((ij)Δn)=O(1Δnα) under Assumption 2. Hence |εn|8nj=1n1i=j+1nρ((ij)Δn)Ef22=O(1Δnα). So the series εn=O(1Δnα) under Assumption 2.

In conclusion,

Var[ε2,n]=1nhnΔnVar[1ni=1nfi]=O(1nhnΔn1+α)0asn

by Assumption 5. So Lemma 3 (ii) can be deduced from the above considerations.

Finally, write

ε3,n=Bn(x)Bn0(x)=1ni=1nK(X~(i1)Δn)(X~(i+1)ΔnX~iΔn)2Δn1ni=1nK(X(i1)Δn)(X~(i+1)ΔnX~iΔn)2Δn.

Lemma 3 (iii) now follows:

E[|ε3,n|]E[1ni=1n|K(X~(i1)Δn)K(X(i1)Δn)|E[(X~(i+1)ΔnX~iΔn)2Δn|F(i1)Δn]]CE[K(ξn,2)|X~ΔnXΔn|(σ2(XΔn)+Ec2(XΔn,z)f(z)dz+Op(Δn))]Δnlog(1/Δn)hnE[hnK(ξn,2)(σ2(XΔn)+Ec2(XΔn,z)f(z)dz+Op(Δn))]0

by (6) and Assumption 3 and Assumption 5, which implies Lemma 3 (iii).   □

A.3 The proof of Theorem 1

Proof.

One can easily prove Theorem 1 by Lemma 2 and Lemma 3 (i).   □

A.4 The proof of Theorem 2

Proof.

(i) From Theorem 1 we get p^n(x)pp(x), hence to prove

a^n(x)=An(x)p^n(x)pμ(x)

it is sufficient to verify that

An(x)pμ(x)p(x).

From Lemma 2. we obtain

An0(x)pμ(x)p(x).

Now we prove

(21)An(x)An0(x)p0.

By Lemma 3 (ii) An(x)An0(x) has the same limit in probability as

δ1,n(x):=1ni=1nK(X(i1)Δn)X~(i+1)ΔnX~iΔnΔn1ni=1nK(X(i1)Δn)XiΔnX(i1)ΔnΔn=1ni=1nK(X(i1)Δn)(X~(i+1)ΔnX~iΔnΔnXiΔnX(i1)ΔnΔn).

We now prove that δ1,n(x)p0. Using Assumption 1, Assumption 2, Assumption 4 and Lemma 1, we find

E[δ1,n(x)]=E[K(X(i1)Δn)E[E[((X~(i+1)ΔnX~iΔn)Δn(XiΔnX(i1)Δn)Δn)|FiΔn]|F(i1)Δn]]=Δn2E[K(X(i1)Δn)(μ(X(i1)Δn)μ(X(i1)Δn)+12σ2(X(i1)Δn)μ(X(i1)Δn)+E{μ(X(i1)Δn+c(X(i1)Δn,z))μ(X(i1)Δn)μ(X(i1)Δn)c(X(i1)Δn,z)}f(z)dz)]=O(Δn)

and

Var[δ1,n(x)]=1nΔnhnVar[1ni=1nhnK(X(i1)Δn)Δn((X~(i+1)ΔnX~iΔn)Δn(XiΔnX(i1)Δn)Δn)]=:1nΔnhnVar[1ni=1ngi].

We have Var[1ni=1ngi]=1ni=1nVar[gi]+εn where εn represents the sum of 2nj=1n1i=j+1n terms involving the autovariances. Under Assumption 2, the series εn=O(1Δnα), and one easily obtains Var[1ni=1ngi]=O(1Δnα) if E[gi2]< with the same procedures as ε2, n in proof of Lemma 3 (ii). In fact

E[gi2]=E[hnK2(X(i1)Δn)Δn((X~(i+1)ΔnX~iΔn)Δn(XiΔnX(i1)Δn)Δn)2]

is finite because

E[Δn((X~(i+1)ΔnX~iΔn)Δn(XiΔnX(i1)Δn)Δn)2]=23(E[σ2(X0)]+E[Ec2(X0,z)f(z)dz])+O(Δn)

by Lemma 1 and Assumption 3. In conclusion,

Var[δ1,n]=1nΔnhnVar[1ni=1ngi]=O(1nhnΔn1+α)0asn

by Assumption 7.

(ii) By Lemma 2, for “interior x”, if hn=O((nΔn)2/5),

Un0(x):=nΔnhn1/2(a^n0(x)μ(x)hnBa^n0(x))dN(0,M(x)2πx1/2p(x))

and for “boundary x”, if hn=O((nΔn)1/3),

Un0(x):=nΔnhn(a^n0(x)μ(x)hnBa^n0(x))dN(0,M(x)Γ(2κ+1)22κ+1Γ2(κ+1)p(x)).

By the asymptotic equivalence theorem, it suffices to prove that

Un(x)Un0(x)p0

where Un(x)=hnnΔn(a^n(x)μ(x)).

From part (i) we know that

Un(x)Un0(x)=hnnΔn(δ1,n(x)p^n(x))

and

δ1,n(x)p^n(x)=Op(Δn)=op(hn),

thus the assumption hnnΔn30 leads to the result.   □

A.5 The proof of Theorem 3

Proof.

(i) From Theorem 1 we know p^n(x)pp(x), hence to prove

b^n(x)=Bn(x)p^n(x)pσ2(x)+Ec2(x,z)f(z)dz

it is sufficient to verify that

Bn(x)p(σ2(x)+Ec2(x,z)f(z)dz)p(x).

Following Lemma 2 under the conditions of the theorem we obtain

Bn0(x)p(σ2(x)+Ec2(x,z)f(z)dz)p(x).

So, we only need to prove that

(22)Bn(x)Bn0(x)p0.

By Lemma 3 (iii) Bn(x)Bn0(x) has the same limit in probability as

δ2,n(x):=1ni=1nK(X(i1)Δn)32(X~(i+1)ΔnX~iΔn)2Δn1ni=1nK(X(i1)Δn)(XiΔnX(i1)Δn)2Δn=1ni=1nK(X(i1)Δn)(32(X~(i+1)ΔnX~iΔn)2Δn(XiΔnX(i1)Δn)2Δn).

We now prove that δ2,n(x)p0. Let qi=32(X~(i+1)ΔnX~iΔn)2Δn(XiΔnX(i1)Δn)2Δn.

By Lemma 1

E[E[32(X~(i+1)ΔnX~iΔn)2Δn|FiΔn]|F(i1)Δn]=σ2(X(i1)Δn)+Ec2(X(i1)Δn)f(z)dz+O(Δn),E[(XiΔnX(i1)Δn)2Δn|F(i1)Δn]=σ2(X(i1)Δn)+Ec2(X(i1)Δn)f(z)dz+O(Δn).

One has

E[qi]=E[(32(X~(i+1)ΔnX~iΔn)2Δn(XiΔnX(i1)Δn)2Δn)]=O(Δn).

Similarly as δ1,n(x), we can easily verify that E[δ2,n(x)]0 by stationarity and the above equations.

On the other hand,

Var[δ2,n(x)]=1nΔnhnVar[1ni=1nhnK(X(i1)Δn)Δn(32(X~(i+1)ΔnX~iΔn)2Δn(XiΔnX(i1)Δn)2Δn)]=:1nΔnhnVar[1ni=1nsi].

Using the same arguments as in the proof of Theorem 2 (i), it is easy to conclude

E[Δn(32(X~(i+1)ΔnX~iΔn)2Δn(XiΔnX(i1)Δn)2Δn)2]=O(1)

assures Var[δ2,n(x)]0.

(ii) By Lemma 2, for “interior x”, if hn=O((nΔn)2/5),

Un0(x):=nΔnhn1/2(b^n0(x)M(x)hnBb^n0(x))dN(0,Ec4(x,z)f(z)dz2πx1/2p(x))

and for “boundary x”, if hn=O((nΔn)1/3),

Un0(x):=nΔnhn(b^n0(x)M(x)hnBb^n0(x))dN(0,Ec4(x,z)f(z)dzΓ(2κ+1)22κ+1Γ2(κ+1)p(x)).

By the asymptotic equivalence theorem, it suffices to prove that

Un(x)Un0(x)p0

where Un(x)=hnnΔn(b^n(x)(σ2(x)+Ec2(x,z)f(z)dz)).

From part (i) we know that

Un(x)Un0(x)=hnnΔn(δ2,n(x)p^n(x))

and

δ2,n(x)p^n(x)=Op(Δn)=op(hn),

Thus the assumption hnnΔn30 leads to the result.   □

References

Bandi, F., and T. Nguyen. 2003. “On the Functional Estimation of Jump Diffusion Models.” Journal of Econometrics 116: 293–328.10.1016/S0304-4076(03)00110-6Search in Google Scholar

Barndorff-Nielsen, O. E., and N. Shephard. 2006. “Econometrics of Testing for Jumps in Financial Economics using Bipower Variation.” Journal of Financial Econometrics 4: 1–30.10.1093/jjfinec/nbi022Search in Google Scholar

Chen, S. X. 2000. “Probability Density Function Estimation using Gamma Kernels.” Annals of the Institute of Statistical Mathematics 52: 471–480.10.1023/A:1004165218295Search in Google Scholar

Cont, R., and P. Tankov. 2004. Financial Modeling with Jump Processes. New York: Chapman and Hall/CRC.Search in Google Scholar

Chen, Y., and L. Zhang. 2015. “Local Linear Estimation of Second-Order Jump-Diffusion Model.” Communications in Statistics-Theory and Methods 44: 3903–3920.10.1080/03610926.2013.788718Search in Google Scholar

Ditlevsen, S., and M. Sørensen. 2004. “Inference for Observations of Integrated Diffusion Processes.” Scandinavian Journal of Statistics 31: 417–429.10.1111/j.1467-9469.2004.02_023.xSearch in Google Scholar

Fan, J., and I. Gijbels. 1996. Local Polynomial Modeling and its Applications. London: Chapman and Hall.Search in Google Scholar

Florens-Zmirou, D. 1989. “Approximate Discrete Time Schemes for Statistics of Diffusion Processes.” Statistics 20: 547–557.10.1080/02331888908802205Search in Google Scholar

Funke, B., and E. Schmisser. 2018. “Adaptive Nonparametric Drift Estimation of an Integrated Jump Diffusion Process.” Forthcoming in ESAIM: Probability and Statistics. DOI: 10.1051/ps/2018005.Search in Google Scholar

Gloter, A. 2001. “Parameter Estimation for a Discret Sampling of an Integrated Ornstein-Uhlenbeck Process.” Statistics 35: 225–243.10.1080/02331880108802733Search in Google Scholar

Gloter, A. 2006. “Parameter Estimation for a Discretly Observed Integrated Diffusion Process.” Scandinavian Journal of Statistics 33: 83–104.10.1111/j.1467-9469.2006.00465.xSearch in Google Scholar

Gospodinovy, N., and M. Hirukawa. 2012. “Nonparametric Estimation of Scalar Diffusion Processes of Interest Rates using Asymmetric Kernels.” Journal of Empirical Finance 19: 595–609.10.1016/j.jempfin.2012.04.001Search in Google Scholar

Gugushvili, S., and P. Spereij. 2012. “Parametric Inference for Stochastic Differential Equations: A Smooth and Match Approach.” Latin American Journal of Probability & Mathematical Statistics 9: 609–635.Search in Google Scholar

Hanif, M. 2015. “Nonparametric Estimation of Second-Order Diffusion Equation by using Asymmetric Kernels.” Communications in Statistics-Theory and Methods 44: 1896–1910.10.1080/03610926.2012.756915Search in Google Scholar

Hanif, M. 2016. “A Nonparametric Approach to the Estimation of Jump-Diffusion Models with Asymmetric Kernels.” Cogent Mathematics 3: 1–15.10.1080/23311835.2016.1179247Search in Google Scholar

Jacod, J., and A. Shiryaev. 2003. Limit Theorems for Stochastic Processes. 2nd ed. Berlin: Springer.10.1007/978-3-662-05265-5Search in Google Scholar

Johannes, M. S. 2004. “The Economic and Statistical Role of Jumps to Interest Rates.” Journal of Finance 59: 227–260.10.1111/j.1540-6321.2004.00632.xSearch in Google Scholar

Kessler, M. 1997. “Estimation of an Ergodic Diffusion from Discrete Observations.” Scandinavian Journal of Statistics 24: 211–229.10.1111/1467-9469.00059Search in Google Scholar

Kristensen, D. 2010. “Nonparametric Filtering of the Realized Spot Volatility: A Kernel-Based Approach.” Econometric Theory 26: 60–93.10.1017/S0266466609090616Search in Google Scholar

Lin, Z., and Z. Bai. 2010. Probability Inequalities. Beijing: Science Press.10.1007/978-3-642-05261-3Search in Google Scholar

Madan, D., and E. Seneta. 1990. “The Variance Gamma (V.G.) Model for Share Market Returns.” Journal of Business 63: 511–524.10.1086/296519Search in Google Scholar

Nicolau, J. 2007. “Nonparametric Estimation of Second-Order Stochastic Differential Equations.” Econometric theory 23: 880–898.10.1017/S0266466607070375Search in Google Scholar

Nicolau, J. 2008. “Modeling Financial Time Series Thriugh Second-Order Stochastic Differential Equations.” Statistics and Probability Letters 78: 2700–2704.10.1016/j.spl.2008.03.024Search in Google Scholar

Protter, P. 2004. Stochastic Integration and Differential Equations. 2nd ed. Berlin: Springer.10.1007/978-3-662-10061-5Search in Google Scholar

Shimizu, Y., and N. Yoshida. 2006. “Estimation of Parameters for Diffusion Processes with Jumps from Discrete Observations.” Statistical Inference for Stochastic Processes 9: 227–277.10.1007/s11203-005-8114-xSearch in Google Scholar

Song, Y. 2017. “Nonparametric Estimation for Second-Order Jump-Diffusion Model in High Frequency Data.” Forthcoming in Singapore Economic Reviews. DOI: 10.1142/S0217590817500102.Search in Google Scholar

Song, Y., Z. Lin, and H. Wang. 2013. “Re-Weighted Nadaraya-Watson Estimation of Second-Order Jump-Diffusion Model.” Journal of Statistical Planning and Inference 143: 730–744.10.1016/j.jspi.2012.09.010Search in Google Scholar

Wang, H., and Z. Lin. 2011. “Local Linear Estimation of Second-Order Diffusion Models.” Communications in Statistics-Theory and Methods 40: 394–407.10.1080/03610920903391345Search in Google Scholar

Wang, H., and L. Zhou. 2017. “Bandwidth Selection of Nonparametric Threshold Estimator in Jump-Diffusion Models.” Computers & Mathematics with Applications 73: 211–219.10.1016/j.camwa.2016.11.011Search in Google Scholar

Xu, K., and P. Phillips. 2011. “Tilted Nonparametric Estimation of Volatility Functions with Empirical Applications.” Journal of Business & Economic Statistics 29: 518–528.10.1198/jbes.2011.09012Search in Google Scholar

Yan, T., and C. Mei. 2016. A Test for a Parametric form of the Volatility in Second-Order Diffusion Models. Computational Statistics. DOI: 10.1007/s00180-016-0685-z.Search in Google Scholar


Supplementary Material

The online version of this article offers supplementary material (DOI: https://doi.org/10.1515/snde-2018-0001).


Published Online: 2019-04-09

©2019 Walter de Gruyter GmbH, Berlin/Boston

Downloaded on 12.9.2025 from https://www.degruyterbrill.com/document/doi/10.1515/snde-2018-0001/html
Scroll to top button