Home Time-varying threshold cointegration with an application to the Fisher hypothesis
Article
Licensed
Unlicensed Requires Authentication

Time-varying threshold cointegration with an application to the Fisher hypothesis

  • Lixiong Yang EMAIL logo
Published/Copyright: February 22, 2021

Abstract

This paper extends the threshold cointegration model developed by Gonzalo, J., and J. Y. Pitarakis. 2006. “Threshold Effects in Cointegrating Relationships.” Oxford Bulletin of Economics & Statistics 68: 813–33 and Chen, H. 2015. “Robust Estimation and Inference for Threshold Models with Integrated Regressors.” Econometric Theory 31 (4): 778–810 to allow for a time-varying threshold, which is a function of candidate variables that affect the separation of regimes. We derive the asymptotic distribution of the proposed least-square estimator of the threshold, and study the convergence rate of the threshold estimator. We also suggest test statistics for threshold effect and threshold constancy. Monte Carlo simulations point out that the convergence rate of the threshold estimator is consistent with the asymptotic theory, and the proposed tests have good size and power properties. The empirical usefulness of the proposed model is illustrated by an application to the US data to investigate the Fisher hypothesis.

JEL classification: C12; C13; C22; C51

Corresponding author: Lixiong Yang, School of Management, Lanzhou University, 222 South Tianshui Road, Lanzhou 730000, China. Phone: +86 13669327501, E-mail:

Award Identifier / Grant number: 71803072

Award Identifier / Grant number: Unassigned

Acknowledgements

The author would like to thank the editor and anonymous referees for very valuable comments and suggestions that result in a substantial improvement in this paper. Remaining errors and omissions are my own. The author acknowledges the financial support from the National Natural Science Foundation of China (Grant No. 71803072).

  1. Author contribution: All the authors have accepted responsibility for the entire content of this submitted manuscript and approved submission.

  2. Research funding: This study was supported by the National Natural Science Foundation of China (Grant No. 71803072).

  3. Conflict of interest statement: The authors declare no conflicts of interest regarding this article.

Appendix

This appendix provides the Proof of Theorem 1 in the paper. We first collect some lemmas that are useful in proving Theorem 1. Define 1 t ( u ) = 1 U t u , in which U t has a marginal U[0, 1] distribution, the partial-sum process

W t ( u ) = s = 1 t 1 s 1 ( u ) e t ,

W T ( s , u ) = 1 σ e T W [ T s ] ( u ) = 1 σ e T t = 1 [ T s ] 1 t 1 ( u ) e t ,

where σ e 2 = E ( e t 2 ) < .

Definition 1

W(s, u) is a two-parameter Brownian motion on (s, u) ∈ [0,1]2 if W(s, u) ∼ N(0, su) and

E W ( s 1 , u 1 ) W ( s 2 , u 2 ) = ( s 1 s 2 ) ( u 1 u 2 ) .

Lemma 1

Under Assumptions 1–3, W T (s, u) ⇒ W(s, u) on (s, u) ∈ [0,1]2 as T → ∞.

Proof

This result follows Theorem 1 of Caner and Hansen (2001). □

Lemma 2

If the joint probability density function of (q t , z t ) is f(q, z), then q t * q t γ 1 z t has a cumulative distribution function F γ 1 ( q * ) = q * f γ 1 ( u ) d u , where the probability density function f γ 1 ( q * ) = · · · + f ( q * + γ 1 z , z ) d z 1 d z k , in which z = (z 1, …, z k )′.

Proof

(A.1) F γ 1 ( q * ) = Pr ( Q * q * ) = · · · q γ 1 z q * f ( q , z ) d q d z 1 d z k = · · · + q * + γ 1 z f ( q , z ) d q d z 1 d z k = · · · + q * f ( u + γ 1 z , z ) d u d z 1 d z k = q * · · · + f ( u + γ 1 z , z ) d z 1 d z k d u q * f γ 1 ( u ) d u .

Lemma 3

Under Assumptions 1–4, for any γ ∈ Γ, we have 1 T t = 1 [ T s ] 1 ( q t γ t ) e t σ e 2 W ( s , F γ 1 ( γ 0 ) ) .

Proof

This result follows Lemma 1 by replacing u with F γ 1 ( γ 0 ) .

Lemma 4

Under Assumptions 1–4, for γ ∈ Γ, we have

1 T t = 1 [ T s ] x t 1 ( q t γ t ) e t σ e B v ( s ) d W ( s , F γ 1 ( γ 0 ) ) ,

in which 1 T t = 1 [ T s ] v t B v ( s ) is a Brownian motion with a positive definite long-run covariance matrix.

Proof

The results follows Theorem 2 of Caner and Hansen (2001), Lemma A.2 Chen (2015). □

Lemma 5

Under Assumptions 1–4, for any γ ∈ Γ, as T → ∞, define x t ( γ t ) = [ x t , x t 1 ( q t γ t ) ] , we have

1 T 2 t = 1 T x t ( γ t ) x t ( γ t ) = 1 T 2 x t x t x t x t 1 ( q t γ t ) x t 1 ( q t γ t ) x t x t 1 ( q t γ t ) x t 1 ( q t γ t ) B v ( s ) B v ( s ) d s F γ 1 ( γ 0 ) B v ( s ) B v ( s ) d s F γ 1 ( γ 0 ) B v ( s ) B v ( s ) d s F γ 1 ( γ 0 ) B v ( s ) B v ( s ) d s M ( γ 0 , γ 1 ) ,

1 T t = 1 T x t ( γ t ) e t = 1 T x t e t x t 1 ( q t γ t ) e t σ e B v ( s ) d W ( s ) B v ( s ) d W ( s , F γ 1 ( γ 0 ) ) .

Proof

The result follows Lemma 3 of Chen (2015). □

Lemma 6

Under Assumptions 1–5, β = [ β 2, β 1 β 2]′ = [ β 2, δ ′]′, in which δ = δ T = δ 0 T −1/2−α . When γ t = γ t 0 , we have T β ̂ β = O ( 1 ) . When γ t γ t 0 , we have T α + 1 / 2 β ̂ β = O ( 1 ) .

Proof

When γ t = γ t 0 ,

(A.2) T β ̂ β = 1 T 2 t = 1 T x t ( γ t 0 ) x t ( γ t 0 ) 1 T t = 1 T x t ( γ t 0 ) e t M ( γ 0 0 , γ 1 0 ) σ e B v ( s ) d W ( s ) B v ( s ) d W ( s , F γ 1 ( γ 0 ) ) .

When γ t γ t 0 and 1 2 < α < 1 2 , we note that

(A.3) y t = β 2 x t + δ x t 1 ( γ t 0 ) + e t = β x t ( γ t 0 ) + e t = β x t ( γ t ) + e t δ ( x t 1 ( γ t ) x t 1 ( γ t 0 ) ) ,

hence we have

(A.4) T α + 1 / 2 β ̂ β = T α + 1 / 2 t = 1 T x t ( γ t ) x t ( γ t ) 1 t = 1 T x t ( γ t ) y t = T α + 1 / 2 t = 1 T x t ( γ t ) x t ( γ t ) 1 t = 1 T x t ( γ t ) e t T α + 1 / 2 t = 1 T x t ( γ t ) x t ( γ t ) 1 t = 1 T x t ( γ t ) ( x t 1 ( γ t ) x t 1 ( γ t 0 ) ) δ = T α 1 / 2 O p ( 1 ) M ( γ t ) 1 1 T 2 t = 1 T x t ( γ t ) ( x t 1 ( γ t ) x t 1 ( γ t 0 ) ) δ 0 + o ( 1 ) M ( γ 0 , γ 1 ) 1 F γ 1 ( γ 0 ) F γ 1 0 ( γ 0 0 ) B v ( s ) B v ( s ) d s F γ 1 ( γ 0 ) F γ 1 ( γ 0 ) F γ 1 0 ( γ 0 0 ) B v ( s ) B v ( s ) d s δ 0 .

Lemma 7

Under Assumptions 1–5, ( γ ̂ 0 , γ ̂ 1 ) ( γ 0 0 , γ 1 0 ) .

Proof

For notational simplicity, we rewrite the model y t = β 2 x t + δ x t 1(γ t ) + e t in matrix form Y = β 2 X + δ X(γ t ) + e, in which X ( γ t ) = X 1 ( γ t ) .

Define X*(γ t ) = [X(γ t ), XX(γ t )], and P γ t * = X * ( γ t ) [ X * ( γ t ) X * ( γ t ) ] 1 X * ( γ t ) . Then Y and X lies in the space spanned by P γ t * , and we have

(A.5) SS R T ( γ t ) = Y ( I P γ t * ) Y = δ X ( γ t 0 ) ( I P γ t * ) X ( γ t 0 ) δ + 2 δ X ( γ t 0 ) ( I P γ t * ) e + e ( I P γ t * ) e .

When γ t = γ t 0 , we have SS R T ( γ t 0 ) = e ( I P γ t 0 * ) e . Using Lemmas 3 and 4, we have

(A.6) T 2 α 1 SS R T ( γ t ) SS R T ( γ t 0 ) = 1 T δ 0 X ( γ t 0 ) ( I P γ t * ) X ( γ t 0 ) δ 0 + o p ( 1 ) .

Using a similar argument of Theorem 1 in Chen (2015), we can show, for any γ t γ t 0 ,

(A.7) 1 T δ 0 X ( γ t 0 ) ( I P γ t * ) X ( γ t 0 ) δ 0 p F γ 1 0 ( γ 0 0 ) F γ 1 0 ( γ 0 0 ) F γ 1 ( γ 0 ) 1 F γ 1 0 ( γ 0 0 ) δ 0 B v ( s ) B v ( s ) d s δ 0 b 1 ( γ 0 , γ 1 ) .

Since F γ 1 0 ( γ 0 0 ) F γ 1 0 ( γ 0 0 ) F γ 1 ( γ 0 ) 1 F γ 1 0 ( γ 0 0 ) > 0 , and B v ( s ) B v ( s ) d s is a positive definite matrix, hence b 1(γ 0, γ 1) ≥ 0 and the equality holds if and only if ( γ 0 , γ 1 ) = ( γ 0 0 , γ 1 0 ) (i.e., γ t = γ t 0 ).

Similarly, for γ t γ t 0 we have

(A.8) 1 T δ 0 X ( γ t 0 ) ( I P γ t * ) X ( γ t 0 ) δ 0 p F γ 1 0 ( γ 0 0 ) F γ 1 ( γ 0 ) δ 0 B v ( s ) B v ( s ) d s δ 0 b 2 ( γ 0 , γ 1 ) ,

in which b 2(γ 0, γ 1) ≥ 0 and the equality holds if and only if γ t = γ t 0 since F γ 1 0 ( γ 0 0 ) F γ 1 ( γ 0 ) 0 .

Define b ( γ 0 , γ 1 ) = b 1 ( γ 0 , γ 1 ) 1 ( γ t γ t 0 ) + b 2 ( γ 0 , γ 1 ) 1 ( γ t γ t 0 ) . Combining the above results we have T 2 α 1 SS R T ( γ t ) SS R T ( γ t 0 ) p b ( γ 0 , γ 1 ) . Also, noting that γ ̂ t = arg min γ 0 , γ 1 Γ SS R T ( γ t ) , we have γ ̂ t = γ ̂ 0 + γ ̂ 1 z t minimizes SSR T (γ t ), and hence γ ̂ 0 + γ ̂ 1 z t p γ 0 0 + γ 1 0 z t . For any z t , define λ = ( 1 z t 2 + 1 , z t z t 2 + 1 ) . Clearly we have λ λ = 1, λ ( γ ̂ 0 , γ ̂ 1 ) p λ ( γ ̂ 0 0 , γ ̂ 1 0 ) . By the Cramér–Wold device theorem, we have ( γ ̂ 0 , γ ̂ 1 ) p ( γ 0 0 , γ 1 0 ) . □

Proof of Theorem 1

We first prove the following result a T ( γ ̂ 0 , γ ̂ 1 ) γ 0 0 , γ 1 0 = O p ( 1 ) , in which a T = T 1−2α .

To prove this, we need to prove that, for any v ̄ > 0 , we have

(A.9) lim T Pr ( γ ̂ 0 , γ ̂ 1 ) γ 0 0 , γ 1 0 v ̄ / a T = 1 .

For any B > 0, define V B = { γ 0 , γ 1 : ( γ 0 , γ 1 ) ( γ 0 0 , γ 1 0 ) < B } . When the sample size T is large enough, we have v ̄ / a T < B . Since ( γ ̂ 0 , γ ̂ 1 ) p ( γ 0 0 , γ 1 0 ) as proved before, we have lim T Pr ( γ ̂ 0 , γ ̂ 1 ) V B p 1 . Therefore, we only need to consider the limiting behavior of ( γ ̂ 0 , γ ̂ 1 ) in V B . Define a subset of V B : V B ( v ̄ ) = { γ 0 , γ 1 : v ̄ / a T < ( γ 0 , γ 1 ) ( γ 0 0 , γ 1 0 ) < B } . Thus, to prove Pr ( γ ̂ 0 , γ ̂ 1 ) γ 0 , γ 1 v ̄ / a T = 1 , we just need to prove Pr ( γ ̂ 0 , γ ̂ 1 ) V B ( ν ̄ ) = 0 .

Let S T * ( γ t ) = SSR T ( β ̂ , δ ̂ , γ t ) and S T * ( γ t 0 ) = SSR T ( β ̂ , δ ̂ , γ t 0 ) , where SSR T (.) is the sum of squared errors function (6) defined in the paper. From the estimation of ( γ ̂ 0 , γ ̂ 1 ) , we have S T * ( γ ̂ t ) S T * ( γ t 0 ) . Thus, to prove Pr ( γ ̂ 0 , γ ̂ 1 ) V B ( ν ̄ ) = 0 , it suffices to prove that for any ( γ 0 , γ 1 ) V B ( ν ̄ ) , we have

(A.10) lim T Pr S T ( β ̂ , δ ̂ , γ t ) > S T ( β ̂ , δ ̂ , γ t 0 ) = 1 .

To this end, we first consider the case of γ t γ t 0 . In this case, it is equivalent to prove

(A.11) S T * ( γ t ) S T * ( γ t 0 ) a T ( γ t γ t 0 ) > 0 .

Since Y = X β + X ( γ t 0 ) δ + e , we have

(A.12) S T * ( γ t ) S T * ( γ t 0 ) = Y X β X ( γ t ) δ Y X β X ( γ t ) δ Y X β X ( γ t 0 ) δ Y X β X ( γ t 0 ) δ = δ ̂ Δ X γ Δ X γ δ ̂ 2 δ ̂ Δ X γ e + 2 δ ̂ Δ X γ Δ X γ ( β ̂ β ) = δ n Δ X γ Δ X γ δ n 2 δ ̂ Δ X γ e + 2 δ ̂ Δ X γ Δ X γ ( β ̂ β ) + ( δ n + δ ̂ ) Δ X γ Δ X γ ( δ ̂ δ n ) = S 1 S 2 + S 3 + S 4 ,

in which Δ X γ = X ( γ t ) X ( γ t 0 ) . We next prove S 1 S 2 + S 3 + S 4 a T ( γ t γ t 0 ) converges to a positive random variable with probability one.

For the first term, we have

(A.13) S 1 a T = δ 0 T α Δ X γ Δ X γ δ 0 T α a T = δ 0 X ( γ t 0 ) X ( γ t ) X ( γ t 0 ) X ( γ t ) δ 0 T δ 0 F γ 1 ( γ 0 ) F γ 1 0 ( γ 0 0 ) B v ( s ) B v ( s ) d s δ 0 > 0 .

And the second term

(A.14) S 2 a T ( γ t γ t 0 ) = 2 δ ̂ 0 T α Δ X γ e a T ( γ t γ t 0 ) = 2 δ ̂ 0 T 1 / 2 α X ( γ t 0 ) X ( γ t ) e T 1 2 α ( γ t γ t 0 ) = 2 δ ̂ 0 X ( γ t 0 ) X ( γ t ) e T T 1 / 2 α ( γ t γ t 0 ) = O p 1 a n | γ t γ t 0 | = o p ( 1 ) .

By Lemma 5, T α + 1 / 2 δ ̂ = O ( 1 ) and T α + 1 / 2 ( β ̂ β ) = O p ( 1 ) . Hence, for the third and fourth terms we have

(A.15) S 3 a T ( γ t γ t 0 ) = 2 δ ̂ Δ X γ Δ X γ ( β ̂ β ) a T ( γ t γ t 0 ) = 2 T α + 1 / 2 δ ̂ Δ X γ Δ X γ T α + 1 / 2 ( β ̂ β ) T 2 ( γ t γ t 0 ) = o p ( 1 ) ,

(A.16) S 4 a T ( γ t γ t 0 ) = ( δ + δ ̂ ) Δ X γ Δ X γ ( δ ̂ δ ) T 1 2 α ( γ t γ t 0 ) = T α + 1 / 2 ( δ + δ ̂ ) Δ X γ Δ X γ T α + 1 / 2 ( δ ̂ δ ) T 2 ( γ t γ t 0 ) = o p ( 1 ) .

Thus, when γ t γ t 0 we have

(A.17) lim T Pr S T ( β ̂ , δ ̂ , γ t ) > S T ( β ̂ , δ ̂ , γ t 0 ) = 1 .

Using a similar procedure, it is easily to show that, when γ t γ t 0 and ( γ 0 , γ 1 ) V B ( ν ̄ ) , we also have lim T Pr S T ( β ̂ , δ ̂ , γ t ) > S T ( β ̂ , δ ̂ , γ t 0 ) = 1 . As discussed above, this is sufficient to establish the consistency in Theorem 1.

We next derive the asymptotic distribution:

(A.18) a T γ ̂ 0 γ 0 0 d ψ arg max r ( , ) ( W ( r ) r 2 ) ,

and

(A.19) a T γ ̂ 1 γ 1 0 d ψ argmax r ( , ) ( Λ ( r ) r 2 ) .

Since the threshold parameters are consistent with convergence rate T 1−2α , thus we can study their asymptotic behavior in the neighborhood of the true thresholds.

Let γ t = γ t 0 + w a T . By definition of threshold estimates, we have

(A.20) a T ( γ ̂ t γ t 0 ) w * = arg min w ( , ) S T * γ t 0 + w a n S T * ( γ t 0 ) .

Since γ t = γ 0 + γ 1 z t , we can rewrite the above results as

(A.21) a T γ ̂ 0 γ 0 0 w 1 * = arg min w 1 ( , ) S T * γ t 0 + w 1 a n S T * ( γ t 0 ) ,

(A.22) a T γ ̂ 1 γ 1 0 w 2 * = arg min w 2 ( , ) k S T * γ t 0 + w 2 a n z t S T * ( γ t 0 ) .

As in the proof of  the consistency in Theorem 1, we have S T * ( γ 0 0 + w 1 a T , γ 1 0 ) S T * ( γ 0 0 , γ 1 0 ) = S 1 S 2 + S 3 + S 4 . Then we can consider the limiting behavior of each S i . We only provide the proof for the case w 1 > 0, as the proof for the case w 1 ≤ 0 is similar.

For the first term we have

(A.23) S 1 = δ T X γ t 0 + w 1 a T X ( γ t 0 ) X γ t 0 + w 1 a T X ( γ t 0 ) δ T = δ 0 T 1 2 α 1 T 2 X γ t 0 + w 1 a T X ( γ t 0 ) X γ t 0 + w 1 a T X ( γ t 0 ) δ 0 = T 1 2 α δ 0 F γ 1 0 γ 0 0 + w 1 a T F γ 1 0 ( γ 0 0 ) B v ( s ) B v ( s ) d s δ 0 + o p ( 1 ) = T 1 2 α w 1 a T f γ 1 0 ( γ 0 0 ) δ 0 B v ( s ) B v ( s ) d s δ 0 + o p ( 1 ) p f γ 1 0 ( γ 0 0 ) w 1 δ 0 B v ( s ) B v ( s ) d s δ 0 .

For the second term we have

(A.24) S 2 = 2 δ ̂ 0 T 1 / 2 α 1 T X γ t 0 + w 1 a T X ( γ t 0 ) e 2 a T σ e δ 0 B v ( s ) d W s , F γ t 0 + w 1 a T W ( s , F ( γ 0 0 ) ) = 2 a T σ e δ 0 J 1 γ t 0 + w 1 a T J 1 ( γ t 0 ) ,

in which J 1(u) =∫B v (s)dW(s, u) is a zero-mean Gaussian process with an almost surely continuous sample path and with the covariance kernel

(A.25) E ( J 1 ( u 1 ) J 1 ( u 2 ) ) = ( u 1 u 2 ) B v ( s ) B v ( s ) d s .

It is easily to show that S 2 + S 3 = o p (1). Define D T ( w 1 ) = a T J 1 ( γ t 0 + w 1 a n ) J 1 ( γ t 0 ) . Then, by Lemma A.11 of Hansen (2000), D T (w 1) is a vector Brownian motion with covariance matrix f γ 1 0 ( γ 0 0 ) B v ( s ) B v ( s ) d s .

Combing the above convergence results we have

(A.26) S T * ( γ t ) S T * ( γ t 0 ) = S T * γ t 0 + w 1 a T S T * ( γ 0 0 ) = S 1 S 2 + S 3 + S 4 f γ 1 0 ( γ 0 0 ) w 1 δ 0 B v ( s ) B v ( s ) d s δ 0 2 σ e f γ 1 0 ( γ 0 0 ) δ 0 B v ( s ) B v ( s ) d s δ 0 W 1 ( w 1 ) ,

in which W 1(w 1) is a standard Brownian motion on [0, ∞]. Making the change-of-variables w 1 = σ e 2 f γ 1 0 ( γ 0 0 ) δ 0 B v ( s ) B v ( s ) d s δ 0 r , and noting that W(a 2 r) = aW(r), we have

(A.27) S T * γ t 0 + w 1 a T S T * ( γ t 0 ) 2 σ e 2 ( r 2 W 1 ( r ) ) .

Recall that a n γ ̂ 0 γ 0 0 w 1 * = arg min w 1 ( , ) S T * ( γ 0 0 + w 1 a n , γ 1 0 ) S T * ( γ 0 0 , γ 1 0 ) . Using continuous mapping theorem, the asymptotic distribution of γ ̂ 0 is

(A.28) a T γ ̂ 0 γ 0 0 r * p ψ 0 argmax r ( 0 , ) ( W 1 ( r ) r 2 ) ,

in which ψ 0 = σ e 2 f γ 1 0 ( γ 0 0 ) δ 0 B v ( s ) B v ( s ) d s δ 0 .

For w 1 < 0, we can prove it similarly. Hence, for γ ̂ 0 we have

(A.29) a T γ ̂ 0 γ 0 0 r * p ψ 0 argmax r ( , ) ( W ( r ) r 2 ) ,

in which W ( r ) = W 1 ( r ) , r > 0 0 , r = 0 W 2 ( r ) , r < 0

Using a similar procedure we can show that the asymptotic distribution of γ ̂ 1 = ( γ ̂ 11 , , γ ̂ 1 k ) . For any γ ̂ 1 j , j = 1 , , k , we have

(A.30) S T * γ t 0 + w 2 j a T z j t S T * ( γ t 0 ) = S 1 j S 2 j + S 3 j + S 4 j .

We only provide the proof for the case w 2j > 0, as the proof for the case with w 2j < 0 is analogous. For the first term, we have

(A.31) S 1 j = δ T X γ t 0 + w 2 j a T z j t X ( γ t 0 ) X γ t 0 + w 2 j a T z j t X ( γ t 0 ) δ T = δ 0 T 1 2 α 1 T 2 X γ t 0 + w 2 j a T z j t X ( γ t 0 ) X γ t 0 + w 2 j a T z j t X ( γ t 0 ) δ 0 = T 1 2 α δ 0 F γ 1 0 + w 2 j a T ( γ 0 0 ) F γ 1 0 ( γ 0 0 ) B v ( s ) B v ( s ) d s δ 0 + o p ( 1 ) ,

in which (as in Lemma 2)

(A.32) F γ 1 0 + w 2 j a T γ 0 0 F γ 1 0 ( γ 0 0 ) = γ 0 0 · · · + f u + γ 1 0 z + w 2 j a T z j , z d z 1 d z k d u γ 0 0 · · · + f ( u + γ 1 0 z , z ) d z 1 d z k d u = γ 0 0 · · · + f u + γ 1 0 z + w 2 j a T z j , z f ( u + γ 1 0 z , z ) d z 1 d z k d u p γ 0 0 · · · + f 1 ( u + γ 1 0 z , z ) w 2 j a T z j d z 1 d z k d u = w 2 j a T γ 0 0 · · · + f 1 ( u + γ 1 0 z , z ) z j d z 1 d z k d u w 2 j a T g j ( γ 0 0 , γ 1 0 ) .

Thus, we have

(A.33) S 1 j = T 1 2 α δ 0 F γ 1 0 + w 2 j a n ( γ 0 0 ) F γ 1 0 ( γ 0 0 ) B v ( s ) B v ( s ) d s δ 0 + o p ( 1 ) = T 1 2 α w 2 j a T g ( γ 0 0 , γ 0 1 ) δ 0 B v ( s ) B v ( s ) d s δ 0 + o p ( 1 ) p w 2 j g j ( γ 0 0 , γ 1 0 ) δ 0 B v ( s ) B v ( s ) d s δ 0 .

For the second term, we have

(A.34) S 2 j = 2 δ ̂ T 1 / 2 α 1 T X γ t 0 + w 2 j a T z j t X ( γ t 0 ) e = 2 a T σ e δ J 1 γ t 0 + w 2 j a T z j t J 1 ( γ t 0 ) + o p ( 1 ) .

Using a similar argument of Lemma 11 of Hansen (2000), we can show D ( w 2 j ) = a T J 1 ( γ t 0 + w 2 j a T z j t ) J 1 ( γ t 0 ) is vector Brownian motion with covariance matrix g j ( γ 0 0 , γ 1 0 ) B v ( s ) B v ( s ) d s .

And similarly, we have S 3j + S 4j = o p (1). Therefore, we have

(A.35) S T * γ t 0 + w 2 j a T z j t S T * ( γ t 0 ) = S 1 j S 2 j + S 3 j + S 4 j w 2 j g j ( γ 0 0 , γ 1 0 ) δ 0 B v ( s ) B v ( s ) d s δ 0 2 σ e g j ( γ 0 0 , γ 1 0 ) δ 0 B v ( s ) B v ( s ) d s δ 0 W 3 j ( w 2 ) .

Making the change-of-variables w 2 j = σ e 2 g j ( γ 0 0 , γ 1 0 ) δ 0 B v ( s ) B v ( s ) d s δ 0 r ( j ) , and using continuous mapping theorem, we have

(A.36) a T γ ̂ 1 j γ 1 j 0 ψ j argmax r ( , ) ( Λ j ( r ( j ) ) r ( j ) 2 ) ,

where Λ j ( r ( j ) ) = W 3 j ( r ) , r ( j ) > 0 0 , r ( j ) = 0 W 4 j ( r ( j ) ) , r ( j ) < 0 in which W 1(r), W 2(r), W 3 j ( r ) and W 4 j ( r ) are two independent standard Brownian motions on [0, ∞). ψ j = σ e 2 g j ( γ 0 0 , γ 1 0 ) δ 0 B v ( s ) B v ( s ) d s δ 0 , g j ( γ 0 0 , γ 1 0 ) = γ 0 0 · · · + f 1 ( u + γ 1 0 z , z ) z j d z 1 d z k d u . □

References

Andrews, D. W. K., and W. Ploberger. 1994. “Optimal Tests when a Nuisance Parameter is Present Only under the Alternative.” Econometrica 62 (6): 1383–414. https://doi.org/10.2307/2951753.Search in Google Scholar

Balke, N. S., and T. B. Fomby. 1997. “Threshold Cointegration.” International Economic Review 38: 627–45. https://doi.org/10.2307/2527284.Search in Google Scholar

Caner, M., and B. E. Hansen. 2001. “Threshold Autoregression with a Unit Root.” Econometrica 69: 1555–96. https://doi.org/10.1111/1468-0262.00257.Search in Google Scholar

Chen, H. 2015. “Robust Estimation and Inference for Threshold Models with Integrated Regressors.” Econometric Theory 31 (4): 778–810. https://doi.org/10.1017/s0266466614000553.Search in Google Scholar

Davies, R. B. 1977. “Hypothesis Testing when a Nuisance Parameter is Present Only under the Alternative.” Biometrika 64: 247–54. https://doi.org/10.1093/biomet/64.2.247.Search in Google Scholar

Davies, R. B. 1987. “Hypothesis Testing when a Nuisance Parameter is Present Only under the Alternative.” Biometrika 74: 33–43. https://doi.org/10.1093/biomet/74.1.33.Search in Google Scholar

Dueker, M. J., Z. Psaradakis, M. Sola, and F. Spagnolo. 2013. “State-Dependent Threshold Smooth Transition Autoregressive Models.” Oxford Bulletin of Economics & Statistics 75: 835–54. https://doi.org/10.1111/j.1468-0084.2012.00719.x.Search in Google Scholar

Elliott, G., T. J. Rothenberg, and J. H. Stock. 1996. “Efficient Tests for an Autoregressive Unit Root.” Econometrica 64: 813–36. https://doi.org/10.2307/2171846.Search in Google Scholar

Engle, R. F., and C. W. J. Granger. 1987. “Co-integration and Error Correction: Representation, Estimation, and Testing.” Econometrica 55: 251–76. https://doi.org/10.2307/1913236.Search in Google Scholar

Gonzalo, J., and J. Y. Pitarakis. 2006. “Threshold Effects in Cointegrating Relationships.” Oxford Bulletin of Economics & Statistics 68: 813–33. https://doi.org/10.1111/j.1468-0084.2006.00458.x.Search in Google Scholar

Hansen, B. E. 1996. “Inference when a Nuisance Parameter is Not Identified under the Null Hypothesis.” Econometrica 64 (2): 413–30. https://doi.org/10.2307/2171789.Search in Google Scholar

Hansen, B. E. 2000. “Sample Splitting and Threshold Estimation.” Econometrica 68: 575–603. https://doi.org/10.1111/1468-0262.00124.Search in Google Scholar

Hansen, B. E. 2017. “Regression Kink with an Unknown Threshold.” Journal of Business & Economic Statistics 35: 228–40. https://doi.org/10.1080/07350015.2015.1073595.Search in Google Scholar

Kwiatkowski, D., P. C. B. Phillips, P. Schmidt, and Y. Shin. 1992. “Testing the Null Hypothesis of Stationarity against the Alternative of a Unit Root: How Sure are We that Economic Time Series Have a Unit Root?” Journal of Econometrics 54: 159–78. https://doi.org/10.1016/0304-4076(92)90104-y.Search in Google Scholar

MacKinnon, J. G. 2010. “Critical Values for Cointegration Tests.” Queen’s Economics Department Working Paper.Search in Google Scholar

Million, N. 2004. “Central Bank’s Interventions and the Fisher Hypothesis: A Threshold Cointegration Investigation.” Economic Modelling 21: 1051–64. https://doi.org/10.1016/j.econmod.2004.03.002.Search in Google Scholar

Mishkin, F. S. 1991. “Is the Fisher Effect for Real? A Reexamination of the Relationship between Inflation and Interest Rates.” Journal of Monetary Economics 30: 195–215. https://doi.org/10.1016/0304-3932(92)90060-F.Search in Google Scholar

Petruccelli, J. D. 1992. “On the Approximation of Time Series by Threshold Autoregressive Models.” Sankhya: The Indian Journal of Statistics, Series B: 106–13, https://doi.org/10.2307/25052727.Search in Google Scholar

Porter, J., and P. Yu. 2015. “Regression Discontinuity Designs with Unknown Discontinuity Points: Testing and Estimation.” Journal of Econometrics 189: 132–47. https://doi.org/10.1016/j.jeconom.2015.06.002.Search in Google Scholar

Yang, L. 2019. “Regression Discontinuity Designs with Unknown State-dependent Discontinuity Points: Estimation and Testing.” Studies in Nonlinear Dynamics & Econometrics 23: 1–18. https://doi.org/10.1515/snde-2017-0059.Search in Google Scholar

Yang, L. 2020. “State-dependent Biases and the Quality of Chinas Preliminary GDP Announcements.” Empirical Economics 59: 2663–87. https://doi.org/10.1007/s00181-019-01751-z.Search in Google Scholar

Yang, L., and J.-J. Su. 2018. “Debt and Growth: is There a Constant Tipping Point?” Journal of International Money and Finance 87: 133–43. https://doi.org/10.1016/j.jimonfin.2018.06.002.Search in Google Scholar

Yu, P., and X. Fan. 2020. “Threshold Regression with a Threshold Boundary.” Journal of Business & Economic Statistics, https://doi.org/10.1080/07350015.2020.1740712. .Search in Google Scholar


Supplementary Material

The online version of this article offers supplementary material (https://doi.org/10.1515/snde-2018-0101).


Received: 2018-10-12
Revised: 2021-01-22
Accepted: 2021-01-28
Published Online: 2021-02-22

© 2021 Walter de Gruyter GmbH, Berlin/Boston

Downloaded on 7.10.2025 from https://www.degruyterbrill.com/document/doi/10.1515/snde-2018-0101/html
Scroll to top button