Home Estimation of Partially Specified Spatial Autoregressive Model
Article Publicly Available

Estimation of Partially Specified Spatial Autoregressive Model

  • Yuanqing Zhang EMAIL logo
Published/Copyright: June 25, 2014
Become an author with De Gruyter Brill

Abstract

In this paper, we study estimation of a partially specified spatial autoregressive model with heteroskedasticity error term. Under the assumption of exogenous regressors and exogenous spatial weighting matrix, we propose an instrumental variable estimation. Under some sufficient conditions, we show that the proposed estimator for the finite dimensional parameter is root-n consistent and asymptotically normally distributed and the proposed estimator for the unknown function is consistent and also asymptotically distributed though at a rate slower than root-n. Monte Carlo simulations verify our theory and the results suggest that the proposed method has some practical value.

1 Introduction

Since Paelinck coined the term “spatial econometrics” in the early 1970s to refer to a set of methods that explicitly handles with spatial dependence and spatial heterogeneity, this field has grown rapidly. The books by Cliff and Ord[1], Anselin[2], Cressie[3], contribute significantly to the development of the field. For a recent survey about the subject, see Anselin and Bera[4].

Among the class of spatial models, spatial autoregressive (SAR) models have attracted a lot of attention. Many methods have been proposed to estimate the SAR models, which include the method of maximum likelihood (ML) by Ord[5] and Smirnov and Anselin[6], the method of moments (MM) by Kelejian and Prucha[79], and the method of quasi-maximum likelihood estimation (QMLE) by Lee[10]. A common characteristic of these methods is that they are all developed to estimate finite dimensional parameters in the SAR models which are assumed to be linear. Linearity is widely used in the parametric statistical inference and it needs many model assumptions. Although their properties are very well established, linear models are often unrealistic in applications. Moreover, mis-specification of the data generation by a linear model could lead to a large bias. To achieve greater realism, Su and Jin[11] propose partially specified spatial autoregressive model in the following (hereafter PLSAR):

(1)yi=λ0j=1Nwijyj+xiβ0+g0(zi)+εi,i=1,2,,n

where yi denotes the dependent variable of individual i, xi is q-dimensional explanatory variable and zi denotes p-dimensional explanatory variable, wij denotes the spatial weight between individual i and j, εi denotes random noise, and λ0 and β0 denote the unknown true parameter value.

Recently, Su and Jin[11] give a profiled quasi-maximum likelihood based estimator (PQMLE) and show that the rates of consistency for the finite dimensional parameters in the model depend on some general features of the spatial weight matrix. Unfortunately, A major weakness for Su and Jin[11]’s PQMLE seems to be that the error term εi’s in their model are required to be homoscedastically distributed. In the presence of heteroscedasticity, Lin and Lee[12] have demonstrated that the QML-based estimator is usually inconsistent. Their estimator does not have a closed form expression. More recently, Su[13] proposes semi-parametric GMM (SPGMM) estimation of semi-parametric SAR models where the spatial lag effect exists the model linearly and the exogenous variables enter the model nonparametrically, but the model of Su[13] is absence of xiβ.

However, both estimators mentioned above seem to be computationally challenging for applied scientists and are not easy to use in practice. In this article, an alternative computationally simple estimator is proposed for consistently estimating PLSAR model (1) for both finite dimensional parameters and nonparametric component under heteroskedasticity in the model. We give the asymptotic normality and consistent covariance matrices for the parametric and nonparametric part. The estimators all have closed form expressions in the paper, so, it is easy to implement in practice. Monte Carlo simulation results indicate that our estimators perform well in finite samples.

The paper is organized as follows: Section 2 proposes an estimation of the model (1); Section 3 gives the asymptotic properties of the proposed estimators; Section 4 reports some Monte Carlo simulation results. All technical proofs are relegated to the appendices.

2 Estimation

There are two issues with estimation of the model (1). First is the correlation between the spatial term j=1nwijyj and the error term εi (i.e., endogeneity). Second is the infinite dimensionality of the unknown parameter g0(z), one can approximate it with a sieve method. Such an method was taken by Ai and Chen[14] who show that the sieve estimator of the nonparametric component (g0 here) is consistent with a rate faster than n–1/4 under certain metric; the estimator of the parametric component is nconsistent and asymptotically normally distributed; the estimator for the asymptotic covariance of the parametric component estimator is consistent and easy to compute; and the optimally weighted minimum distance estimator of parametric component attains the semi-parametric efficiency bound. Specifically, let pK(z) = (p1(z),p2(z), · · · , pK(z))′ denote a sieve estimator. For each K, there exists some constant π0 such that g0(z) = pK(z)⁚π0 + v0(z), with v0(z) satisfying ||v0(z)|| → 0 as K → ∞.

Let wi = (wi1, wi2, · · ·, win)T, Y = (y1, y2, · · ·, yn)T, We rewrite the model (1) yields

(2)yi=λwiTY+ziTβ+g(xi)+εi

Then taking conditional expectation yields

E(yi|xi)=λE(wiTY|xi)+E(ziT|xi)β+g(xi)+E(εi|xi).

Since E(εi|xi) = 0, we get

(3)yiE(yi|xi)=λ{wiTYE(wiTY|xi)}+{ziTE(ziT|xi)}β+εi

Finally, we deal with the endogeneity problem by applying the 2SLS. Let hi denote a column vector of instrumental variables. The instrumental variables include all exogenous regressors in the model. The instrument matrices H have full column rank rq + 1.

Denote pi = pK(zi), P = (p1, p2, · · ·, pn)′, δ0=(λ0,β0),y^i=yiE^(yi|xi),Y^=(y^1,y^2,,y^n)Tb^i=(wiTYE^(wiTY|xi),ziTE^(ziT|xi))T,B^=(b^1,b^2,,b^n)T, where

E^(zi|xi)=pK(xi)T(PTP)1i=1npK(xi)Tzi,E^(yi|xi)=pK(xi)T(PTP)1i=1npK(xi)Tyi,E^(wiTZ|xi)=pK(xi)T(PTP)1i=1npK(xi)TwiTZ,E^(wiTY|xi)=pK(xi)T(PTP)1i=1npK(xi)TwiTY.

The 2SLS estimators for the unknow parameters δ0, π0, g0(z) are respectively given by

(4)δ^=[B^TH^(H^TH^)1H^TB^]1B^TH^(H^TH^)1H^TY^
(5)π^=PP1P{YBδ^}
(6)g^(z)=pK(z)π^

3 Asymptotic properties

The following assumptions will be maintained throughout the paper.

Assumption 1

(i) {(Xi, Zi, εi), i = 1, 2, · · ·, n} are independently and identically distributed; (ii) for all i, E[εi|Xi, Zi, W] = 0; (iii) there exist some constants σmin2,σmax2>0 such that σmax2σi2σmin2 holds, where σi2=E(εi2|Xi,Zi,W).(iv) for all i, E[εi4|Xi,Zi,W] is finite and bounded.

Assumption 2

(i) The diagonal element of the spatial weighting matrix W is zero; (ii) the matrix Iλ0W is nonsingular with |λ0| < 1, where I is an n × n identity matrix; (iii) the row and column sums of the matrices W, (Iλ0W)–1 are bounded uniformly in absolute value.

Assumption 1 is needed for establishing the asymptotic distribution of the nonparametric estimator g^(z) and was used in [15]. Assumption 2 imposes restrictions on the spatial weighting matrix. These restrictions are commonly imposed in the spatial regression literature (e.g., Lee[16]).

Assumption 3

For every K; (i) the smallest eigenvalue of Epipi is bounded away from zero uniformly in K and (ii) there is a sequence of constants ζ0(K) satisfying supzZpK(z)ζ0(K), and K = K (n) such that ζ0(K)2K/n → 0 as n → ∞.

Assumption 4

For any function g(·) satisfying the normalization of g0(·), (i) there exist some α(>0), π = π(K), supzZ|g(z)pK(z)π|=O(Kα)asK; (ii) nKα0 and n → ∞.

Assumption 5

The limits QH~TB~=plimnn1H~TB~,QH~TH~=plimnn1H~TH~,QH~TΛH~=plimnn1H~TΛH~=plimnn1i=1nh~ih~iTσi2. exist and nonsingular.

Assumption 3 are imposed on the sieve approximations. Since the constant is not identified, we must impose some normalization on g0(·) such as g0(z0) = 0 at some point z0 so that g0(·) can be identified. The basis functions pK(z) shall be constructed to satisfy this normalization. In addtion, the assumption imposes a normalization on the approximating functions, bounding the second moment matrix away from singularity, and restricting the magnitude of the series terms. This condition is needed to ensure the convergence in probability of the sample second moment matrix of the approximating functions to their expectations. Without loss of generality, we will assume Epipi=I.Basis functions satisfying this condition can always through appropriate normalization. Assumption 5 is a stability condition; the quantities are used in the expression of asymptotic variance of the estimator.

Define

B=CG0+Xβ0,X;Γn=B(IS)M(IS);Λn=diag{σ12,σ22,,σn2};Ωn=(QH~TB~TQH~TH~1QH~TB~)1QH~TB~TQH~TH~1QH~TΛH~QH~TH~1QH~TB~(QH~TB~TQH~TH~1QH~TB~)1.

The following assumptions will be maintained throughout.

Theorem 1

Suppose assumptions 1~5 are satisfied, then we can establish

Ωn1/2(δ^δ0)N(0,I)indistribution.

For any given value z0, denote

σn2(z0)=pK(z0)PP1PΛnPPP1pK(z0).

We can establish the following theorem regarding the nonparametric component.

Theorem 2

Suppose Assumptions 1~5 are satisfied, then we can establish: (i) π^=π0+Op(K/n+Kα); (ii) supzZg^(z)g0(z)=Op(ζ0(K)[K/n+Kα]); (iii) E[g^(z)g0(z)2]=Op([K/n+K2α]);(iv) g^(z0)g0(z0)σn(z0)N(0,1)in distribution.

We estimate the residuals by ε^i=YiBiδ^Piπ^ and estimate σi2byσ^i2=(YiBiδ^Piπ^)2

4 Monte Carlo simulation

In this section, we conduct a monte carlo experiment to evaluate finite sample performance of the proposed method. The data generating process is given by

(7)yi=λ0j=1Nwijyj+xiβ0+g0(zi)+εi,i=1,2,,n

where zi follows the uniform distribution on [0,1], xi follows the exponential distribution with parameter 3, and the error term εi is normally distributed with mean zero and variance σ2. We take the spatial weights matrix as the spatial scenario in [17] with R number of districts, m members in each district, and with each neighbor of a member in a district given equal weight, i.e., W = IRBm, where Bm=(lmlmIm)/(m1),m=10, where lm denotes m-dimensional column vector with all elements of 1, Im denotes identity matrix. We consider g(z) = 6 cos(2πz), λ = –0.5, 0, or 0.5, and β = –6 or 6. In the simulations, we take K = 6 and pK(z) = (1, z, z2, z3, z4, z5)′.

Table 1 summarizes the simulation results for estimating λ, β and g(z) under different selections of the true parameters for n = 200, 500, and 1000. Each entry is based on 1000 repetitions.

Table 1

when σ2 = 1, Summary of Bias, SEE, ESE and CP for λ and β, the estimator for σ2 under model (8)

σ2λßg(z)
nσ̂2BiasSEEESECPBiasSEEESECPRISE
λ = –0.5β = 6
2000.97060.00920.07200.07420.9460.00560.21290.21860.9440.2466
5000.9952–0.00090.04340.04360.9450.00500.13850.13710.9430.1740
10001.00200.00080.02930.03010.9600.00100.09630.09570.9490.1458
λ = 0β = 6
2000.96990.00720.05040.05010.9480.00480.21190.21760.9420.2457
5000.99510.00070.03060.02990.9400.00130.13790.13670.9430.1738
10001.0020–0.00060.02070.02070.9550.00110.096110.09540.9490.1456
λ = 0.5β = 6
2000.96910.00420.02650.02650.9510.00460.21230.21780.9440.2456
5000.9950.00040.01610.01600.9420.00360.13810.1370.9420.1739
10001.0020–0.00030.01090.01120.9570.00150.09640.09570.9480.1456
λ = –0.5β = –6
2000.97060.00860.06880.07360.9440.00610.21210.21860.9470.2446
5000.99520.00060.04210.04380.9570.00230.13860.13720.9430.1721
10001.0020–0.00050.02890.03000.9540.00190.09630.09570.9470.1453
λ = 0β =-6
2000.96990.00680.04820.04980.9420.00600.21160.21770.9440.2438
5000.99520.00050.02970.03000.9560.00430.13840.13670.9400.1720
10001.0020–0.00020.02040.02070.9500.00210.09620.09540.9480.1452
λ = 0.5β = –6
2000.96910.00400.02530.02640.9510.00700.21230.21790.9440.2438
5000.99520.00030.01570.01610.9590.00610.13900.13690.9380.1722
10001.0020–3.8e–050.01080.01110.9520.00230.09660.09570.9520.1452

For each estimator, Bias is the average of estimation biases from 1000 repetitions, SSE is the sample standard error of the estimates, ESE is the average of the estimated standard error, let SEE denote the sampling standard error of the estimates, CP is the coverage probability of a 95 percent confidence interval for the estimator, RISE is defined as the average of the square root of integrated square error of ĝ(z) where for each repetition,

RISE(g^(z))=Z{g^(z)g(z)}2dF(z),

where Ƶ is the support of z, F(z) is the distribution function of z.

Table 1 shows that σ̂2 is close to the true value 1 as n becomes larger. The biases of λ and β become smaller as n becomes larger. The SEE and ESE for λ̂ and β̂ are very close. The coverage probabilities of λ̂ and β̂ are close to 0.95 nominal level. The RISE of g(z) becomes smaller as n becomes larger. Figure 1 shows the comparison of the true function of g(z) and the estimating function of ĝ(z).

Figure 1 The plots of the estimates of g(z) = 6cos(2πz) under model (8) with n = 200 and σ = 1 for different values of λ and β.
Figure 1

The plots of the estimates of g(z) = 6cos(2πz) under model (8) with n = 200 and σ = 1 for different values of λ and β.

5 Appendix technicalities

We denote tr(A) = trace(A) for a square matrix A.

Before we prove the main results, we first give some useful convergence results. By Assumption 3(i), we assume without loss of generality that: E {pkK(zi)pjK(zi)} = Ijk, where Ijk denotes the (j, k)th element of the identity matrix. Let ||·|| denotes the Euclidean norm. Denote Q=PPn, we obtain

(8)QI=Op(ζ0(K)K/n)=op(1)

Let 1n denote the indicator function for the smallest eigenvalue of Q being greater than 1/2. Then limnP(1n=1)=1. By Assumption 1, we obtain

(9)1nQ1/2Pε/n=Op(K/n)

and

(10)1nQ1Pε/n=Op(K/n)
(11)1nQ1PV0/n=Op(Kα)

Proof of Theorem 1

For the sake of convenience, we take the instrument variables as hi=(xiwiTZ,z~iT)T in the proof, Η = (h1, h2, · · · , hn)T. m(ui) = E[hi|ui], vi = him(ui), we will use the shorthand notation: mi = m(ui), the variables without the subscript represent the matrices, e.g, m = (m1, m2, · · · , mn)T. For any matrix A, we define A=P(PTP)1PTA, applying this definition to , , ε, , we get Y,m,ε,ν. From hi = mi + vi, we get hi=mi+νi, or in vector matrix notation, H = m + v, H=m+ν. For scalars and vectors A1i and A2i, let us define SA1,A2=n1iA1iA2iT,Also we let SA1,A1=SA1 Because

ESν,ε|X2=n2tr[P(PTP)1PTννTP(PTP)1PTE(εεT|X)]Cn2tr[P(PTP)1PTννTP(PTP)1PT]Cn2tr[ννT]Cn1Sν=O(K/n2).

Hence,

Sν,ε=Op(K/n).

Furthermore,

Smm,ε{SmmSε}1/2=Op(Kα)Op(K/n),

and

Sν,ε{SνSε}1/2=Op(K/n).

Because

SHH,ε=Sν+(mm)ν,ε=Sν,ε+Smm,ε+Sν,ε,

So,

SHH,ε=op(n1/2).

Because

ESmm,ε|X2=n2tr[(mm)(mm)TE(εεT|X)]Cn2tr[(mm)(mm)T]Cn1Smm=op(n1),

Hence,

Smm,ε=op(n1/2);

Because

ESν,ε|X2Cn1tr[Sν]=op(n1);

Hence,

Sν,ε=op(n1/2);SHH,ε=Sν+(mm)ν,ε=Sν,ε+Smm,εSν,ε.nSHH,ε=nSν,ε+op(1).

So,

(12)nSHH,εN(0,QH~TΛH~).[B^TH^n(H^TH^n)1H^TB^n]1B^TH^n(H^TH^n)1
(13){QH~TB~TQH~TH~1QH~TB~}1QH~TB~TQH~TH~1n[δ^δ]=[B^TH^n(H^TH^n)1H^TB^n]1B^TH^n(H^TH^n)1n{SHH,εSHH,ε}

Thus,

n[δ^δ]={QH~B~TQH~TH~1QH~TB~}1QH~TB~TQH~TH~1N(0,QH~TΛH~)+op(1),

The proof of Theorem is completed following an application of the Crémer-Wald device.

Proof of Theorem 2

We write

π^={PP}1P{YBδ^}=PP1PYBδ0+PP1PB(δ^δ0).

Consider the second term. We have

(PP)1PB(δ^δ0)2=(δ^δ0)BPnPPn2PBn(δ^δ0)=Op1n(δ^δ0)BSB(δ^δ0)=Op1n   by Assumption 5.

Thus,

π^=PP1PYBδ0+Op(n1/2)=π0+PP1Pε+PP1PV0+Op(n1/2)=π0+Op(K/n+Kα)

by applying (11). Notice that

1n|g^(z)g0(z)||pK(z)(π^π0)|+|pK(z)π0g0(z)|ζ0(K)1n||π^π0||+O(Kα)=Op(ζ0(K)[K/n+Kα]).

We obtain

g^(z)=g0(z)+Op(ζ0(K)[K/n+Kα])

uniformly over z and

supzZ|g^(z)g0(z)|=Op(ζ0(K)[K/n+Kα]).

Furthermore, applying (12) and Assumption 3(ii), we obtain

1nE(g^(z)g0(z))2=1nE[pK(z)(π^π0)+(pK(z)π0g0(z))]221n||π^π0||2+21nE[pK(z)π0g0(z)]2=Op(K/n+K2α)+O(K2α)=Op(K/n+K2α).

Notice that

pK(z0)(PP)1PB(δ^δ0)=Op(ζ0(K)n1/2).

By substitution and at a fixed z0, we have

g^(z0)g0(z0)=pK(z0)π^[pK(z0)π0+OKα]=pK(z0)[π0+PP1Pε+PP1PV0+Op(n1/2)][pK(z0)π0+OKα]=pK(z0)PP1Pε+Op(ζ0(K)[n1/2+Kα]).

Let ξin=pK(z0)PP1Piεiσn(z0), note that for each n, E[ξin]=0,sn2=i=1nE[ξin2]=1. For every > 0,

1sn2i=1nE[1(|ξin|>ϵsn)ξin2]=i=1nϵE[1(|ξin/ϵ|>1)ξin2]=i=1nϵ2E[1(|ξin/ϵ|>1)(ξin/ϵ)2]i=1nϵ2E[(ξin/ϵ)4]i=1nϵ2pK(z0)σn(z0)4×(PTP/n)14ζ0(K)2E{Pi2E[εi4|Z]}/n2ϵ4Cζ0(K)2K/n0by Assumptions 2, 3, 4.

where 1(·) denotes indicative function. Then, by the Lindbergh-Feller central limit theorem, i=1nξinN(0,1) distribution. So,

pK(z0)PP1Pεσn(z0)N(0,1)in distribution.

Thus,

g^(z)g0(z)σn(z0)N(0,1)in distribution.

6 Acknowledgements

We thank the referees for their time and comments.


Supported by National Natural Science Foundation of China (Grant No. 71371118); the Graduate Innovation Fund Project of Shanghai University of Finance and Economics (Grant No. CXJJ-2011-444)


References

[1] Cliff A D, Ord J K. Spatial autocorrelation. London: Pion Ltd, 1973.Search in Google Scholar

[2] Anselin L. Spatial econometrics: Methods and models. The Netherlands: Kluwer Academic Publishers, 1988.10.1007/978-94-015-7799-1Search in Google Scholar

[3] Cressie N. Statistics for spatial data. JohnWiley Sons, New York, 1993.10.1002/9781119115151Search in Google Scholar

[4] Anselin L, Bera A K. Spatial dependence in linear regressionmodels with an introduction to spatial econometrics. In: Ullah A, Giles D E A. Handbook of Applied Economics Statistics. Marcel Dekker, New York, 2002.Search in Google Scholar

[5] Ord J K. Estimation methods for models of spatial interaction. Journal of the American Statistical Association, 1975, 70: 120–126.10.1080/01621459.1975.10480272Search in Google Scholar

[6] Smirnov O, Anselin L. Fast maximum likelihood estimation of very large spatial autoregressive models: A characteristic polynomial approach. Computational Statistics and Data Analysis, 2001, 35: 301–319.10.1016/S0167-9473(00)00018-9Search in Google Scholar

[7] Kelejian H H, Prucha I R. A generalized spatial two-stage least squares procedure for estimating a spatial autoregressive model with sutoregressive disturbance. Journal of Real Estate Finance and Economics, 1998, 17: 99–121.10.1023/A:1007707430416Search in Google Scholar

[8] Kelejian H H, Prucha I R. A generalized moments estimator for the autoregressive parameter in a spatial model. International Economic Review, 1999, 40: 509–533.10.1111/1468-2354.00027Search in Google Scholar

[9] Kelejian H H, Prucha I R. Specification and estimation of spatial autoregres-sive models with autoregressive and heteroskedastic disturbances. Journal of Econometrics, 2010, 157: 53–67.10.1016/j.jeconom.2009.10.025Search in Google Scholar PubMed PubMed Central

[10] Lee L F. Asymptotic distribution of quasi-maximum likelihood estimators for spatial autoregressive models. Econometrica 2004, 72: 1899–1925.10.1111/j.1468-0262.2004.00558.xSearch in Google Scholar

[11] Su L, Jin S. Profile quasi-maximum likelihood estimation of spatial autoregressive models. Journal of Econometrics, 2010, 157: 18–33.10.1016/j.jeconom.2009.10.033Search in Google Scholar

[12] Lin X, Lee L F. GMM estimation of spatial autoregressive models with unknown heteroskedasticity. Journal of Econometrics, 2010, 157(1): 34–52.10.1016/j.jeconom.2009.10.035Search in Google Scholar

[13] Su L. Semi-parametric GMM estimation of spatial autoregressive models. Journal of Econometrics, 2012, 167: 543–560.10.1016/j.jeconom.2011.09.034Search in Google Scholar

[14] Ai C, Chen X. Efficient estimation of models with conditional moment restrictions containing unknown functions. Econometrica 2003, 71: 1795–1843.10.1111/1468-0262.00470Search in Google Scholar

[15] Newey W K. Convergence rates and asymptotic normality for series estimators. Journal of Econometrics, 1997, 79: 147–168.10.1016/S0304-4076(97)00011-0Search in Google Scholar

[16] Lee L F. GMM and 2SLS estimation of mixed regressive, spatial autoregressive models. Journal of Econometrics, 2007, 137: 489–514.10.1016/j.jeconom.2005.10.004Search in Google Scholar

[17] Case A C. Spatial patterns in household demand. Econometrica, 1991, 59: 953–965.10.2307/2938168Search in Google Scholar

Received: 2013-11-12
Accepted: 2014-1-3
Published Online: 2014-6-25

© 2014 Walter de Gruyter GmbH, Berlin/Boston

Downloaded on 20.11.2025 from https://www.degruyterbrill.com/document/doi/10.1515/JSSI-2014-0226/html?lang=en
Scroll to top button