Next Article in Journal
Personalized Sliding Window Recommendation Algorithm Based on Sequence Alignment
Previous Article in Journal
Deep Learning-Based Segmentation of Head and Neck Organs-at-Risk with Clinical Partially Labeled Data
Previous Article in Special Issue
Evolution of Cohesion between USA Financial Sector Companies before, during, and Post-Economic Crisis: Complex Networks Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Variable Selection of Spatial Logistic Autoregressive Model with Linear Constraints

School of Science, China University of Petroleum, Qingdao 266580, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(11), 1660; https://doi.org/10.3390/e24111660
Submission received: 18 August 2022 / Revised: 25 October 2022 / Accepted: 7 November 2022 / Published: 15 November 2022
(This article belongs to the Special Issue Statistical Methods for Complex Systems)

Abstract

:
In recent years, spatial data widely exist in various fields such as finance, geology, environment, and natural science. These data collected by many scholars often have geographical characteristics. The spatial autoregressive model is a general method to describe the spatial correlations among observation units in spatial econometrics. The spatial logistic autoregressive model augments the conventional logistic regression model with an extra network structure when the spatial response variables are discrete, which enhances classification precision. In many application fields, prior knowledge can be formulated as constraints on the parameters to improve the effectiveness of variable selection and estimation. This paper proposes a variable selection method with linear constraints for the high-dimensional spatial logistic autoregressive model in order to integrate the prior information into the model selection. Monte Carlo experiments are provided to analyze the performance of our proposed method under finite samples. The results show that the method can effectively screen out insignificant variables and give the corresponding coefficient estimates of significant variables simultaneously. As an empirical illustration, we apply our method to land area data.

1. Introduction

Spatial econometrics, developed to deal with spatial correlation and spatial heterogeneity of data, has become a standard analytical tool for spatial data and has begun to enter the mainstream of econometrics. Spatial models have a long history in econometrics. Much progress has been made in the estimation of spatial models, please refer to the special works of Anselin (1988) [1]; LeSage and Pace (2009) [2]. Nowadays, the spatial autoregressive (SAR) model developed by Cliff and Ord in 1973 [3] is the most studied and widely applied modeling method for dealing with spatial correlation. Additionally, the model can be widely applied in many fields including social networks (Ma et al. (2020) [4]), real estate (Dubin (1999) [5]; Osland (2010) [6]), crime incidents (Kakamu et al. (2008) [7]), sampled network data (Zhou et al. (2017) [8]), artificial neural networks (Wang et al. (2019) [9]), and geospatial data (Khalfi et al. (2021)). We can consider the spatial autoregressive model as an ordinary regression model that additively takes the spatial spillover effect of the dependent variable into account. Thus, this model can model both traditional covariates and network dependence simultaneously in a convenient manner.
However, most of the conventional spatial analyses were designed to address the problem of estimation or prediction based on continuous observations. In the case of discrete or binary variables, for instance, in pathological diagnosis, there are only two possible outcomes, positive (denoted as 1) or negative (denoted as 0). The logistic regressive model is a popular method to deal with discrete binary responses. As one of the most popular classification methods, the logistic regressive model has been studied extensively. Essentially, the model assumes that an individual’s class label is influenced by a set of predictors. In practical use, observational data can be taken from different places. In other words, the data generated can be cross-section data. The cross-section data involve several locations. Therefore, it is possible that spatial effect influences the model. In the presence of spatial effects, the usual logistic regression does not sufficiently model the data. Thus, the spatial logistic regression model will be better to model data that contains spatial effects [10]. Theoretical economists and practical researchers are interested in the spatial logistic autoregressive model, which investigates how covariates affect the correlation response of spatial discrete values. The study of spatial logistic regression models, which use categorization technologies to model spatial data, is a relatively new area of spatial econometrics, and research in this area is still quite restricted. Calabrese and Elkink (2014) [11] introduced the binary spatial autoregressive model for the first time. Hilwin Nisa et al. (2019) [12] proposed the spatial logistic regression model which is obtained by the logistic regression model and spatial binary regression model.
High-dimensional spatial data appear frequently in many fields of social life and scientific research, such as biomedical imaging, X-ray tomography, finance, and geoscience. In recent years, a variety of regression methods have been proposed to model high-dimensional data in spatial statistics. For example, Piribauer et al. (2016) [13] proposed a Bayesian variable selection procedure in a spatial autoregressive model. A penalized quasi-maximum likelihood method was put forth by Liu et al. (2018) [14] for variable selection in the spatial autoregressive model. Model selection in spatial autoregressive models with varying coefficients was studied by Wei et al. (2019) [15]. Variable selection for the spatial autoregressive models with a diverging number of parameters was considered by Xie et al. (2020) [16]. Cai et al. (2020) [17] considered variable selection and estimation for a high-dimensional spatial autoregressive model. Li et al. (2020) [18] proposed a variable selection method for the partially linear varying coefficient spatial autoregressive model. More recently, Li et al. (2021) [19] proposed a new variable selection method for a higher-order partially linear spatial autoregressive model with a diverging number of parameters. Liu et al. (2021) [20] studied variable selection for the spatial autoregressive model with autoregressive disturbances. Song et al. (2021) [21] proposed a new robust variable selection method with an exponential squared loss for the spatial autoregressive model.
The above methods mainly focus on the variable selection of continuous response variables based on the penalty regression technique. Penalized regression techniques shrink the insignificant coefficients to 0, which has attracted increasing attention to high-dimensional data analysis, such as least absolute shrinkage and selection operator (LASSO) (Tibshirani (1996) [22]), smoothly clipped absolute deviation (SCAD) (Fan and Li (2001) [23]), and minimax concave penalty (MCP) (Zhang (2010) [24]) for mean regression. LASSO minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant, for which it tends to produce some coefficients that are exactly 0 and hence give interpretable models. However, LASSO has some bias in the estimation of the coefficients. SCAD attempts to mitigate this bias and produce nearly unbiased estimates for large coefficients while still retaining the continuous penalty for sparsity. MCP provides the convexity of the penalized loss in sparse regions to the greatest extent given certain thresholds for variable selection and unbiasedness. However, penalized spatial logistic autoregression has been rarely studied. For high-dimensional spatial data, there are several problems using spatial logistic autoregressive modeling such as endogeneity and including too many variables in the model. First, the spatial lag term in the model will make it endogenous. In the presence of endogeneity, the ordinary least squares (OLS) method can produce biased and inconsistent parameter estimates. Second, as the dimensionality of the variables increases, redundant variables will bring challenges to the estimation in the modeling process.
Furthermore, as the penalized spatial logistic regression does not account for any prior information, and then we can consider how to incorporate prior information into the modeling procedure. Statistical models with linear constraints on variables have gained widespread applications recently. These constraints on regression coefficients reflect the prior information and structure, which can help us to find the optimal parameters with the given information. To incorporate the prior information in the modeling process, we add linear constraints to the penalized spatial logistic autoregressive model. As far as we are aware, no previous research has investigated penalized spatial logistics autoregressive models with linear constraints. Thus, in this paper, we will study a penalized spatial logistic autoregressive model with linear constraints. For the spatial logistic regression model (4), we estimate β by solving the following optimization
min ln [ L ( β , ρ ) ] + 2 n j p p λ β j subject to C β d E β = f
where L ( β , ρ ) is the likelihood function, ρ R , β R p , n is the sample size, p λ ( · ) is the penalty function, and C R q × p , d R q , E R s × p , and f R s are determined concretely according to the experience and knowledge of practical problems. In this paper, our contribution is summarized as follows:
  • Propose a penalized spatial logistics autoregression with linear constraints. These constraints contain the prior information and structure, which can improve the robustness of the model and help us to find the optimal parameters. Thus, the model performs better in the variable selection and estimation under a high-dimensional data space.
  • Provide the formula for degrees of freedom, and then construct the model selection criteria to select the optimal tuning parameter.
  • Simulation results show that the performance of the proposed method is more explanatory and reasonable than penalized quasi-maximum likelihood without linear constraints, and an empirical application illustrates the usefulness of the methods in practical work. The effectiveness of the penalized quasi-maximum likelihood with linear constraints algorithm is demonstrated.
The following is how the paper is arranged. In Section 2, we introduce the general form of the problem we study and present our penalized quasi-maximum likelihood without linear constraints algorithm. The formula for degrees of freedom using Stein’s unbiased risk estimation (SURE) lemma is derived in detail in Section 3. Some Monte Carlos results on the performance of the proposed method are discussed in Section 4. Section 5 shows our method for analyzing real data sets. Section 6 presents the conclusions.

2. Models

2.1. Spatial Autoregressive Model (SAR)

Think of a network that has n nodes. The matrix A R n × n can be used to characterize how the network is structured. Define a i j = 1 when node i follows node j, and a i j = 0 otherwise. With a n × 1 vector of observations on the dependent variable Y and a n × p matrix of regressors X, we can establish the following SAR [3] model:
Y = ρ W Y + X β + ϵ ,
where ρ R is network autocorrelation coefficient and β = ( β 1 , , β p ) T R p is the regression coefficient vector. W is the row-normalized version of A such that w i j = a i j / j = 1 n a i j . Let θ = ( ρ , β T ) T R p + 1 and denote ϵ = ( ϵ 1 , , ϵ n ) T as the error vector of independent disturbances with mean zero and finite variances σ 2 .
Denote G = I ρ W , S = Y ρ W Y X β . Then the SAR model’s log-likelihood function is shown as follows:
ln L ( θ , σ 2 ) = n 2 ln ( 2 π ) n 2 ln σ 2 + ln | G | 1 2 σ 2 S T S .

2.2. Spatial Logistic Regression Model

The spatial logistic regression model is a combination of the spatial autoregressive model and the logistic regression model. Binary classification or multi-classification are both acceptable response variables for a logistic regression model. However, we solely take into account the case of the binary of the response variable.
The model (2) can be written as:
y * = ( I ρ W ) 1 ( X β + ε ) = ( I ρ W ) 1 X β + ( I ρ W ) 1 ε = H X β + e , e M V N ( 0 , Ω )
where M V N denotes the multivariate normal distribution. For simplicity, we use y * instead of Y where H = ( I ρ W ) 1 is an ( n × n ) matrix, and define e = ( I ρ W ) 1 ε as an ( n × 1 ) vector. Latent variable y * has a binary category which is defined as variable y:
y i = 1 , for y i * > 0 0 , for y * 0 .
Therefore, the probability of P y i = 1 and P y i = 0 is:
P y i = 1 X i = P y i * > 0 = P [ H X β ] i + e > 0 = P e [ H X β ] i = 1 1 + exp [ H X β ] i
P y i = 0 X i = P y i * 0 = P [ H X β ] i + e 0 = P e > [ H X β ] i = 1 P e [ H X β ] i = 1 1 1 + exp [ H X β ] i
When we assume the mean value of e is 0 and the variance is Ω , then we get
P y i = 1 = 1 1 + exp [ H X β ] i Ω i i
where Ω i i is the diagonal element of Ω , which is formed as Ω = ( I ρ W ) ( I ρ W ) 1 . The same idea applies to P y i = 0 .
The parameter estimation of spatial logistic regression can be obtained by maximum likelihood estimation (MLE). The parameter is estimated by maximizing the likelihood function of random variable y i , which follows a Bernoulli distribution:
L ( β , ρ ) = i = 1 n 1 1 + exp [ H X β ] i Ω i i y i 1 1 1 + exp [ H X β ] i Ω i i 1 y i
Then, the natural log( l n ) is used to transform the likelihood function as follows:
ln [ L ( β , ρ ) ] = i = 1 n y i ln [ 1 + exp [ H X β ] i Ω i i + i = 1 n 1 y i ln 1 1 [ 1 + exp [ H X β ] i Ω i i
To estimate β and ρ , we use the maximization formula (10), then define
( β ^ , ρ ^ ) = arg max ( β , ρ ) ln [ L ( β , ρ ) ] .

3. Main Results

3.1. Variable Selection with Linear Constraints

In many application fields, prior knowledge can be formulated as constraints on parameters to improve the effectiveness of variable selection and estimation. In this section, we consider the variable selection of the spatial logistic regression model with linear constraints.
We will concentrate on the variable selection for the spatial logistic regression model with linear constraints, that is
min ln [ L ( β , ρ ) ] + 2 n j = 1 p p λ β j s . t C β d , E β = f .
where ρ R , β R p , n is the sample size, and C R q × p , d R q , E R s × p , and f R s are determined concretely according to practical knowledge and experience. p λ ( ) is the penalty function, and the shrinkage degree of the penalty is determined by the tuning parameter λ in the penalty term. There are many popular choices for the penalty function p λ ( ) in the statistics literature: (1) the LASSO penalty with p λ ( t ) = λ | t | ; (2) the SCAD penalty with p λ ( t ) = λ 0 | t | min 1 , ( a t / λ ) + / ( a 1 ) d t , a > 2 where v + denotes its positive part, that is, v I ( v 0 ) ; (3) the MCP with p λ ( t ) = λ 0 | t | ( 1 t / ( λ a ) ) + d t , a > 1 . Penalty functions can provide estimators with three properties which include unbiasedness, sparsity, and continuity according to Fan and Li (2001) [23].
LASSO is not unbiased, and MCP calculation is relatively complex. Fan and Li (2001) [23] demonstrated the oracle properties for the SCAD in the variable selection aspect, and pointed out that the LASSO penalty does not possess the oracle properties. Compared with ridge regression, the SCAD penalty method reduces the prediction variance of the model. Moreover, the SCAD penalty method outperforms the LASSO penalty ones, which reduces the deviation of parameter estimation. Thus, we choose to use the SCAD penalty here.

3.2. Selection of the Tuning Parameter

We decided to choose the SCAD [23] penalty, relying on the analyses mentioned above. The penalty function is defined as:
p λ ( | β | ) = λ β j , 0 β j < λ , β j 2 2 a λ β j + λ 2 / ( 2 a 2 ) , λ β j < a λ , ( a + 1 ) λ 2 / 2 , β j a λ ,
where λ 0 and a > 2 are tuning parameters. Here, a is usually taken to be 3.7, a fact that is elucidated in the work of Fan and Li (2001) [23]. At the same time, they have also shown that λ determines the shrinkage strength of parameter estimation. In this paper, the selection of tuning parameter λ by Bayesian information criterion (BIC) is also related to degrees of freedom.
The number of degrees of freedom measures the number of effective parameters in the regression model and the complexity of the model. It plays an important role in model assessment and selection. There are different ways to measure degrees of freedom. Assume that y follows a distribution y ( μ , σ 2 ) , where μ is the true mean and σ 2 is the variance. Ye ( 1998 ) [25] and Efron ( 2004 ) [26] defined the number of degrees of freedom as
df ( μ ^ ) = 1 σ 2 i = 1 n cov y i , μ ^ i .
where μ ^ ( y ) = y ^ = X * β ^ is the fitted response for y R n .
Under the framework of Stein’s unbiased risk estimation (SURE) theory (Stein (1981) [27]), c o v y i , μ ^ i can be estimated by σ 2 E μ i ^ y i , if μ ^ ( y ) is continuous and almost differentiable. Then the expression for degrees of freedom of fitted μ ^ can be calculated as
df ( μ ^ ) = E i = 1 n μ ^ i y i .
to apply it. We need to assume that the response is normally distributed, that is, y N μ , σ 2 I n .
Degrees of freedom are used effectively while selecting the tuning parameter λ . In this study, the model selection criteria will be based on the Bayesian information criterion (BIC) (Schwarz (1978) [28]):
BIC ( λ ) = ln 1 n i = 1 n δ τ y i t ^ i , λ + ln n n d f t ^ λ .
Since we have calculated the corresponding ρ ^ ( λ ) and β ^ ( λ ) at each tuning parameter λ , we can use the corresponding fitted value t ^ ( λ ) and degrees of freedom d f ( t ^ ( λ ) ) to select the optimal λ that minimizes BIC ( λ ) .

4. Simulation Studies

4.1. Simulation Experiment Design

In the simulation experiment, we test the performance of the model through Monte Carlo simulation. The random sample is generated by model (2.1) combined with model (2.7), in which the covariate is considered when the ( q + 3 ) -dimensional normal distribution with zero mean and covariance matrix σ i j , where σ i j = 0 . 5 | i j | . Therefore, X is an n × ( q + 3 ) matrix. In the following simulation, we set the number of samples n { 60 , 90 , 120 } , and the number of insignificant covariates q { 5 , 10 , 25 } . In this paper, we show the cases of q = 5 and q = 10 in the simulation results.
In the spatial autoregressive model, the network autocorrelation coefficient ρ is generated by the uniform distribution on the interval ρ 1 1 , ρ 1 + 1 , where ρ 1 { 0.8 , 0.5 , 0.2 } . Define the spatial weight matrix W = I R B m , where B m = ( 1 / ( m 1 ) ) 1 m · 1 m T I m , and ⊗ is the Kronecker product. Denote 1 m as the m-dimensional column vector of 1. In the simulation experiment, we consider m = 3 , and differet values of R, where R { 20 , 30 , 40 } . The regression coefficients are set to β = β 1 , β 2 , β 3 , 0 q T , where β 1 , β 2 , β 3 is generated from the 3-dimensional normal distribution with the mean vector of ( 3 , 2 , 1.6 ) , and the covariance matrix is 0.001 I 3 .
To simplify the calculation, the regression coefficients are set to 3 , 2 , 1.6 , 0 q T , where 0 q is a q-dimensional zero vector. The response variable is given by the following formula:
y * = I n ρ W 1 X β + ε n
Then use the following formula to convert the response variable into a category variable:
Y i = 1 , for y i * > 0 0 , for y i * 0 .
Thus, we can obtain the binary classification of the response variable Y. For purpose of confirming the robustness of the model, the error terms ε i s are independently generated from the following two distributions: (a) normal distribution ε i N 0 , σ 2 I n , denoted as ε 0 and (b) Gaussian mixture distribution ε i 0.5 N 1 , 2.5 2 + 0.5 N 1 , 0.5 2 , denoted as ε 1 . σ 2 is generated by the uniform distribution on the interval σ 1 0.1 , σ 1 + 0.1 , where σ 1 { 1 , 2 } . In the simulation experiment, we set σ = 1.5 .
In order to verify that the effect of the model with linear constraints is better, we will compare it with the model without constraints. In one case, we can set the constraints as:
β 3 + β 6 = 1.6 β 1 + β 5 = 3 β 1 + β 3 4 β 2 + β 6 2.5
Obviously, we can find C, d, E, and f.

4.2. Evaluation Indicators

According to the above simulation experiments, we set the number of Monte Carlo repetitions at 1000. We define the following three indicators to evaluate the performance of variable selection in different settings.
Correct: the average number of coefficients of the true zeros correctly set to zero;
Incorrect: the average number of coefficients of the true nonzeros incorrectly set to zero;
ME: the mean error between the true value and the estimate, which is defined by:
1 1000 i = 1 1000 θ i θ ^ i 1

4.3. Simulation Results

Table 1 and Table 2 show the results of models with linear constraints without linear constraints, respectively. The constrained model is recorded as "Const”, and the unconstrained model is recorded as “Unconst”. The results in Table 1 clearly show that the spatial logistic model performs better with linear constraints, which also confirms the effectiveness of our model. Most significantly, when ρ 1 = 0.8 , the error of the models with constraints and without constraints are very different, which shows that when the spatial effect is strong the constraints can greatly improve the accuracy of model parameter estimations. Moreover, we find that the effect of the model tends to become bad by increasing the network autocorrelation coefficient ρ 1 , which indicates the importance of the spatial effect. Moreover, by setting two types of errors, we observe that the model has good robustness. Additionally, in most cases, the model with an error term of ε 1 performs better than the one with an error term of ε 0 . Similarly, with the increase in sample size n, the incorrect rate of variable selection and the estimation error both decrease. This situation is in line with our prediction of the effect of the model. Moreover, it can be seen in Table 1 that the mean error is minimum when n = 120, ρ 1 = 0.2, and ε = ε 1 , which also confirms our analysis above.
In Table 2, by increasing the dimension to q = 10, we find that with the increase in the network autocorrelation coefficient ρ 1 , the effect of the model becomes bad. Compared to the case of q = 5, we find an increase in the mean error of estimation. After analysis, the possible reason is that the proportion of data size to dimension is bad, that is, the dimension of the sample is higher than the sample size. suppose the network complexity increases, it will have a certain negative impact on the effect of variable selection. For high-dimensional samples, the model results can be optimized by increasing the sample size. Simultaneously, according to the simulation results, compared with the unconstrained case, the spatial logistic model has stronger robustness, higher accuracy, and lower estimation error rate.
We compare the variable selection using the SCAD penalty and LASSO penalty under constraints. The simulation results are shown in Table 3. We can clearly observe that in the case where ρ 1 = 0.2 and ρ 1 = 0.5 , the performance of the SCAD penalty is significantly better than the LASSO penalty, which is shown in higher correct selection rate and lower estimation error. As the network autocorrelation coefficient increases to ρ 1 = 0.8 , the correct rate of variable selection of the SCAD penalty is higher than the LASSO penalty, while the incorrect rate of the LASSO penalty is lower than the SCAD penalty. Additionally, the estimation error between them is not much different. The reason may be that the tuning parameter λ in the LASSO penalty is too large and the shrinkage strength is stronger.

5. Real Data Example

In this section, we provide a real-world example to demonstrate the performance of the variable selection procedure proposed in this paper for spatial logistic regression models with linear constraints.

5.1. The Land Area Utilization Data

Land area utilization is analyzed by the spatial logistic model. The data set is different types of land area data from 48 states in the United States from 1954 to 2012 (recorded every five years). The dependent variables are binary, with “1” denoting a low land utilization rate, which means that most of the land has not been properly developed, and “0”, denoting a high land utilization rate, which means that most of the land has been efficiently developed and exploited. As for the independent variables, there are eight properties, which are Cropland used for crops, Cropland used for pasture, Cropland idled, Grassland pasture and range, Forest-use land grazed, Land in rural transportation facilities, Land in urban areas, and Other idle land (shown in Table 4).

5.2. Variable Selection and Estimation

For the above land area utilization data sets, we constructed a spatial logistic autoregressive model. We use the land utilization rate as a response variable, and take eight variables, Cropland used for crops, Cropland used for pasture, Cropland idled, Grassland pasture and range, Forest-use land grazed, Land in rural transportation facilities, Land in urban areas, and Other idle land, as the independent variables.
According to theoretical knowledge, the idle area of cropland and other idle land areas have a significant impact on the probability of classification results. Through the fitting of the model, the following two cases are considered: the first is the parameter estimation without constraints, and the other is the parameter estimation with linear constraints, as shown in Table 5 and Table 6.
According to the results of parameter estimation based on the spatial logistic model, it is found that the performance of variable selection is not obvious under unconstrained conditions. Among them, “Forest-use land grazed” and “Land in rural transportation facilities” have little impact on land use efficiency, which can be almost ignored. However, “Other idle land” has a great influence on the classification effect. Considering the greater relationship between land use area and idle land area, the performance of model selection is greatly improved with constraints. According to the table below, it is found that “Cropland idle” and “Other idle land” have a great impact on the classification results, and the fitting parameters of other attributes are very small and can be ignored.

6. Conclusions

In the paper, we obtain a spatial logistic model from the spatial autoregressive model (SAR) and logistic regression model. In order to improve the accuracy of the model, we take the prior information into account, and finally, propose a variable selection method with linear constraints for the spatial logistic model. According to the simulation results, by comparing the constrained and unconstrained variable selection models, we find that the performance of variable selection is more stable with the increase in sample size in the case of limited samples. When we fix the sample size at a constant value, the performance of the model tends to improve with the increase in network complexity. At the same time, the model has strong robustness to noise. In order to verify the superiority of the SCAD penalty, we compare the performance of the SCAD and LASSO penalties in the case of the linear constraint model and find that the SCAD penalty has a better effect on variable selection.
In practical application, most data show the characteristics of small sample size and high dimensions. For purpose of verifying the wide adaptability of the model, we verify that the proposed model can be effectively applied to the data set of high-dimensional and small samples through simulation experiments. However, we find that in this case, when the network complexity is very high, the performance effect of the model is not very good. The sample size being too small might be the cause of this.

Author Contributions

Methodology, Y.S. (Yunquan Song); Software, Y.S. (Yuqi Su); Writing—original draft, Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

The researches are supported by the National Key Research and Development Program of China (2021YFA1000102), NSF project (ZR2019MA016) of Shandong Province of China.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Anselin, L. Spatial Econometrics: Methods and Models; Kluwer Academic: Dordrecht, The Netherlands, 1988. [Google Scholar]
  2. LeSage, J.; Pace, R.K. Introduction to Spatial Econometrics; Chapman & Hall: New York, NY, USA, 2009. [Google Scholar]
  3. Ord, K. Estimation methods for models of spatial interaction. J. Am. Stat. Assoc. 1975, 70, 120–126. [Google Scholar] [CrossRef]
  4. Ma, Y.; Pan, R.; Zou, T. A Naive Least Squares Method for Spatial Autoregression with Covariates. Stat. Sin. 2020, 30, 653–672. [Google Scholar]
  5. Dubin, R.; Pace, K.; Thibodeau, T. Spatial autoregression techniques for real estate data. J. Real Estate Lit. 1999, 7, 79–95. [Google Scholar] [CrossRef]
  6. Osland, L. An application of spatial econometrics in relation to hedonic house price modelling. J. Real Estate Res. 2010, 32, 289–320. [Google Scholar] [CrossRef]
  7. Kakamu, K.; Polasek, W.; Wago, H. Spatial interaction of crime incidents in Japan. Math. Comput. Simul. 2008, 78, 276–282. [Google Scholar] [CrossRef] [Green Version]
  8. Zhou, J.; Tu, Y.; Chen, Y.; Wang, H. Estimating Spatial Autocorrelation with Sampled Network Data. J. Bus. Econ. Stat. 2017, 35, 130–138. [Google Scholar] [CrossRef]
  9. Wang, W.; Andrews, B. Partially Specified Spatial Autoregressive Model with Artificial Neural Network. arXiv 2019, arXiv:1801.07822. [Google Scholar] [CrossRef]
  10. Madsen, H.; Thyregod, P. Introduction to General and Generalized Linear Models; CRC Press: New York, NY, USA, 2010. [Google Scholar]
  11. Calabrese, R.; Elkink, J.A. Estimators of Binary Spatial Autoregressive Models: A Monte Carlo Study. J. Reg. Sci. 2014, 54, 664–687. [Google Scholar] [CrossRef] [Green Version]
  12. Nisa, H.; Mitakda, M.B.; Astutik, S. Estimation of propensity score using spatial logistic regression. IOP Conf. Ser. Mater. Sci. Eng. 2019, 546, 052048. [Google Scholar] [CrossRef]
  13. Piribauer, P.; Crespo Cuaresma, J. Bayesian Variable Selection in Spatial Autoregressive Models. Spat. Econ. Anal. 2016, 11, 457–479. [Google Scholar] [CrossRef] [Green Version]
  14. Liu, X.; Chen, J.; Cheng, S. A penalized quasi-maximum likelihood method for variable selection in the spatial autoregressive model. Spat. Stat. 2018, 25, 86–104. [Google Scholar] [CrossRef]
  15. Wei, H.; Sun, Y.; Hu, M. Model Selection in Spatial Autoregressive Models with Varying Coefficients. Front. Econ. China 2019, 13, 559–576. [Google Scholar]
  16. Xie, T.; Cao, R.; Du, J. Variable selection for spatial autoregressive models with a diverging number of parameters. Stat. Pap. 2020, 61, 1125–1145. [Google Scholar]
  17. Cai, L.; Maiti, T. Variable selection and estimation for high-dimensional spatial autoregressive models. Scand. J. Stat. 2020, 47, 587–607. [Google Scholar] [CrossRef]
  18. Li, T.; Yin, Q.; Peng, J. Variable selection of partially linear varying coefficient spatial autoregressive model. J. Stat. Comput. Simul. 2020, 90, 2681–2704. [Google Scholar] [CrossRef]
  19. Li, T.; Kang, X. Variable selection of higher-order partially linear spatial autoregressive model with a diverging number of parameters. Stat. Pap. 2021, 63, 243–285. [Google Scholar] [CrossRef]
  20. Liu, X.; Chen, J. Variable Selection for the Spatial Autoregressive Model with Autoregressive Disturbances. Mathematics 2021, 9, 1448. [Google Scholar] [CrossRef]
  21. Song, Y.; Liang, X.; Zhu, Y.; Lin, L. Robust variable selection with exponential squared loss for the spatial autoregressive model. Comput. Stat. Data Anal. 2021, 155, 107094. [Google Scholar] [CrossRef]
  22. Tibshirani, R.J. Regression shrinkage and selection via the LASSO. J. R. Stat. Soc. Ser. B 1996, 58, 267–288. [Google Scholar] [CrossRef]
  23. Fan, J.; Li, R. Variable selection via nonconcave penalized likelihood and its oracle properties. J. Am. Stat. Assoc. 2001, 96, 1348–1360. [Google Scholar] [CrossRef]
  24. Zhang, C.H. Nearly unbiased variable selection under minimax concave penalty. Ann. Statist. 2010, 38, 894–942. [Google Scholar] [CrossRef] [Green Version]
  25. Ye, J. On ineasuring and correcting the effects of data mining and model selection. J. Am. Statist. Assoc. 1998, 93, 120–131. [Google Scholar] [CrossRef]
  26. Efron, B. The estimation of prediction error: Covariance penalties and cross-validation. J. Am. Statist. Assoc. 2004, 99, 619–632. [Google Scholar] [CrossRef]
  27. Stein, C.M. Estimation of the mean of a multivariate normal distribution. Ann. Statist. 1981, 9, 1135–1151. [Google Scholar] [CrossRef]
  28. Schwarz, G. Estimating the dimension of a model. Ann. Statist. 1978, 6, 461–464. [Google Scholar] [CrossRef]
Table 1. Simulation results of the SCAD penalty model with constraints and without constraints ( q = 5 ) .
Table 1. Simulation results of the SCAD penalty model with constraints and without constraints ( q = 5 ) .
Method n = 60, q = 5n = 90, q = 5n = 120, q = 5
ConstUnconstConstUnconstConstUnconst
ρ 1 = 0.2 Correct4.7034.6374.7554.8094.7894.812
ε 0 Incorrect00.18900.09900.068
ME0.3234.0520.2903.0130.2732.542
ρ 1 = 0.2 Correct4.7804.6344.7834.7824.7844.837
ε 1 Incorrect00.11400.05000.032
ME0.3145.6510.2884.7890.2623.959
ρ 1 = 0.5 Correct4.6824.4214.7524.5894.7014.664
ε 0 Incorrect00.13500.06200.044
ME0.4026.9650.3315.4340.3194.995
ρ 1 = 0.5 Correct4.6924.3854.6754.5484.7004.672
ε 1 Incorrect00.08400.05000.029
ME0.3749.3350.3387.7350.3197.073
ρ 1 = 0.8 Correct3.8942.9984.1013.6194.1793.809
ε 0 Incorrect0.0750.0650.0210.0180.0180.015
ME2.36245.9411.44229.5961.09323.867
ρ 1 = 0.8 Correct3.8652.8374.1263.5424.1743.876
ε 1 Incorrect0.0740.0520.0160.0140.0280.007
ME2.24351.8301.36535.2131.09229.309
Table 2. Simulation results of the SCAD penalty model with constraints and without constraints ( q = 10 ) .
Table 2. Simulation results of the SCAD penalty model with constraints and without constraints ( q = 10 ) .
Method n = 60, q = 10n = 90, q = 10n = 120, q = 10
ConstUnconstConstUnconstConstUnconst
ρ 1 = 0.2 Correct9.2558.7409.3749.3639.5719.614
ε 0 Incorrect00.16600.07400.036
ME0.3833.7130.3173.0720.2812.487
ρ 1 = 0.2 Correct9.7779.2829.8619.1789.6429.546
ε 1 Incorrect00.10200.02100.015
ME0.3285.6790.2934.6970.2624.134
ρ 1 = 0.5 Correct8.0367.2208.3447.9708.5988.298
ε 0 Incorrect00.13300.05200.022
ME0.6716.6410.4935.2720.4464.853
ρ 1 = 0.5 Correct9.5716.6328.3958.1608.6908.576
ε 1 Incorrect00.11700.03400.009
ME0.4149.1010.4887.2170.4196.829
ρ 1 = 0.8 Correct5.6884.0495.8895.2046.2219.755
ε 0 Incorrect0.1290.0700.0710.0200.0802.937
ME10.52257.6966.46534.6755.0396.697
ρ 1 = 0.8 Correct5.6694.0466.0195.1526.1949.759
ε 1 Incorrect0.1340.0430.0760.0230.0792.935
ME9.90157.5846.34537.5094.6276.690
Table 3. Simulation results of the constrained model with the SCAD penalty and LASSO penalty.
Table 3. Simulation results of the constrained model with the SCAD penalty and LASSO penalty.
Method n = 60, q = 5n = 90, q = 5n =120, q = 5
SCADLASSOSCADLASSOSCADLASSO
ρ 1 = 0.2 Correct4.7033.4964.7553.4954.7893.483
ε 0 Incorrect000000
ME0.3232.1820.2902.1620.2732.152
ρ 1 = 0.2 Correct4.7803.4824.7833.4724.7843.397
ε 1 Incorrect000000
ME0.3142.1570.2882.1460.2622.096
ρ 1 = 0.5 Correct4.6823.6334.7523.6464.7013.625
ε 0 Incorrect000000
ME0.4022.2270.3312.2120.3192.219
ρ 1 = 0.5 Correct4.6923.6054.6753.6834.7003.624
ε 1 Incorrect000000
ME0.3742.2100.3382.2220.3192.221
ρ 1 = 0.8 Correct3.8943.5164.1013.5164.1793.509
ε 0 Incorrect0.07500.02100.0180
ME2.3622.3271.4422.2851.0932.307
ρ 1 = 0.8 Correct3.8653.4784.1263.5384.1743.485
ε 1 Incorrect0.07400.01600.0280
ME2.2432.3481.3652.2741.0922.272
Table 4. Summary of predictor variables for the land area utilization model.
Table 4. Summary of predictor variables for the land area utilization model.
Variable NameDescription
CLand_CCropland used for crops
CLand_PCropland used for pasture
CLand_ICropland idled
Grass_PGrassland pasture and range
Land_GForest-use land grazed
Land_TLand in rural transportation facilities
Land_ULand in urban areas
Land_IOther idle land
Table 5. Parameter estimation results without constraints.
Table 5. Parameter estimation results without constraints.
YearCLand_CCLand_PCLand_IGrass_PLand_GLand_TLand_ULand_I
1954000.9820.4230000.602
195901.7000−1.740000.30501.27
196401.280−0.6150000.892
196901.550.715−1.490−0.38501.06
19740.8541.521.39−2.260−0.596−0.6511.52
19780.66301.90−1.72−0.249−0.2870.6141.26
19820.3021.451.12−1.1200−0.9841.14
198701.150.664000−1.111.25
19920.98100.1742.820−0.949−2.411.80
19970.56800.2622.750−0.877−1.931.87
20022.4201.61−1.210−1.0002.39
20071.300.8521.76−1.2000−0.3832.47
2012002.58−0.4680−0.79901.02
Table 6. Parameter estimation results with linear constraints.
Table 6. Parameter estimation results with linear constraints.
YearCLand_CCLand_PCLand_IGrass_PLand_GLand_TLand_ULand_I
1954000.9820.4190000.604
1959000.660−0.2980.600−0.46401.16
19640000.6630000.691
1969001.03−0.3.160.311−0.45200.814
1974001.74−1.210.640−0.63901.10
1978001.83−0.7410001.41
1982000.92200.604−0.49900.884
1987000.90300.585−0.54700.945
1992000.5140.5960.447−0.54500.981
1997000.5860.7700.408−0.61001.05
2002001.4800002.24
2007002.0500001.04
2012001.7600000.694
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Song, Y.; Su, Y.; Wang, Z. Variable Selection of Spatial Logistic Autoregressive Model with Linear Constraints. Entropy 2022, 24, 1660. https://doi.org/10.3390/e24111660

AMA Style

Song Y, Su Y, Wang Z. Variable Selection of Spatial Logistic Autoregressive Model with Linear Constraints. Entropy. 2022; 24(11):1660. https://doi.org/10.3390/e24111660

Chicago/Turabian Style

Song, Yunquan, Yuqi Su, and Zhijian Wang. 2022. "Variable Selection of Spatial Logistic Autoregressive Model with Linear Constraints" Entropy 24, no. 11: 1660. https://doi.org/10.3390/e24111660

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop