Next Article in Journal
Experimental Study on the Shear Performance of Reinforced Concrete Beams Strengthened with Bolted Side-Plating
Next Article in Special Issue
Acquiring and Geo-Visualizing Aviation Carbon Footprint among Urban Agglomerations in China
Previous Article in Journal
Assessing Airbnb Logistics in Cities: Geographic Information System and Convenience Theory
Previous Article in Special Issue
Spatial Dynamics of Intercity Technology Transfer Networks in China’s Three Urban Agglomerations: A Patent Transaction Perspective
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Integrating Cellular Automata with Unsupervised Deep-Learning Algorithms: A Case Study of Urban-Sprawl Simulation in the Jingjintang Urban Agglomeration, China

1
College of Land Science and Technology, China Agriclutural University, No.17 Tsing Hua East Road, Haidian District, Beijing 100083, China
2
Institute of Electronics, China Academy of Sciences, No.19 North Fourth Ring West Road, Haidian District, Beijing 100190, China
*
Author to whom correspondence should be addressed.
Sustainability 2019, 11(9), 2464; https://doi.org/10.3390/su11092464
Submission received: 20 March 2019 / Revised: 14 April 2019 / Accepted: 23 April 2019 / Published: 26 April 2019

Abstract

:
An effective simulation of the urban sprawl in an urban agglomeration is conducive to making regional policies. Previous studies verified the effectiveness of the cellular-automata (CA) model in simulating urban sprawl, and emphasized that the definition of transition rules is the key to the construction of the CA model. However, existing simulation models based on CA are limited in defining complex transition rules. The aim of this study was to investigate the capability of two unsupervised deep-learning algorithms (deep-belief networks, DBN) and stacked denoising autoencoders (SDA) to define transition rules in order to obtain more accurate simulated results. Choosing the Beijing–Tianjin–Tangshan urban agglomeration as the study area, two proposed models (DBN–CA and SDA–CA) were implemented in this area for simulating its urban sprawl during 2000–2010. Additionally, two traditional machine-learning-based CA models were built for comparative experiments. The implementation results demonstrated that integrating CA with unsupervised deep-learning algorithms is more suitable and accurate than traditional machine-learning algorithms on both the cell level and pattern level. Meanwhile, compared with the DBN–CA, the SDA–CA model had better accuracy in both aspects. Therefore, the unsupervised deep-learning-based CA model, especially SDA–CA, is a novel approach for simulating urban sprawl and also potentially for other complex geographical phenomena.

1. Introduction

Urban areas have been dominant players in the world’s socioeconomic, political, cultural, and environmental spheres, and the global shift from rural to urban living has been a defining trend. In the mean time, as the most significant land-use change processes, the urban sprawl has had an important impact on Earth’s surface, ecosystem, and environmental sustainability, and is closely related with the life of almost half of the world’s population [1]. In this context, urban-sprawl simulations have played a key role in understanding its spatial-evolution process and has became a powerful tool for supporting urban planning and sustainable development in urban agglomeration. In recent decades, several models were proposed to better understand and forecast urban-sprawl processes. Models evolved from early empirical statistical models or aspatially oriented models that are based on equations to reach static status in the current spatial dynamic, to spatially oriented or integrated models that are based on the principle of spatial interaction. Comparing the two main models, the major difference is that the former models just predict population and land demand in the future from the perspective of economics or statistics, while the latter models further consider the spatial interaction in the process of urban sprawl. Among them, cellular automata (CA) are examples of mathematical systems constructed from many identical components, each simple, but together capable of complex behavior [2]. Because of the simplicity, flexibility, and intuitiveness of CA [3], it has been widely adopted as a typical spatial dynamic model to simulate urban sprawl.
When we use CA as a “bottom–up” model to simulate a complex geographical phenomenon like urban sprawl, the key step is how to define the transition rules of CA that determine the state conversion of geographical processes [4]. With many driving factors involved, there exists a nonlinear relationship between driving factors and geographical structure, and it is difficult to obtain appropriate transition rules. A variety of mathematical methods were used to obtain the appropriate transition rules of CA. Although a traditional method like multicriteria evaluation [5,6,7] is simple and its mechanism is clear, it is still not efficient and reliable because the determination of parameters has a certain subjectivity and randomness. It is also difficult to tackle a series of complex behaviors associated with natural systems. To overcome this problem, a series of machine-learning algorithms have been proposed, assuming that the historic geographical processes remain stable for a certain period in the future through local interaction between cells, such as logistic regression [8,9,10], artificial neural networks [11,12,13,14], support vector machines [15,16,17], decision trees [18,19], random forests [20,21], genetic algorithms [22,23], and swarm-intelligence algorithms [24,25,26,27,28]. Although the algorithms above show significant improvement in defining nonlinear transition rules for CA, there still remain many problems like overfitting, easily resulting in local optima and weak in global searching. Moreover, the geographical simulation of large-scale regions with fine-resolution units has become an inevitable trend [29], which further makes the implementation of big-sample-oriented CA simulations difficult with these methods.
In recent years, as a feature-learning method that can yield more nonlinear, more abstract representations [30], deep learning has become the dominant technique to learning more information from multivariable nonlinear systems in the era of Big Data. Compared to traditional machine-learning algorithms, deep-learning-based methods attempt to model high-level abstractions in data by using multiple processing layers with complex structures, resulting in better representations from the point of view of simplifying a learning task from input examples [31]. A deep architecture consists of feature-detector units arranged in multiple layers: lower layers detect simple features and feed into higher layers, which in turn detect more complex features [32]. In particular, two typical unsupervised deep-learning algorithms for training, deep-belief networks (DBN) [33] and stacked denoising autoencoders (SDA) [34], are regarded as a breakthrough in feature learning due to their effective training strategies. They are all based on a similar central idea: greedy layerwise unsupervised pretraining, followed by supervised fine-tuning [35]. Specifically, this idea provides a better approach to (pre)train each layer in turn, initially using a local unsupervised criterion [36] with the aim of learning to produce useful higher-level representations from lower-level-representation output of the previous layer, which leads to much better solutions in terms of generalization performance. Due to such characteristics, DBNs and SDAs were successfully implemented in many nonlinear systems like dimensionality reduction [37,38,39], time-series forecasting [40,41,42], acoustic modeling [43,44,45], and digit recognition [46,47,48]. Therefore, we think the above-mentioned algorithms also have the potential to be applied in urban-sprawl simulations.
In summary, urban-sprawl simulation is crucial to understand and assess the sustainable development of urban land use changes. In defining the transition rules of CA, current studies tend to introduce many varied machine-learning algorithms to tackle the difficulty in automatically extracting nonlinear relationship between driving factors and spatial dynamic processes. However, most traditional machine-learning-based CA models are small-sample-problem-oriented approaches by using feature mapping, which is not suitable for representation learning in the era of Big Data. Therefore, we applied two typical unsupervised deep-learning algorithms (DBN and SDA) for representation learning from a complex geographical phenomenon like urban sprawl in this study. In addition, we noticed that there were a few studies [49,50,51] that applied deep learning to modeling geographical phenomena, but only when comparing the proposed deep-learning algorithm with traditional machine-learning algorithms, without a comparison of a series of similar deep-learning algorithms in the same case. No relevant models have been applied to large-scale cases such as urban agglomerations. Therefore, the purpose of this study is to incorporate unsupervised deep-learning algorithms into CA, making it possible to more accurately simulate urban sprawl. First, we integrated DBN and SDA, two typical unsupervised deep-learning algorithms, into the CA. Then, we selected the Beijing–Tianjin–Tangshan (JJT) urban agglomeration to test the proposed models because of its dramatic land-use changes and rapid urbanization in the past two decades. Finally, we compared the performance of the unsupervised deep-learning algorithms between themselves, and with the obtained results under traditional machine-learning algorithms like an artificial neural network (ANN) and logistic regression (LR) by using the figure of merit (FoM), the hit-miss-false alarm approach, and a series of landscape indices.

2. Methodology

2.1. DBN

A DBN is a typical multilayer generative neural network, and it exhibits particular strength for learning representations of large amounts of data with multiple levels of abstraction [52]. As shown in Figure 1a, a typical DBN architecture contains an unsupervised learning subpart by using Restricted Boltzmann Machines (RBMs), which is trained in a greedy manner, followed by a supervised fine-tuning subpart in the top level for prediction. The basic idea of DBN is to use an unsupervised learning method to train each RBM layer by layer, and finally fine-tune the whole network with supervised learning.
As the basic DBN component, the RBM is in essence a stochastic neural network and an energy-based model. It is generally regarded as a kind of unsupervised learning algorithm to extract features from high-dimensional data by estimating their probability distribution. The structure of a typical RBM is shown in Figure 1b. It is a feed-forward graph network with two layers, the so-called visible layer (or input layer) of the i dimension, representing observable data, and the hidden (or output) layer of the j dimension, representing extracted features from observable data. Moreover, it allows full connection within the different layers, but no connections within the same layer. The purpose of unsupervised learning by RBM is to continuously adjust the connection weight and bias parameters through Gibbs sampling, so as to minimize the error between v and v after the restoration and reconstruction of implicit characteristic signal h. Given the difficulty of determining the step size for calculating the slope of a probability change, an algorithm called contrastive divergence (CD) [53] is proposed to speed up the RBM training process and maintain accuracy.
From Figure 1b, we can see that each unit of the visible layer has a symmetric connection weight with the corresponding units of the hidden layer. Weight matrix W(size: m × n ) encodes a statistical relationship by the joint distribution between the visible and the hidden layer, which can be mathematically described as the following equations:
p ( v , h ) = e E ( v , h ) v , h e E ( v , h )
where v is the visible vector, and h is the hidden vector. If the visible units are binary-valued, energy function E(v,h) of certain configuration can be defined as equation
E ( v , h ) = i = 1 m a i v i j = 1 n b i h i i = 1 m j = 1 n h i W i j v j
where a i , b j are the biases of the visible and hidden layer, respectively, and W i j is the combined weights.
In summary, an RBM contains five parameters: v, h, W, a, and b. Except for v being the visible (input) vector and h the hidden (output) vector, W, a, and b can be learned and updated by computing the extract gradient of the log probability of the training data. According to the CD algorithm, the updating rule for W, a, and b is as follows:
Δ W i j = ε ( < v i h j > d a t a < v i h j > r e c o n s t r u c t i o n )
Δ a i = ε ( < v i > d a t a < v i > r e c o n s t r u c t i o n )
Δ b j = ε ( < h j > d a t a < h j > r e c o n s t r u c t i o n )
where ε is the learning rate, < > d a t a is the expectation of the observed data, and < > r e c o n s t r u c t i o n is the expectation under the reconstructed visible and hidden units, respectively, at every iteration.
To overcome the limited learning ability of the single-layer network, the DBN is proposed as a multilayer learning structure that is built in the form of stacks using individual restricted Boltzmann machines. The DBN training process can be divided into two main stages as follows [54]:
  • Unsupervised pretraining stage: (1) Train the bottom RBM with CD on training data v being the visible units; (2) freeze weights matrix W and bias a, b of the first RBM, and the state of the hidden units is inferred and used as the input of the next higher RBM; (3) next higher RBM is stacked on top of the previous lower RBM after training; (4) iterate Steps 2 and 3 for the desired number of layers, each time upward propagating either samples or mean values.
  • Supervised fine-tuning stage: After an unsupervised pretraining stage, all parameters need to be slightly adjusted in supervised manner until DBN loss function reaches its minimum. In this paper, a logistic regression layer periodically works in the top-level RBM during the supervised fine-tuning stage.

2.2. SDA

Similar to DBN, which consists of a series of RBMs, an SDA is established based on a series of denoising autoencoders (DAEs; see Figure 2a) and learns to use a deep network architecture in a layer-by-layer fashion. The key function of SDA is unsupervised pretraining; once each layer is pretrained for feature selection and the extraction of input from the previous layer, the second phase of supervised fine-tuning can occur.
The autoencoder (AE) learns from a distributed representation (encoding) for a set of training data, and then reconstructs the data back to themselves from the encoder (decoding). However, it simply retains the information of the original input during the learning process, thereby failing to ensure that useful feature representation is extracted for the input data. To avoid this, the DAE adds some statistical noise to the original input and “corrupts” the original input, so that the autoencoder not only learns the characteristics of the original data, but also learns degraded features after being “corrupted”, which greatly improved the overfitting problem and the generalization ability of the AE.
From Figure 2b, we can see that a DAE consists of an encoder, a decoder, and a hidden layer. It is trained to reconstruct a clean “repaired” input from a “corrupted” version. First, add statistical noise to original input x to obtain partially “corrupted” data x ˜ by a stochastic mapping:
x ˜ q D ( x | x )
where D represents the dataset. After being corrupted, the input to the encoder is x after the original data are corrupted, and hidden layer y is computed by
y = f θ ( x ˜ ) = s ( W x ˜ + b )
where s represents a nonlinear mapping function; W is the weight matrix; b is the bias vector; and θ is the collection parameter for the mapping. Next, the reconstruction layer is generated by
z = g θ ( y ) = s ( W y + b )
where W is the inverse weight matrix; b is the inverse bias vector; and θ is the collection parameter for the inverse mapping. Finally, reconstruction loss L measured by the squared in:
L ( x , x ˜ ) = x x ˜ 2
The training process of a DAE aims at minimizing the summation of squared errors over all training data. After minimizing the overall squared loss function, DAEs are trained.
In summary, an SDA is formed by stacking several DAEs through a layer-by-layer overlay mechanism similar to DBN. Its training mechanism is essentially a layer-by-layer greedy training mechanism that is also divided into two stages: unsupervised pretraining and supervised fine-tuning. The details are as follows:
  • Unsupervised pretrain stage: (1): Train the bottom DAE through the above steps, and an encoder f θ is obtained when the first layer is trained; (2) the feature representation vector is obtained by this encoder f θ on the original input data and regarded as the hidden layer vector, which is used to obtain and train the encoder of the second SDA layer; (3) iterate Steps 1 and 2 for the desired number of SDA layers.
  • Supervised fine-tuning stage: When the entire pretraining stage is over, the top layer is the final output layer. With this output as the base layer for logistic regression errors throughout the SDA structure, fine-tuning global parameters are adjusted.

2.3. Proposed Geographical CA Model

A standard CA consists of a cell space and a transition function that defines the space [55]. The basic CA components are the state of the cell, cell space, neighborhood, transition rules, and discrete time. So, a cellular automaton can be described by the following five-tuple model:
C A = S , L , N , R , T
where S is the state of the cell, all cells are mutually discrete, and a cell can only have one state at a certain time, which is taken from a finite set; L represents the lattice, the cell that is distributed in the set of spatial sites, and the commonly used two-dimensional cellular automaton includes a triangular mesh, a square mesh, and a hexagon mesh; N describes the spatial neighborhood of the cell, which is a collection of cells delimited by a certain shape around the cell, and it can affect the state of the cell at the next moment; R represents the transition rules, which are the core of the cellular automaton, and it expresses the logical relationship of the simulated process and determines the result of a spatial change; Time interval T is an integer value and may be equally spaced, and assuming that the initial moment is T 0 , then the state of a cell at time T 0 +T depends on the state of the cell and its neighbor cells at time.
Geographical CA includes geospatial-change information acquisition, global spatial-variable acquisition, transition-rule definition, model construction, and process simulation [11]. Change-information acquisition and global spatial-variable acquisition are usually done by GIS/RS tools. Model construction and process simulation are usually done by GIS tools or an open-source programming framework. Transition rules are a key part of building a model and directly determining the quality and effectiveness of a geographical CA. For an urban sprawl, the cell’s state is guided by transition rules, which reflect the complexity of the environment. These rules act as a link between spatial patterns and the underlying spatial process [56]. This is deeply affected by a series of driving factors like the initial state of a cell, neighborhood conditions, global spatial variables, and some constraints on urban development. Therefore, an urban-sprawl simulation model can be formulated as follows:
S i j T 0 + T = R ( S i j T 0 , N i j T 0 , G i j T 0 , C i j T 0 ) = u r b a n , p i j T 0 > p t h r e s h o l d n o n u r b a n , p i j T 0 p t h r e s h o l d
where S i j T 0 and S i j T 0 + T represent the state (urban/nonurban) of cell i j at time T 0 and T 0 + T , respectively. p i j T 0 is the probability of urban sprawl of cell i j at T 0 , p t h r e s h o l d is the threshold value generally predefined in the range of [0,1]. N i j T 0 represents the state of neighboring cells. G i j T 0 represents global spatial variables, such as socio-economic factors and natural environmental conditions. C i j T 0 are the constraint conditions for urban sprawl. R are the urban-sprawl transition rules.
In addition, the various global spatial variables in the transition rules correspond to many weights that reflect the degree of contribution of different variables to the model. The setting of these weights has a great influence on the model simulation results. In practical implementations, it is difficult to effectively define the transition rules of these different categories of variables. Therefore, we used DBN and SDA to simplify this complex process with a training network to directly obtain the transition probability of every cell.
The proposed model consists of two parts (see Figure 3): The first part of the DBN (or SDA)-based CA model is to train the DBN (or SDA) to obtain the transition probability from a training dataset that is the sample of historic data and obtain a well-trained model. The simulation is cell-based, and each cell is composed of a set of attributes as inputs to the DBN (or SDA) after converting them in the range of [ 0 , 1 ] . The attributes during the training phase can be expressed by:
X = [ x 1 , x 2 , x 3 , , x n ] T
where x i is the ith attribute and T is transposition. As described above for DBN and SDA, the attribute inputs of these series of cellular automaton are directly used as the observed data for the unsupervised pretraining stage by using the DBN energy function (see Equation (2)) or the SDA loss function (see Equation (9)). Accordingly, each cell has its own state at the next time that can be expressed by:
Y = [ y 1 , y 2 , y 3 , , y n ] T
where y i is the ith state and T is transposition, which are regarded as the outputs to slightly adjust all parameters of DBN (or SDA) for the supervised fine-tuning stage.
The second part is to simulate (or predict) an urban sprawl by applying a prediction dataset to the well-trained model. Just like the training phase of the networks, the network input is the state of each cell itself, the state of the neighborhood, and other global spatial variables, and the output is the state of each cell at the next moment.

2.4. Accuracy Assessment

Ten indices were used as the criteria to evaluate the predictive performance of the proposed CA models in this study, which can be grouped into two types: criteria for assessing position accuracy at the cell level, and measures of pattern accuracy at the pattern level. At the cell level, we used FoM and the hit-miss-false alarm approach. Specifically, FoM is the ratio of the intersection of simulated and observed changes to the union of the simulated and observed changes [57], which can be calculated as follows:
F o M = Δ s i m u l a t e d Δ o b s e r v e d Δ s i m u l a t e d Δ o b s e r v e d = B A + B + C + D
where Δ s i m u l a t e d is the numbers of changed cells by the simulated map compared to the initial map; Δ o b s e r v e d is the numbers of changed cells by the observed map compared to the initial map; A is the numbers of incorrect cells due to observed changes simulated as persistence; B is the numbers of correct cells due to observed changes simulated as change; C is the numbers of incorrect cells due to observed changes simulated as a wrong category; and D is the numbers of incorrect cells due to observed persistence simulated as change.
The hit-miss-false alarm approach is a method used to quantify the goodness of fit of a land-change projection along a gradient of an explanatory variable [58], and it classified the pixels of a simulated map in this study as one of four types: correct due to observed urban sprawl predicted as urban sprawl (hits), error due to observed urban sprawl predicted as nonurban persistence (miss), error due to observed nonurban persistence predicted as urban sprawl (false alarms), and correct due to observed nonurban persistence predicted as nonurban persistence (correct rejection).
At the pattern level, a series of landscape indices, including the number of urban patches (NP), area-weighted mean patch fractal dimension (AWMPFD), edge density (ED), landscape shape index (LSI) and aggregation index (AI) were used in this study. Among them, NP, AWMPFD, and ED can be used to evaluate the fragmentation of the urban landscape, AWMPFD and LSI can be used to measure the complexity of the urban landscape, and AI can be used to assess the compactness of the urban landscape. The spatial urban form was measured based on a concise formula using the above-mentioned indices, which can be expressed as follows:
S i = 1 1 n i = 1 n Δ I i
Δ I i = | I i , s I i , o | I i , o × 100 % , I = N P , A W M P F D , E D , L S , A I
where I i , s and I i , o are the ith landscape indices of the simulated and observed map, respectively, and Δ I i is the normalized result of all landscape indices. S i is calculated by means of Δ I i , and can effectively reflect the similarity between the simulated and the observed map at the pattern level. n is the number of landscape indices, set as 5 in this study since five types of landscape indices were used.

3. Proposed-Model Implementation

3.1. Study Area

The study area that we selected for implementing the proposed model is the JJT urban agglomeration, also known as Beijing–Tianjin–Tangshan, which is the national capital region of China and the largest urbanized region in northern China. The area is approximately 48,100 km 2 and contains a population of approximately 50 million. It contains two municipalities (Beijing and Tianjin) and two prefecture-level cities (Tangshan and Langfang; see Figure 4). Beijing is the capital of China, the world’s third most populous city, and most populous capital city. Tianjin is the largest coastal metropolis and the third largest city in northern China. Beijing and Tianjin are regarded as the “dual core” of the JJT urban agglomeration and have accepted most migrants from other regions of northern China. The JJT urban agglomeration has been experiencing rapid urbanization since the implementation of reform and opening-up policy in 1978. During 2000–2010, the urban population and the GDP of the JJT urban agglomeration increased by about 5.34 million and RMB2383.01 billion (Yuan; approximately USD354.83 billion).
To construct a multiperiod CA model, this study adopted urban dynamics in 2000, 2005, and 2010. Specifically, the urban area in the study area occupies 4.13% of the total area in 2000, 5.53% in 2005 and 5.92% in 2010, generating an average sprawl area of 428.32 km 2 over this decade. With this rapid urban development, the land use of the JJT urban agglomeration has become very complex, leading to a series of environmental, economic, and urban-development issues. Although China has formulated the strictest land-use control and farmland-protection policies around the world, the influx of large populations inevitably leads to the continued sprawl of this area, and its land-use structure has experienced and will continue to experience dramatic changes. Therefore, an effective simulation of the urban sprawl in such an area can assist future urban planning and infrastructure construction, and guide the development of cities in a reasonable trajectory.

3.2. Data Preprocessing

3.2.1. Land-Use Data

A series of land-use maps (2000, 2005, and 2010) for the JJT were used in this study. All of these maps were produced from the classification of Landsat TM/ETM+ images (https://earthexplorer.usgs.gov/) with a resolution of 30 m, and have six land-use classes (urban land, cropland, grassland, forest, water, and others). They were used to derive past and current actual urban land, and detect urban sprawl (see Figure 5), which serve as the initial and final state of each cell during simulations in CA models. In the simulation phase, the target variable represents if the state of a nonurban cell remained (0), if a nonurban cell developed into an urban state (1), and if an urban cell remained (2) (here, we do not consider the case of an urban cell converting to a nonurban cell because this study only simulates the progress of the urban sprawl.)

3.2.2. Neighborhood Conditions

The neighborhood conditions include both nonurban and urban neighborhoods, which is also important for the determination of an urban sprawl during the simulation. Although 3 × 3 , 5 × 5 , and 7 × 7 windows have been widely used in previous study, we performed a test in which we found that the 7 × 7 window performed best, and it may be because it contains more neighborhood interactions. Therefore, we chose the number of surrounding urban cells in 7 × 7 with the extended Moore neighborhood in the implementation. The states of the surrounding cells were derived from land-use maps using the FocalStatistics function of ArcPy.

3.2.3. Global Spatial Variables

Global spatial variables, as we mentioned above, have great influence on an urban sprawl. In this implementation, we chose a series of global spatial variables, including physical property and distance-based variables. The physical property mainly considered the influence of topography on urban sprawl, which included the digital elevation model (DEM) and slope. The DEM was derived from SRTM data (https://earthexplorer.usgs.gov/), and the slope was calculated from the DEM using the Slope function of ArcPy. Distance-based variables include distance to the nearest airport, distance to the nearest city center, distance to the nearest county center, distance to the nearest town center, distance to the nearest railway station, distance to the nearest reservoir, distance to the nearest motorway, distance to the nearest railway, distance to the nearest river, and distance to the nearest trunkway. This study accounts for distance-based variables from the aspects of accessibility, government/consumer decisions, and living environment. Among them, the vector data for airport, city center, county center, and town center were derived from the National Geomatics Center of China (http://www.ngcc.cn/), and the vector data for reservoir, river, railway station, motorway, trunkway, and railway were derived from OpenStreetMap (http://www.openstreetmap.org). All distance-based variables were calculated from the vector data using the EucDistance function of ArcPy (see Figure 6).
As listed in Table 1, all data mentioned above were resampled to a spatial resolution of 100 × 100 m, with 3325 columns and 2910 rows, and restored them as the full dataset in CSV format to facilitate the proposed models. Then, using a systematic sampling approach, we selected 1,000,000 samples from the full dataset to serve as inputs to the proposed models for defining the CA transition rules. Because a large amount of data are available and efficient computing power in the era of Big Data, we did not have to compromise as much and could use a greater portion to train the model. Thus, samples were divided into three parts: 980,000 samples as training data used to fit DBN or SDA, 10,000 as valid data used to provide an unbiased evaluation of the DBN or SDA fit on the training data while tuning model hyperparameters, and 10,000 as test data used to provide an unbiased evaluation of a final model fit on the training data.

3.3. Model Design and Computational Environment

The entire implemented process, including data preprocessing, training of the proposed models, and simulation, is based on Python. Some basic scientific-computation libraries like Numpy, Pandas, and Rasterio were used for spatial-data organization, such as raster-data reading and their conversion to CSV. Arcpy was used for spatial-data analysis, such as neighborhood or distance analysis. Scikit-learn (https://scikit-learn.org/stable/) was used to develop the artificial-neural-network (ANN) and random-forest (RF) methods as method comparison.
In addition to the above scientific computation libraries, it is particularly important to emphasize that the computational framework for DBN and SDA in our study was based on Theano (http://deeplearning.net/software/theano/), which is an open-source project that has the following characteristics: (1) symbolic computation for tensors, (2) highly expressive, transparent GPU acceleration, (3) easily switches between CPU and GPU, and (4) easily integrates with Python ecosystems. When running DBN and SDA based on Theano, there are some parameters that need to be preset (see Table 2).

3.4. Simulation Results and Comparisons

The DBN–CA and SDA–CA models proposed in this study were established through known data from 2000, 2005, and 2010. The initial year was 2000, and the 2000–2005 samples were used to calibrate the proposed models for discovering the transition rules; the simulation result of urban sprawl during 2000–2005 was conducted simultaneously. Next, the simulated urban-land status of 2005 was acquired to conduct the simulation result of the urban sprawl during 2005–2010 based on the calibrated models. Land-use data in 2010 were regarded as the observed urban-land status to confirm the validation of the calibrated DBN–CA (or SDA–CA) model. Figure 7 and Figure 8 show the consistency between the initial, observed, and simulated urban-land status and enlarged area of Beijing in 2005 and 2010 based on the four models mentioned above.
Additionally, FoM were used for more quantitative assessments of the proposed models at the cell level. As shown in Figure 9, SDA–CA obtain the best simulated results in both periods. Compared with DBN–CA, the FoM of SDA-CA increased by 9.8% in 2005 and 14.2% in 2010, which demonstrated that SDA is more suitable for discovering the transition rules for CA than DBN at cell level. Meanwhile, compared with the other two traditional machine-learning algorithms, the FoM of the SDA–CA increased by 16.6–31.6 in 2005, and 17.9–19.7% in 2010, which corroborated that SDA can learn more information from multivariable nonlinear systems than traditional machine-learning algorithms at cell level, thus obtaining better recognition of urban sprawl.
In Table 3, we outline the FoM of the major city’s simulation results in JJT by the four models for 2005 and 2010. In general, the minimum FoM of DBN–CA and SDA–CA in all cities for 2005 and 2010 reached 0.332, which is a reasonable result of the relevant research. The FoM of unsupervised deep-learning algorithms are generally better than traditional machine-learning algorithms in every city. Comparing the FoM of four models, SDA–CA achieved the best result, and more specifically, the SDA–CA FoM of Beijing increased by 16.8–34.0% in 2005 and 12.5–19.1% in 2010, the SDA–CA FoM of Tianjin increased by 7.6–30.1% in 2005 and 8.3–21.0% in 2010, and the SDA-CA FoM of Tangshan increased by 21.2–26.4% in 2005 and 10.0–23.3% in 2010. These results demonstrate that the implementation effect of the model varied in different years or in different cities. From the perspective of development trends of various cities, taking the SDA-CA FoM as an example, Beijing and Tianjin, which are the municipality of China and a metropolis in northern China that developed at a high speed due to the effects of population aggregation, generally have a higher FoM of around 0.4–9.1% than Tangshan. These results illustrate that areas with a fierce urban sprawl are often able to achieve better simulation results because they provide more samples of the urban sprawl for training the models.
In addition to using FoM to compare the simulation results of the four models, we also introduced the hit-miss-false alarm approach to measure the similarity between the simulated map and the observed map at the cell level. Four different types of results by four models for 2005 and 2010 are summarized in Table 4. Generally speaking, unsupervised deep-learning-based models have better performance than shallow machine-learning-based models in all four different types of results, and SDA–CA obtained the best results from the four models. Specifically, compared with DBN–CA, the numbers of SDA–CA hits increased by 6.11% in 2005 and 7.80% in 2010, the numbers of SDA-CA’s miss decreased by 41.02% in 2005 and 39.21% in 2010, the numbers of SDA–CA’s false alarm decreased by 22.48% in 2005 and 9.30% in 2010, and the numbers of SDA–CA’s correct rejection increased by 0.08% in 2005 and 0.08% in 2010. Compared with the two other shallow machine-learning models, the numbers of SDA–CA’s hits increased by 5.94–11.66% in 2005 and 8.41–10.34% in 2010, the numbers of SDA–CA’s misses decreased by 43.04–55.77% in 2005 and 40.88–45.50% in 2010, the numbers of SDA–CA’s false alarm decreased by 10.48% (but increased by 63.03% compared with ANN–CA) in 2005 and 22.50–49.56% in 2010, and the numbers of SDA–CA’s correct rejection increased by 0.51% (but decreased by 0.03% compared with ANN–CA) in 2005 and 0.21–0.73% in 2010. These analysis results confirm the validity of the proposed models at cell level from another aspect.
To further demonstrate the results of our proposed models, here we enlarge the simulation results of SDA–CA to three major cities (Beijing, Tianjin, and Tangshan) in Figure 10. Visual inspection suggests that the hit area mainly concentrated on the vicinity of the initial urban area, the miss area mainly focused on where there is a sudden change in urban sprawl, and false alarm exists in urban fringe and nonurban areas. So, we can infer that the proposed model is more suitable for urban sprawling than a leap-forward development of urban areas.
At pattern level, Table 5 lists a series of landscape indices for four different models in 2005 and 2010 that can distinguish the differences between the simulated and observed maps. In general, SDA–CA had the best performance in pattern similarity. As seen from the table, In the simulation results for 2005, the S i between the simulated map of SDA–CA and the observed map is 88.37 % , with an NP value difference of merely 37, an ED value difference of merely 0.164, an LSI value difference of merely 3.504, an AWMPFD value difference of merely 0.016, and an AI value difference of merely 0.627. Compared with DBN–CA, the S i of SDA–CA increased by 3.29% in 2005 and 2.05% in 2010, which demonstrated that SDA is more suitable for urban sprawl than DBN at the pattern level. Moreover, compared with the two other traditional machine-learning models, the S i of SDA–CA increased by 7.64–38.08% in 2005 and 9.5–11.75% in 2010, which demonstrated that SDA is more suitable for urban sprawl than DBN at pattern level. On the whole, these results indicate the capability of SDA–CA for simulating the urban sprawl, and the SDA–CA model behaves well during the simulation at pattern level.

4. Discussion and Conclusions

The urban sprawl has many negative consequences for residents and the environment, and how to accurately simulate this process for future planning has been a hot research topic of geographical simulations. Prior works documented the effectiveness of the CA model in simulating urban sprawls [3,9,11] and introduced a series of algorithms into the CA model to more efficiently discover its transition rules. However, these CA-based models either have remaining problems like overfitting, or have resulted in a local optimum. In addition, from the perspective of the development of machine learning, traditional shallow machine-learning algorithms can no longer adapt to current big-sample-oriented learning ideas in the era of Big Data. Therefore, we introduced two unsupervised deep-learning algorithms that achieved great power and flexibility in representation learning, for defining CA transition rules, and testing the effectiveness of the proposed models in an urban-sprawl simulation.
To test the effectiveness of the proposed models in different cities, we selected the JJT urban agglomeration as the study area and simulated its urban sprawl in 2000–2010. First, land-use data, neighborhood conditions, and global spatial variables in 2000, 2005, and 2010 were used as the full dataset for model calibration and validation. Then, based on the sample data from 2000 to 2005, two calibrated CA models were constructed by using DBN and SDA. Last, by using the two calibrated CA models, two simulated results of JJT in 2010 were conducted. Model validation was implemented by assessing consistency between the simulated and observed maps and comparing with two traditional CA models (ANN–CA and LR–CA models). As shown by a series of model accuracy assessments, the SDA–CA model obtained the best simulation results at both cell level and pattern level. Meanwhile, compared with other traditional machine-learning-based CA models, the DBN–CA model also had better performance in most aspects. These findings demonstrated that unsupervised deep-learning algorithms are well suitable for discovering CA transition rules, and the performance of the SDA–CA model may be better than the DBN–CA model for simulating complex geographical systems.
Most notably, this is the first study to our knowledge to investigate the capability of two typical unsupervised deep-learning algorithms for defining the transition rules of CA models in the same case. Our study provided a systematic comparison between unsupervised deep-learning-based CA models and traditional machine-learning-based CA models in a real urban-sprawl case. However, some limitations are worth noting. Although simulated results of the proposed models showed stronger capability in feature learning from urban-land-change data than traditional models, the relationship between an urban-sprawl state and driver factors (such as the neighborhood condition and global spatial variables) is still ambiguous. Future work should therefore not only introduce more deep-learning algorithms, such as RNN or GAN, to enhance the feature-learning ability of the CA model, but also needs to consider integrating the CA model with reinforcement learning [59] to improve simulation interpretability.

Author Contributions

C.O. and J.Y. conceived the study design, developed the models, and drafted the manuscript. J.Y. provided the funding. C.O. and Z.D. was involved in data acquisition and analysis, and worked on aspects of experiment evaluation. X.Z. improved the conceptual framework and updated the manuscript. D.Z. was involved in the design and revision of the manuscript. All authors read and approved the final manuscript.

Funding

This research was funded by the Chinese Universities Scientific Fund “Monitoring the quantity and quality using remote sensing and intelligence analysis of total factor for cropland”, grant number 2019TC117 and the Special Fund for Scientific Research on Public Causes, grant number 201511010-06.

Acknowledgments

The authors would like to thank Wanling Chen and Yiming Liu for their suggestions in improving both the structure and the details of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Stow, D.A.; Chen, D. Sensitivity of multitemporal NOAA AVHRR data of an urbanizing region to land-use/land-cover changes and misregistration. Remote Sens. Environ. 2002, 80, 297–307. [Google Scholar] [CrossRef]
  2. Wolfram, S. Universality and complexity in cellular automata. Int. Sympos. Phys. Des. 1984, 10, 1–35. [Google Scholar] [CrossRef]
  3. Santé, I.; García, A.M.; Miranda, D.; Crecente, R. Cellular automata models for the simulation of real-world urban processes: A review and analysis. Landsc. Urban Plan. 2010, 96, 108–122. [Google Scholar] [CrossRef]
  4. Liu, X.; Li, X.; Liu, L.; He, J.; Ai, B. A bottom-up approach to discover transition rules of cellular automata using ant intelligence. Int. J. Geogr. Inf. Sci. 2008, 22, 1247–1269. [Google Scholar] [CrossRef]
  5. Wu, F. SimLand: A prototype to simulate land conversion through the integrated GIS and CA with AHP-derived transition rules. Int. J. Geogr. Inf. Sci. 1998, 12, 63–82. [Google Scholar] [CrossRef]
  6. Khan, S. Cellular automata-based spatial multi-criteria land suitability simulation for irrigated agriculture. Int. J. Geogr. Inf. Sci. 2011, 25, 131–148. [Google Scholar]
  7. Lai, T.; Dragićević, S.; Schmidt, M. Integration of multicriteria evaluation and cellular automata methods for landslide simulation modelling. Geomatics Nat. Hazards Risk 2013, 4, 355–375. [Google Scholar] [CrossRef]
  8. Arsanjani, J.J.; Helbich, M.; Kainz, W.; Boloorani, A.D. Integration of logistic regression, Markov chain and cellular automata models to simulate urban expansion. Int. J. Appl. Earth Obs. Geoinform. 2013, 21, 265–275. [Google Scholar] [CrossRef]
  9. Almeida, C.M.; Monteiro, A.M.V.; Camara, G.; Soares-Filho, B.S.; Cerqueira, G.C. Modeling The Urban Evolution Of Land Use Transitions Using Cellular Automata And Logistic Regression. In Proceedings of the IEEE International Geoscience & Remote Sensing Symposium, Toulouse, France, 21–25 July 2003. [Google Scholar]
  10. Chen, Y.; Li, X.; Liu, X.; Ai, B. Modeling urban land-use dynamics in a fast developing city using the modified logistic cellular automaton with a patch-based simulation strategy. Int. J. Geogr. Inf. Sci. 2014, 28, 234–255. [Google Scholar] [CrossRef]
  11. Li, X.; Yeh, A.G. Neural-network-based cellular automata for simulating multiple land use changes using GIS. Int. J. Geogr. Inf. Sci. 2002, 16, 323–343. [Google Scholar] [CrossRef]
  12. Huang, W.; Liu, H.; Bai, M. Urban expansion simulation based on constrained Artificial Neural Network cellular automata model. In Proceedings of the International Conference on Geoinformatics, Fairfax, VA, USA, 12–14 August 2009. [Google Scholar]
  13. Basse, R.M.; Omrani, H.; Charif, O.; Gerber, P.; Bódis, K. Land use changes modelling using advanced methods: Cellular automata and artificial neural networks. The spatial and explicit representation of land cover dynamics at the cross-border region scale. Appl. Geogr. 2014, 53, 160–171. [Google Scholar] [CrossRef]
  14. Ozturk, D. Urban Growth Simulation of Atakum (Samsun, Turkey) Using Cellular Automata-Markov Chain and Multi-Layer Perceptron-Markov Chain Models. Remote Sens. 2015, 7, 5918–5950. [Google Scholar] [CrossRef]
  15. Yang, Q.; Li, X.; Shi, X. Cellular automata for simulating land use changes based on support vector machines. Comput. Geosci. 2008, 34, 592–602. [Google Scholar] [CrossRef]
  16. Rienow, A.; Goetzke, R. Supporting SLEUTH—Enhancing a cellular automaton with support vector machines for urban growth modeling. Comput. Environ. Urban Syst. 2015, 49, 66–81. [Google Scholar] [CrossRef]
  17. Okwuashi, O.; Nwilo, P.; Mcconchie, J.; Eyo, E. Enhancing a GIS Cellular Automata model of land use change using Support Vector Machine. In Proceedings of the International Conference on Geoinformatics, Fairfax, VA, USA, 12–14 August 2009. [Google Scholar]
  18. Li, X.; Yeh, A.G. Data mining of cellular automata’s transition rules. Int. J. Geogr. Inf. Sci. 2004, 18, 723–744. [Google Scholar] [CrossRef]
  19. Florencio Ballestores, Z.Q. An integrated parcel-based land use change model using cellular automata and decision tree. Proc. Int. Acad. Ecol. Environ. Sci. 2012, 2, 53. [Google Scholar]
  20. Kamusoko, C.; Gamba, J. Simulating Urban Growth Using a Random Forest-Cellular Automata (RF-CA) Model. ISPRS Int. J. Geo-Inf. 2015, 4, 447–470. [Google Scholar] [CrossRef]
  21. Yao, Y.; Liu, X.; Zhang, D.; Liang, Z.; Zhang, Y. Simulation of Urban Expansion and Farmland Loss in China by Integrating Cellular Automata and Random Forest. arXiv 2017, preprint. arXiv:1705.05651. [Google Scholar]
  22. Jenerette, G.D.; Wu, J. Analysis and simulation of land-use change in the central Arizona—Phoenix region, USA. Landsc. Ecol. 2001, 16, 611–626. [Google Scholar] [CrossRef]
  23. García, A.M.; Santé, I.; Boullón, M.; Crecente, R. Calibration of an urban cellular automaton model by using statistical techniques and a genetic algorithm. Application to a small urban settlement of NW Spain. Int. J. Geogr. Inf. Sci. 2013, 27, 1593–1611. [Google Scholar] [CrossRef]
  24. Xiaoping, L.; Xia, L.; Yeh, A.G.; Jinqiang, H.E.; Jia, T. Discovery of transition rules for geographical cellular automata by using ant colony optimization. Sci. China-Earth Sci. 2007, 50, 1578–1588. [Google Scholar]
  25. Feng, Y.; Liu, Y.; Tong, X.; Liu, M.; Deng, S. Modeling dynamic urban growth using cellular automata and particle swarm optimization rules. Landsc. Urban Plan. 2011, 102, 188–196. [Google Scholar] [CrossRef]
  26. Yang, J.; Tang, G.; Cao, M.; Zhu, R. An intelligent method to discover transition rules for cellular automata using bee colony optimisation. Int. J. Geogr. Inf. Sci. 2013, 27, 1849–1864. [Google Scholar] [CrossRef]
  27. Cao, M.; Tang, G.; Shen, Q.; Wang, Y. A new discovery of transition rules for cellular automata by using cuckoo search algorithm. Int. J. Geogr. Inf. Syst. 2015, 29, 806–824. [Google Scholar] [CrossRef]
  28. Cao, M.; Bennett, S.J.; Shen, Q.; Xu, R. A bat-inspired approach to define transition rules for a cellular automaton model used to simulate urban expansion. Int. J. Geogr. Inf. Sci. 2016, 30, 1–19. [Google Scholar] [CrossRef]
  29. He, C.; Zhao, Y.; Tian, J.; Shi, P. Modeling the urban landscape dynamics in a megalopolitan cluster area by incorporating a gravitational field model with cellular automata. Landsc. Urban Plan. 2013, 113, 78–89. [Google Scholar] [CrossRef]
  30. Bengio, Y.; Courville, A.; Vincent, P. Unsupervised Feature Learning and Deep Learning: A Review and New Perspectives. CoRR 2012, abs/1206.5538. [Google Scholar]
  31. Najafabadi, M.M.; Villanustre, F.; Khoshgoftaar, T.M.; Seliya, N.; Wald, R.; Muharemagic, E. Deep learning applications and challenges in big data analytics. J. Big Data 2015, 2, 1. [Google Scholar] [CrossRef]
  32. Vincent, P.; Larochelle, H.; Bengio, Y.; Manzagol, P.A. Extracting and composing robust features with denoising autoencoders. In Proceedings of the International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; pp. 1096–1103. [Google Scholar]
  33. Chen, Y.; Zhao, X.; Jia, X. Spectral–Spatial Classification of Hyperspectral Data Based on Deep Belief Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  34. Budiman, A.; Fanany, M.I.; Basaruddin, C. Stacked Denoising Autoencoder for feature representation learning in pose-based action recognition. In Proceedings of the 2014 IEEE 3rd Global Conference on Consumer Electronics (GCCE), Tokyo, Japan, 7–10 October 2014. [Google Scholar]
  35. Erhan, D.; Bengio, Y.; Courville, A.C.; Manzagol, P.; Vincent, P.; Bengio, S. Why Does Unsupervised Pre-training Help Deep Learning? J. Mach. Learn. Res. 2010, 11, 625–660. [Google Scholar]
  36. Vincent, P.; Larochelle, H.; Lajoie, I.; Bengio, Y.; Manzagol, P. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion. J. Mach. Learn. Res. 2010, 11, 3371–3408. [Google Scholar]
  37. Hinton, G.E.; Salakhutdinov, R.R. Reducing the Dimensionality of Data with Neural Networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed]
  38. Arsa, D.M.S.; Jati, G.; Mantau, A.J.; Wasito, I. Dimensionality reduction using deep belief network in big data case study: Hyperspectral image classification. In Proceedings of the International Workshop on Big Data & Information Security, Jakarta, Indonesia, 18–19 October 2017. [Google Scholar]
  39. Zabalza, J.; Ren, J.; Zhao, H.; Zhao, H.; Qing, C.; Yang, Z.; Du, P.; Marshall, S. Corrigendum to ‘Novel segmented stacked autoencoder for effective dimensionality reduction and feature extraction in hyperspectral imaging’ Neurocomputing 185 (2016) 1–10. Neurocomputing 2016, 214, 1062. [Google Scholar] [CrossRef]
  40. Romeu, P.; Zamora-Martínez, F.; Botella-Rocamora, P.; Pardo, J. Stacked Denoising Auto-Encoders for Short-Term Time Series Forecasting; Springer: Cham, Switzerland, 2015. [Google Scholar]
  41. Long, W.; Zhang, Z.; Chen, J. Short-Term Electricity Price Forecasting with Stacked Denoising Autoencoders. IEEE Trans. Power Syst. 2017, 32, 2673–2681. [Google Scholar]
  42. Kuremoto, T.; Kimura, S.; Kobayashi, K.; Obayashi, M. Time series forecasting using a deep belief network with restricted Boltzmann machines. Neurocomputing 2014, 137, 47–56. [Google Scholar] [CrossRef]
  43. Ahmed, M.; Shill, P.C.; Islam, K.; Mollah, M.A.S.; Akhand, M.A.H. Acoustic modeling using deep belief network for Bangla speech recognition. In Proceedings of the 2015 18th International Conference on Computer and Information Technology (ICCIT), Dhaka, Bangladesh, 21–23 December 2016. [Google Scholar]
  44. Mohamed, A.R.; Dahl, G.E.; Hinton, G. Acoustic Modeling Using Deep Belief Networks. IEEE Trans. Audio Speech Lang. Process. 2011, 20, 14–22. [Google Scholar] [CrossRef]
  45. Liu, J.H.; Zheng, W.Q.; Zou, Y.X. A Robust Acoustic Feature Extraction Approach Based on Stacked Denoising Autoencoder. In Proceedings of the IEEE International Conference on Multimedia Big Data, Beijing, China, 20–22 April 2015. [Google Scholar]
  46. Gu, F.; Khoshelham, K.; Valaee, S.; Shang, J.; Rui, Z. Locomotion Activity Recognition Using Stacked Denoising Autoencoders. IEEE Internet Things J. 2018, 5, 2085–2093. [Google Scholar] [CrossRef]
  47. Gu, F.; Flórezrevuelta, F.; Monekosso, D.; Remagnino, P. Marginalised Stacked Denoising Autoencoders for Robust Representation of Real-Time Multi-View Action Recognition. Sensors 2015, 15, 17209–17231. [Google Scholar] [CrossRef] [PubMed]
  48. Wicht, B.; Henneberty, J. Mixed handwritten and printed digit recognition in Sudoku with Convolutional Deep Belief Network. In Proceedings of the 2015 13th International Conference on Document Analysis and Recognition (ICDAR), Tunis, Tunisia, 23–26 August 2015. [Google Scholar]
  49. He, J.; Li, X.; Yao, Y.; Hong, Y.; Zhang, J. Mining transition rules of cellular automata for simulating urban expansion by using the deep learning techniques. Int. J. Geogr. Inf. Sci. 2018, 32, 2076–2097. [Google Scholar] [CrossRef]
  50. Zhou, Y.; Zhang, F.; Du, Z.; Ye, X.; Liu, R. Integrating Cellular Automata with the Deep Belief Network for Simulating Urban Growth. Sustainability 2017, 9, 1786. [Google Scholar] [CrossRef]
  51. Du, G.; Yuan, L.; Shin, K.J.; Managi, S. Enhancement of land-use change modeling using convolutional neural networks and convolutional denoising autoencoders. arXiv 2018, preprint. arXiv:1803.01159. [Google Scholar]
  52. Testolin, A.; Piccolini, M.; Suweis, S. Deep Learning Systems as Complex Networks. arXiv 2018, preprint. arXiv:1809.10941. [Google Scholar]
  53. Hinton, G.E. Training Products of Experts by Minimizing Contrastive. Neural Comput. 2002, 14, 1771–1800. [Google Scholar] [CrossRef]
  54. Qin, M.; Li, Z.; Du, Z. Red tide time series forecasting by combining ARIMA and deep belief network. Knowl.-Based Syst. 2017, 125, 39–52. [Google Scholar] [CrossRef]
  55. White, R.; Engelen, G. Cellular Automata and Fractal Urban Form: A Cellular Modelling Approach to the Evolution of Urban Land-Use Patterns. Environ. Plan. A 1993, 25, 1175–1199. [Google Scholar] [CrossRef]
  56. Salapayca, S.; Jankowski, P.; Clarke, K.C.; Kyriakidis, P.C.; Nara, A. A meta-modeling approach for spatio-temporal uncertainty and sensitivity analysis: An application for a cellular automata-based Urban growth and land-use change model. Int. J. Geogr. Inf. Sci. 2018, 32, 637–662. [Google Scholar] [CrossRef]
  57. Pontius, R.G., Jr.; Walker, R.; Yao-Kumah, R.; Arima, E.; Aldrich, S.; Vergara, M.C.D. Accuracy Assessment for a Simulation Model of Amazonian Deforestation. Ann. Assoc. Am. Geogr. 2010, 97, 677–695. [Google Scholar] [CrossRef]
  58. Chen, H.; Pontius, R.G., Jr. Diagnostic tools to evaluate a spatial land change projection along a gradient of an explanatory variable. Landsc. Ecol. 2010, 25, 1319–1331. [Google Scholar] [CrossRef]
  59. Zilinskas, A. Simulation-Based Optimization: Parametric Optimization Techniques and Reinforcement Learning. Interfaces 2003, 35, 535–536. [Google Scholar]
Figure 1. Architecture of (a) Deep Belief Networks (DBNs) and (b) Restricted Boltzmann Machines (RBMs).
Figure 1. Architecture of (a) Deep Belief Networks (DBNs) and (b) Restricted Boltzmann Machines (RBMs).
Sustainability 11 02464 g001
Figure 2. Architecture of (a) Stacked Denoising Autoencoder (SDA) and (b) Denoising Autoencoder (DAE).
Figure 2. Architecture of (a) Stacked Denoising Autoencoder (SDA) and (b) Denoising Autoencoder (DAE).
Sustainability 11 02464 g002
Figure 3. Flowchart of the proposed model.
Figure 3. Flowchart of the proposed model.
Sustainability 11 02464 g003
Figure 4. Location of Jingjintang urban agglomeration.
Figure 4. Location of Jingjintang urban agglomeration.
Sustainability 11 02464 g004
Figure 5. Urban sprawl in the study area from 2000 to 2010.
Figure 5. Urban sprawl in the study area from 2000 to 2010.
Sustainability 11 02464 g005
Figure 6. Spatial variables in metropolitan Jingjintang: (a) digital elevation model (DEM); (b) slope; (c) distance to airport; (d) distance to city center; (e) distance to county center; (f) distance to town center; (g) distance to railway station; (h) distance to reservoir; (i) distance to motorway; (j) distance to railway; (k) distance to river; (l) distance to trunkway.
Figure 6. Spatial variables in metropolitan Jingjintang: (a) digital elevation model (DEM); (b) slope; (c) distance to airport; (d) distance to city center; (e) distance to county center; (f) distance to town center; (g) distance to railway station; (h) distance to reservoir; (i) distance to motorway; (j) distance to railway; (k) distance to river; (l) distance to trunkway.
Sustainability 11 02464 g006
Figure 7. Comparison of simulated results by four models for 2005. (a) Enlarged area of SDA–cellular automata (CA), (b) enlarged area of DBN–CA, (c) enlarged area of ANN-CA, (d) enlarged area of LR-CA.
Figure 7. Comparison of simulated results by four models for 2005. (a) Enlarged area of SDA–cellular automata (CA), (b) enlarged area of DBN–CA, (c) enlarged area of ANN-CA, (d) enlarged area of LR-CA.
Sustainability 11 02464 g007
Figure 8. Comparison of simulated results by four models for 2010. (a) Enlarged area of SDA-CA, (b) enlarged area of DBN-CA, (c) enlarged area of ANN-CA, (d) enlarged area of LR-CA.
Figure 8. Comparison of simulated results by four models for 2010. (a) Enlarged area of SDA-CA, (b) enlarged area of DBN-CA, (c) enlarged area of ANN-CA, (d) enlarged area of LR-CA.
Sustainability 11 02464 g008
Figure 9. Figures of Merit (FoM) of simulated results by four models for 2005 and 2010.
Figure 9. Figures of Merit (FoM) of simulated results by four models for 2005 and 2010.
Sustainability 11 02464 g009
Figure 10. Comparison of enlarged areas in various cities based on detection theory. (A1–2) Enlarged area of Beijing in 2005 and 2010, (B1–2) enlarged area of Tianjin in 2005 and 2010, (B1–2) enlarged area of Tangshan in 2005 and 2010.
Figure 10. Comparison of enlarged areas in various cities based on detection theory. (A1–2) Enlarged area of Beijing in 2005 and 2010, (B1–2) enlarged area of Tianjin in 2005 and 2010, (B1–2) enlarged area of Tangshan in 2005 and 2010.
Sustainability 11 02464 g010
Table 1. List of spatial variables used in urban-sprawl simulation.
Table 1. List of spatial variables used in urban-sprawl simulation.
Spatial VariableValue Range
1 . T a r g e t   v a r i a b l e
Urban sprawl in 2000–2010Remained nonurban area: 0; converted to urban area: 1; remained urban area: 2
2 . L o c a t i o n a l   v a r i a b l e s
Distance to airport0–175,594 m
Distance to city administrative center0–129,142 m
Distance to town administrative center0–47,261 m
Distance to county administrative center0–63,279 m
Distance to reservoir0–44,457 m
Distance to river0–88,911 m
3 . A c c e s s i b i l i t y
Distance to railway station0–109,976 m
Distance to motorway0–70,052 m
Distance to trunkway0–21,979 m
Distance to railway0–33,278 m
4 . N e i g h b o r h o o d   c o n d i t i o n
Urban cells in 7 × 7 neighborhood0–48
5 . C e l l   c h a r a c t e r i s t i c
Land use typeurban area: 1; nonurban area: 0
Elevation−52–2224 m
Slope0–69
Table 2. Set parameters for DBN and SDA.
Table 2. Set parameters for DBN and SDA.
DescriptionSymbolDBNSDA
Input dimensionn_ins1111
Output dimensionn_out33
Intermediate layer sizehidden_layer_sizes[30, 30, 30][30, 30, 30]
Number of epochs for pretrainingpre-training_epochs300300
Maximal number of iterations of running optimizertraining_epochs30003000
Learning rate used in pretrainingpretrain_lr0.010.001
Learning rate used in fine-tuning stagefinetune_lr0.10.01
Table 3. FoM of major city’s simulation results in JJT by four models for 2005 and 2010.
Table 3. FoM of major city’s simulation results in JJT by four models for 2005 and 2010.
SDA-CADBN-CAANN-CALR-CA
2005
Beijing0.5830.3850.2430.415
Tianjin0.5810.4030.2800.505
Tangshan0.5790.3670.3200.315
2010
Beijing0.4680.3430.2770.294
Tianjin0.5230.3640.3130.440
Tangshan0.4320.3320.2990.199
Table 4. Hit-miss-false alarm of simulated results by four models for 2005 and 2010.
Table 4. Hit-miss-false alarm of simulated results by four models for 2005 and 2010.
SDA–CADBN–CAANN–CALR–CA
2005
Hit244528230445219001230811
Miss20247343304577435547
False alarm12555161951136433964
Correcct rejection4514091451045145152824491099
2010
Hit255148236676231238235346
Miss28638471105254848440
False alarm32785361484230364997
Correcct rejection4474850447148744653324442638
Table 5. Comparison of observed and simulated values of landscape indices.
Table 5. Comparison of observed and simulated values of landscape indices.
NPEDLSIAWMPFDAIS_i
2005
observed1421.01923.8891.17595.539-
SDA-CA1790.85520.3851.15996.166 88.37 %
DBN-CA1930.81519.8421.15496.196 85.08 %
ANN-CA2290.81820.6141.15595.905 80.73 %
LR-CA3650.59113.8011.08997.513 50.29 %
2010
observed1791.17426.5791.18695.187-
SDA-CA1910.83118.7251.14596.688 86.21 %
DBN-CA2290.88420.3841.14896.281 84.16 %
ANN-CA2150.63514.6131.12297.390 76.71 %
LR-CA3611.08023.7051.10195.846 74.46 %

Share and Cite

MDPI and ACS Style

Ou, C.; Yang, J.; Du, Z.; Zhang, X.; Zhu, D. Integrating Cellular Automata with Unsupervised Deep-Learning Algorithms: A Case Study of Urban-Sprawl Simulation in the Jingjintang Urban Agglomeration, China. Sustainability 2019, 11, 2464. https://doi.org/10.3390/su11092464

AMA Style

Ou C, Yang J, Du Z, Zhang X, Zhu D. Integrating Cellular Automata with Unsupervised Deep-Learning Algorithms: A Case Study of Urban-Sprawl Simulation in the Jingjintang Urban Agglomeration, China. Sustainability. 2019; 11(9):2464. https://doi.org/10.3390/su11092464

Chicago/Turabian Style

Ou, Cong, Jianyu Yang, Zhenrong Du, Xin Zhang, and Dehai Zhu. 2019. "Integrating Cellular Automata with Unsupervised Deep-Learning Algorithms: A Case Study of Urban-Sprawl Simulation in the Jingjintang Urban Agglomeration, China" Sustainability 11, no. 9: 2464. https://doi.org/10.3390/su11092464

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop