Next Article in Journal
Potential for the Detection of Irrigation Events on Maize Plots Using Sentinel-1 Soil Moisture Products
Previous Article in Journal
A Convolutional Neural Network with Parallel Multi-Scale Spatial Pooling to Detect Temporal Changes in SAR Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mapping Rice Paddy Based on Machine Learning with Sentinel-2 Multi-Temporal Data: Model Comparison and Transferability

1
College of Resources and Environment, Southwest University, Chongqing 400716, China
2
College of Computer and Information Science, Southwest University, Chongqing 400716, China
3
Chongqing Agricultural Technology Extension Station, Chongqing Municipal Committee of Agriculture and Rural Affairs, Chongqing 400121, China
4
State Key Laboratory of Remote Sensing Science, College of Global Change and Earth System Science, Beijing Normal University, Beijing 100000, China
5
Department of Atmospheric and Oceanic Science, Earth System Science Interdisciplinary Center, University of Maryland, College Park, MD 20740, USA
*
Author to whom correspondence should be addressed.
Remote Sens. 2020, 12(10), 1620; https://doi.org/10.3390/rs12101620
Submission received: 9 April 2020 / Revised: 15 May 2020 / Accepted: 17 May 2020 / Published: 19 May 2020
(This article belongs to the Section Remote Sensing in Agriculture and Vegetation)

Abstract

:
Rice is an important agricultural crop in the Southwest Hilly Area, China, but there has been a lack of efficient and accurate monitoring methods in the region. Recently, convolutional neural networks (CNNs) have obtained considerable achievements in the remote sensing community. However, it has not been widely used in mapping a rice paddy, and most studies lack the comparison of classification effectiveness and efficiency between CNNs and other classic machine learning models and their transferability. This study aims to develop various machine learning classification models with remote sensing data for comparing the local accuracy of classifiers and evaluating the transferability of pretrained classifiers. Therefore, two types of experiments were designed: local classification experiments and model transferability experiments. These experiments were conducted using cloud-free Sentinel-2 multi-temporal data in Banan District and Zhongxian County, typical hilly areas of Southwestern China. A pure pixel extraction algorithm was designed based on land-use vector data and a Google Earth Online image. Four convolutional neural network (CNN) algorithms (one-dimensional (Conv-1D), two-dimensional (Conv-2D) and three-dimensional (Conv-3D_1 and Conv-3D_2) convolutional neural networks) were developed and compared with four widely used classifiers (random forest (RF), extreme gradient boosting (XGBoost), support vector machine (SVM) and multilayer perceptron (MLP)). Recall, precision, overall accuracy (OA) and F1 score were applied to evaluate classification accuracy. The results showed that Conv-2D performed best in local classification experiments with OA of 93.14% and F1 score of 0.8552 in Banan District, OA of 92.53% and F1 score of 0.8399 in Zhongxian County. CNN-based models except Conv-1D provided more desirable performance than non-CNN classifiers. Besides, among the non-CNN classifiers, XGBoost received the best result with OA of 89.73% and F1 score of 0.7742 in Banan District, SVM received the best result with OA of 88.57% and F1 score of 0.7538 in Zhongxian County. In model transferability experiments, almost all CNN classifiers had low transferability. RF and XGBoost models have achieved acceptable F1 scores for transfer (RF = 0.6673 and 0.6469, XGBoost = 0.7171 and 0.6709, respectively).

Graphical Abstract

1. Introduction

Food security has become a global and international topic [1,2]. Rice, as one of the most important staple foods, is widely grown in Southern China. However, rice agriculture in China faces great challenges in the coming decade. The growing population has triggered an increase in the demand for rice [3]. Additionally, the acceleration of urbanization may create continual pressure in the rice cultivation area [4]. Other adverse factors such as droughts, soil degradation and economic restructuring should not be overlooked when studying rice production patterns [5]. Rice cultivation plays an important role in water resources utilization as well, since about a quarter to one-third of the world’s freshwater resources are used for rice irrigation [6]. Rice paddies have been identified as an important source of methane (CH4), which has a significant impact on the global greenhouse effect [7]. Therefore, it is essential to map the spatial distribution and planting area of paddy rice at a large scale for guiding rice production, water utilization, climate change and government policy decisions.
Traditionally, the large-scale rice mapping, which relies on massive ground surveys, is time-consuming, laborious and uneconomical. Satellite remote sensing data can provide timely, objective and accurate agricultural information at a regional to global scale [8,9]. Optical data are the most common and easily available satellite imagery data, and have found wide applications in spatial distribution extraction [10,11,12], cropping intensity [13] and the phenology [14] of a rice paddy. The identification of a rice paddy usually requires spectral bands, spectral indices (e.g., the land surface water index (LSWI) [15], normalized difference vegetation index (NDVI) [16], modified normalized difference water index (MNDWI) [17] and enhanced vegetative index (EVI) [18]) or their combinations, which are sensitive to the differences of spectral features between rice paddy and other classes [16,19]. However, the direct distinction of rice and other objects based on a certain phenological phase may be undesirable due to the spectral similarity of the different objects [20]. Extensive research has confirmed the effectiveness of multi-temporal analysis in rice paddy mapping during the specific growth phases [21,22,23]. For example, a pixel- and phenology-(the flooding/open-canopy) based rice growing area extraction algorithm was successfully tested in the Sanjiang Plain, China [21]. Cui et al. used NDVI, EVI and LSWI indexes during flooding/transplanting and ripening phases to identify rice fields, the user and producer accuracies of rice paddy were 90% and 94%, respectively [22].
Over the last decades, MODIS [16,18], Landsat TM/ETM+ [19,22] and GF-1 [24] have been usually used as a source of remote sensing image data in rice paddy mapping, but their spatial–temporal resolution is not sufficient to perform more detailed vegetation classification in areas with complex terrain. Launched by the European Space Agency (ESA) in 2015, the Sentinel-2 remote sensing satellite has a maximum spatial resolution of 10 m and a high frequency of 5 days at the equator [25]. Sentinel-2A optical data have been recently performed in crop classifications [26], water body mapping [27] and crop yield forecasting [28]. The excellent spatial and temporal resolution of the Sentinel-2 can ensure both the acquisition of critical phenological images of rice and the production of more accurate rice maps.
In recent years, many machine learning algorithms, such as decision trees (DTs) [29], support vector machines (SVMs) [30], random forest (RF) [30] and artificial neural networks (ANNs) [31], have been applied to map rice or other crop types. Currently, deep learning (DL) algorithms inspired by how the human brain recognizes and recalls information without outside expert input to guide the process have become a hot topic in the machine learning area [32]. One of the popular deep learning algorithms, convolutional neural networks (CNNs), had exhibited excellent performance levels in computer vision related contests such as ImageNet [33]. CNNs have been applied by the remote sensing (RS) community and shown advantages in pixel-based classification [19,34,35,36,37,38], object detection [39], semantic segmentation [40] and other tasks. Regarding pixel-based classifier, CNNs generally perform better than traditional machine learning algorithms. For example, Zhang et al. [19] showed that the Kappa coefficients of CNN are 7% and 9% higher than SVM and RF classifiers, respectively. CNNs received highest classification accuracy of 91% in the extraction of urban built-up areas [41].
Most traditional machine learning algorithms extract low-level features from the original images (e.g., spectral, terrain and texture). However, using convolutional kernels and the max-pooling method, CNNs can automatically extract high-level features from the original data [42,43], which are more robust and secure for the classification task and thus improve classification accuracy. There are many studies on crop classification using CNNs [19,42,44] according to the following steps: (1) building a dataset by a w × w (w is the size of window) neighboring region from the original image, (2) extracting high-level features from the database by convolution on the spectral, the spatial or the spectral-spatial directions, (3) hierarchical model training and (4) model testing and classification.
Generally, there have been quite a few previous rice paddy or other land cover types classification efforts where various machine learning algorithms were leveraged [19,34,35,36,37,38,41,44]. For example, Zhang et al. [19] used the convolutional neural network to map the spatial distribution of rice in the Dongting Lake Area, China. Zhong et al. [37] utilized remotely sensed time series to classify crops in Yolo County, California based on machine learning models (including RF, SVM, CNN, etc.). Chen et al. [36] extracted urban water bodies from high-resolution remote sensing images based on deep learning. However, most studies lack the comparison of classification effectiveness and efficiency between CNNs and other traditional machine learning algorithms and their transferability.
The main goals of the current study were to (1) explore the feasibility of models for mapping rice paddy with massive labeled data and (2) evaluate the transferability of pretrained classifiers from one study area to another. Specifically, the rice area extraction models were constructed and compared by convolution on three different directions (spectral, spatial and spectral-spatial (both spectral and spatial)) by using Sentinel-2 multi-temporal image dataset in two study areas. Meanwhile, four popular non-CNN models were used for comparison. Finally, the best classification model was selected for mapping a rice paddy.

2. Study Area and Materials

2.1. Study Area

Our two study areas were the Banan district (106°26′–106°59′E, 29°7′–29°45′N) and the county of Zhongxian (107°3′–108°14′E, 30°3′–30°35′N) located in the hilly areas of Chongqing, Southwestern China (Figure 1). Elevations of Banan District and Zhongxian County vary between 154 and 1132 m, and between 117 and 1680 m, respectively. Both of them have a subtropical humid climate. The average annual temperature of Banan District and Zhong County is about 18.7 °C. Annual total precipitation is between 1000 and 1200 mm, which is mainly concentrated in May–July. The frost-free period is more than 300 days. Rice is the main food crop with similar phenology in both regions. Each year, rice is transplanted in April, heading and flowering in July, and harvested in September.

2.2. Datasets

2.2.1. Sentinel-2A Dataset

The available Sentinel-2A data were downloaded from Copernicus Open Access Hub (https://eros.usgs.gov/sentinel-2). The cloud-free (<2% cloud cover) images were acquired within rice growth season (April to October) in 2017. Finally, three images on 14 April 2017 (representing vegetative phase), 23 July 2017 (representing reproductive phase) and 31 October 2017 (representing post-harvest phase) were used for each study area, respectively. Sentinel-2 data were preprocessed using Sen2Cor plug-in and SNAP software, which mainly consists of atmospheric correction and resampling. The downloaded L1C-level top of the atmosphere (TOA) reflectance data were converted into L2A-level bottom of the atmosphere (BOA) reflectance data by the ESA’s Sen2Cor plug-in. Then Sentienl-2 data were resampled to a 10 m resolution by bilinear resampling. Subsequently the original 60 m resolution bands and red-edge bands were culled with SNAP. Finally, six bands (Band 2, Band 3, Band 4, Band 8, Band 11 and Band 12) were used in this work.
Additionally, four spectral indices including normalized difference vegetation index (NDVI), land surface water index (LSWI), modified soil adjusted vegetation index (MSAVI) and modified normalized difference water index (MNDWI) were calculated according to Equations (1)–(4). NDVI is a widely used vegetation index and it is susceptible to soil background noise in low vegetation-covered areas [16]. LSWI is sensitive to the radiation equivalent water thickness and water content of leaf [15]. MSAVI is a vegetation index proposed to reduce the effects of soil noise [45]. MNDWI strengthens the characteristics of open water bodies, and effectively suppresses or even eliminates noise from buildings, vegetation and soil [17].
N D V I = B 8 B 4 B 8 + B 4
L S W I = B 8 B 11 B 8 + B 11
M S A V I = 0.5 × ( 2 × ( B 8 + 1 ) ( 2 × B 8 + 1 ) 2 8 × ( B 8 B 4 ) )
M N D W I = B 3 B 11 B 3 + B 11
Therefore, a total of 30 features derived from remote sensing images of three periods were tested to map rice paddy over the study areas, and each period includes 6 spectral bands and 4 spectral indices.

2.2.2. Training Samples

In order to avoid the influence of mixed pixel samples on classification accuracy, pure pixel samples were selected by 1:10,000 land-use map in 2017 and Google Earth Online image (Google Earth Pro v7.3.2.5776). This land-use map is generated using the visual interpretation approach and is available from the Chinese Land Resources Bureaus (http://www.mnr.gov.cn/). Its first version was generated in 2008 and renewed every year by the local land resources bureaus [19,46,47]. The land-use map comes from the 2017 land use utilization database of Zhongxian County and Banan District, and the report accuracy is over 90%. Here, a grid is a pure pixel if the grid itself and the surrounding pixels within a certain window size belong to the same land-use type. First, the land-use polygons were converted to raster format with a resolution of 10 m by ArcGIS 10.3. Second, consider a w × w square grid window such that the current pixel in consideration is at the center of it. The current pixel is considered “pure” if all w × w pixels in the grid have the same land cover category as the current pixel (at the center). This step was implemented in GDAL (Geospatial Data Abstraction Library, version 2.2.2) and python scripting language. Refer to the supplementary (i.e., Section 1.1 and Section 1.2) for details of this process.
A value of w that is too large might only produce samples distributed in large plots, while the w value that is too small will reduce the accuracy of pure pixel sample extraction. In Banan District and Zhongxian County, the length and width of the paddy field is usually between 50 and 100 m (representing 5–10 pixels at a 10-m spatial resolution). Therefore, w was fixed at 7. After manual correction based on Google Earth Online image, 6359 pure pixels including 1559 rice and 4800 non-rice grids were obtained in Banan District. In Zhongxian County, 6400 pure pixels were obtained in the same way including 1600 rice and 4800 non-rice grids. These data were used as training samples for model construction (Table 1).

2.2.3. Validation Samples

Validation data for this study was assembled by manual inspection of Google Earth satellite images. First, we randomly distributed a set of points with a minimum spacing of 490 m between these points. Then, these points were classified as Rice/Non-Rice by manual inspection via the Google earth interface (imagery provided by TerraMetrics, CNES, etc.). Points that had very ambiguous land covers were omitted. Finally, 1575 pixels in Banan District and 1566 pixels in Zhongxian County were generated for the validation of resultant paddy rice map (Figure 1). Table 1 shows the sample division of the two study areas.

3. Method

3.1. CNN Classifiers

Deep neural networks are composed of a number of hidden layers and are divided into full-connected neural network/multilayer perceptron (FC/MLP) convolutional neural networks (CNNs) and recurrent neural networks (RNNs). CNNs have recently become a popular deep neural networks and have achieved good success in computer vision and image understanding such as ImageNet contest. In convolutional neural networks, every image input is treated as a matrix of pixel values, which represents the gray values at a given pixel in the image. Unlike traditional neural networks, which treat an image as a one-dimensional network, the CNNs consider the spatial relevance between each pixel and its neighboring pixels. CNNs consist of the following layers: A. Input layer, B. Convolutional layer, C. Activation layer, D. Pooling layer, E. Full-connected layer and F. Output layer. The meaning and calculation formula of each layer can refer to [48], the following is only a brief introduction.
  • Input layer: Used for data input.
  • Convolutional layer: In this layer, a kernel (or weight) matrix is used to extract high level features from the images. The kernel with its weights rotates over the image matrix in a sliding window in order to obtain the convolved output. The kernel matrix behaves like a filter in an image extracting particular information from the original image matrix. During the convolution process, the weights are learnt such that the loss function is minimized.
  • Activation layer: The role of the activation layer is to improve the expression capacity of the entire network by increasing nonlinear mapping, and the rectified linear unit (ReLU) f (x) = max (x, 0) is commonly chosen to be an activation function.
  • Pooling layer: Pooling layers are used to extract the most informative features from the generated convolved output.
  • Full-connected layer: Generally located at the end of CNNs, connecting all features and outputting all values to the classifier.
  • Output layer: To generate the final output.
According to the dimension of the convolutional filter, CNNs may be categorized into one-dimensional convolutional neural network (Conv1D), two-dimensional convolutional neural network (Conv2D) and three-dimensional convolutional neural network (Conv3D). Previous studies mostly focused on Conv1D and Conv2D [19,43,44,49]. In this work, Conv1D and Conv2D as well as Conv3D classifiers are constructed to map rice paddy and they are briefly described below:
  • A. Conv-1D classifier
Conv1D, which is a most simplistic form of CNNs, is often used for sequence datasets. This model can be used for extracting local one-dimensional subsequences from the input sequences and identify local patterns within the window of convolution. Spectral information of each pixel is considered in this case. Figure 2a shows the structure of Conv1D, which consists of two convolutional layers, one flatten layer, two full-connected layers, as well as one batch-normalization layer and one dropout layer to prevent overfitting. The popular activation function called rectified linear unit (ReLU) is used to increase the nonlinear expression of neural networks and solve the vanishing gradient problem.
  • B. Conv-2D classifier
Two-dimensional convolutional filters are widely used for image analyses. The main idea of Conv-2D is that the convolutional filter moves in 2-directions (X, Y) to calculate higher-level features from the set of lower-level feature maps. In contrast to the Conv1D, the spatial relationships between pixels rather than spectral information are calculated. In the case of Sentinel-2 (10-m spatial resolution), we have found that the window size of 7 × 7 is able to capture spatially local correlation of a center pixel to the surrounding pixels and limit heterogeneous pixels. Figure 2b illustrates the structure of Conv-2D model. The model includes one flatten layer, two convolutional layers, two maxpooling2D layers and two dense layers. To prevent overfitting, one dropout and one batch-normalization layers are added. Conv-2D classifier also uses ReLU as the activation function.
  • C. Conv-3D classifier
In this section, a CNN with 3D convolution is designed for both spatial and spectral features. A three-dimensional kernel is applied to the cube formed by stacking multiple contiguous feature maps together and moved in three-directions (X, Y and Z) to calculate the high-level feature representations. Figure 2c,d show two Conv-3D architectures (Conv-3D_1 and Conv-3D_2). The inputs (X, Y, Z, Channel) of the two models are different, where X and Y are the size of window, which is consistent with the size of the convolution kernel in Conv-2D, Z is the number of bands for a cube and Channel is the number of cubes. Conv-3D_1 uses three cubes according to three periods, each cube having 10 bands. Conv-3D_2 employs one cube that is composed of 30 features. The neural network structures of the two models are the same except for the input layer, which consists of one flatten layer, two convolutional layers, two full-connected layers and several overfitting layers.

3.2. Other Classifiers

Four non-CNN classifiers including random forest (RF), extreme gradient boosting (XGBoost), support vector machine (SVM) and multilayer perceptron (MLP) are compared with CNNs for mapping a rice paddy. These classification algorithms are known for high performance and have been widely applied in remote sensing community [30,31,50]. RF is an ensemble learning algorithm for the classification and regression based on the bagging technique, which is suitable for processing high-dimensional data and preventing overfitting [50,51]. XGBoost is an extensible end-to-end boosting system, which is widely used in many machine learning competitions and designed as the main component of solutions [52]. SVM is a discriminant classifier for separating hyperplane, in which the input vectors are nonlinearly mapped into a very high dimensional feature space by using kernel functions [53]. MLP is a plain deep neural networks, in which neurons of a layer are fully connected with all neurons of neighboring layers [54].
The models are developed using the Scikit-learn and XGBoost packages established in Python. A grid search method embedded in Python (GridSearchCV) was applied to optimize the models’ parameters. The optimal parameter combinations for RF, SVM, XGBoost and MLP are given in Table 2. The models were executed on a computer with Intel Core i5-9400F processor, NVIDIA GeForce RTX 2060 (6 GB), and RAM 8 Gb. CNN classifiers were run using Python v3.6.6, Tensorflow-gpu v1.13.1 and Keras v2.2.4.

3.3. Model Performance

The classifier performance was evaluated by the overall accuracy (OA), recall, precision and F1 score calculated from the confusion matrix (Table 3). Overall accuracy is a widely used indicator in classification, which represents the ratio between the model and the total number predicted on all test sets. Precision is the fraction of retrieved instances that are relevant, while recall is the fraction of relevant instances that are retrieved. For original sample and predicting results, recall and precision are the ratio of the correct predictions and the total number of actual items or predicted items in the set, respectively. Generally, precision and recall are a pair of contradictory measures, when one is higher, the other is often lower. The F1 score is the harmonic average of the precision and recall with 1 indicating the best and 0 denoting the worst.
The OA, recall, precision and F1 score were calculated using Equations (5)–(8), respectively:
O A = T P + T N T N + T P + F N + F P
R e c a l l = T P T P + F N
P r e c i s i o n = T P T P + F P
F 1 s c o r e = 2 × R e c a l l × P r e c i s i o n R e c a l l + P r e c i s i o n

4. Results and Analysis

4.1. Model Performance at Each Site

Two multi-temporal data sets (30 bands total) were used to train and test CNNs and non-CNN classifiers for the two study areas. Model performance was evaluated at each site (Table 4).
For CNN classifiers, F1 scores of Conv-1D, Conv-2D, Conv-3D_1 and Conv-3D_2 in Banan District were 0.7544, 0.8552, 0.8067 and 0.8196, respectively (Table 4a) and F1 scores of these classifiers in Zhongxian County were 0.7451, 0.8399, 0.8260 and 0.8321, respectively (Table 4b). In Banan District, Conv-2D gave the highest values of recall (0.8264), OA (0.9314) and F1 score (0.8552) while Conv-3D_2 yielded the highest value of precision (0.8906). In Zhongxian County, Conv-2D achieved the highest values of precision (0.8899), OA (0.9253) and F1 scores (0.8399) while Conv-1D gave the highest values of recall (0.9585). However, Conv-2D had the lowest values of recall (0.7953). Besides, Conv-2D and Conv-3D_2 presented similar values of F1 score, but Conv-2D had a lower training time (230 s) and testing time (2320 s) than Conv-3D_2 (training time = 333 s and testing time = 3626 s). Moreover, both in Banan District and Zhongxian County, Conv-1D had the lowest values of precision (0.7827 and 0.5929), OA (0.8838 and 0.8340) and F1 score (0.7544 and 0.7451), indicating that the convolution in the spectral direction only might be unsuitable for rice planting region extraction over the study site.
For non-CNN classifiers, in Banan District, the XGBoost classification model showed the best performance indicated by both accuracy and operation time. This model has the highest recall (0.7150), overall accuracy (0.8973) and F1 score (0.7742) and the lowest training time (0.4933 s) and testing time (113 s). RF and SVM classifiers also gave higher classification accuracy than MLP classifier with F1 scores of 0.7648 and 0.7515, respectively. Besides, the RF classifier was more efficient than the SVM classifier. The testing time of the RF classifier was 759 s, which was much lower than that of SVM (1963 s). MLP model presented the lowest values of recall (0.4482), overall accuracy (0.8571) and F1 score (0.6070) and the highest value of training time (203.48 s). In Zhongxian County, the SVM classifier gained the highest overall accuracy (0.8857) and F1 score (0.7538). MLP presented the lowest accuracy value of precision (0.7293), OA (0.8704) and F1 score (0.7414) and the highest value of precision (0.7539). Besides, RF presented the lowest value of recall (0.6917) and the highest value of precision (0.8042).
On average, CNN classifiers produced higher values of recall, precision, OA and F1 scores than non-CNNs. Among the evaluated methods, the Conv-2D model had the highest accuracy indicated by OA and F1 score. Besides, the operation time (training time and testing time) of the Conv-2D classifier was also acceptable. The two maps of rice paddy of Banan District and Zhongxian County produced by the Conv-2D classifier are shown in Figure 3a,b. In order to compare the detailed mapping information created by the evaluated models, two subregions were selected randomly and illustrated in Figure 4.
Figure 4a,b shows two subregions (5 km × 5 km) of rice paddy maps generated by the eight classifiers in Banan District (Subregion A) and Zhongxian County (Subregion B), respectively. To give a better visual comparison, intersect and erase calculations were performed using ArcGIS on the reference land use map and our classification result. Green color represents the overlap areas between our classification and the reference maps, and red and yellow colors represent missing classification pixels (i.e., actual rice pixels were not identified) and wrong classification pixels (i.e., actual non-rice pixels were identified as paddy pixels), respectively.
Non-CNN classifiers generated massive isolated rice pixels. These pixel-based classification results tended to possess “salt and pepper” noise. Besides, the classification performance of MLP was unstable, which may seriously underestimate the distribution of rice areas in Banan District, but performs well in Zhongxian County. The classification maps of Conv-2D, Conv-3D_1 and Conv-3D_2 were similar. Moreover, compared with the other models, Conv-1D greatly overestimated the distribution of rice planting areas in Subregion B. Furthermore, the classification maps of Conv-2D, Conv-3D_1 and Conv-3D_2 were relatively smoother than that of non-CNNs and a Conv-1D classifier. The main reason may be the use of patch-based method, which selects the image block centered on the target pixel as the input of the classifier, and the spatial associations and semantic links between pixels in the image block are helpful to produce homogeneous and smooth images. Unfortunately, eight classifiers did not perform well in the extraction of narrow rice paddies. Although Conv-2D, Conv-3D_1 and Conv-3D_2 achieved sound classification results but could not accurately identify overly narrow rice paddies.

4.2. Model Transferability

To estimate the transferability of the machine learning classification models, two experiments (A and B) were conducted. Experiment A computed rice paddies in Zhongxian County using the trained classification models based on the dataset of Banan District. Experiment B extracts rice paddies in Banan District using the trained classification models based on the dataset of Zhongxian County. The transferability evaluation results are shown in Table 5.
The classification models with reasonable F1 score of higher than 0.6 in Experiment A included Conv-2D (0.6458), RF (0.6673) and XGBoost (0.7171). Figure 4b shows that the number of yellow pixels in the result maps was larger than the reference map and maps produced by local models. Moreover, by comparing Table 4 and Table 5, the transfer classification accuracy of Conv-2D, RF and XGBoost was lower than that of the local models. In Experiment B, XGBoost presented the highest F1 score of 0.6709 followed by RF (0.6469). However, the values of recall (0.5078, 0.5466) were much lower than the values of precision (0.8909, 0.8683), indicating that the transferred models ignored a large number of actual rice pixels (the number of red pixel increases compared to reference map and result maps produced by local models).
Almost all CNN classifiers had low transferability. According to the F1 score, the Conv-2D model gave the reasonable classification accuracy in Experiment A. However, a huge imbalance between the values of recall (0.9637) and precision (0.4856) indicates that the model recognizes a large number of non-rice pixels as rice paddies (Conv2D-T in subregion B having a large number of yellow pixels). For non-CNN models, SVM and MLP classification models showed low transferability in terms of the performance indicators. RF and XGBoost classification models could be transferred to other research areas, although the transferable accuracy was lower than the classification accuracy at the local site.

5. Discussion

5.1. Evaluation of Classifier’s Local Classification Accuracy

5.1.1. Performance Comparison of CNNs

Deep learning algorithms, which learn the internal rules and presentation hierarchy of sample data, have outperformed most traditional machine learning in land-use or crop classifications [34,37,42,55]. In particular, as the most representative deep learning algorithm, CNNs have been widely used in remote sensing community and have achieved a good performance [19,34,35,42,44]. In this study, three CNN classifiers (Conv-2D, Conv-3D_1 and Conv-3D_2) that used the patch-based method gave higher accuracy than others. RF, XGBoost, SVM, MLP and Conv-1D are pixel-based, where land-use classes are assigned to each individual pixel allocated. For large-scale classification, the spatial context plays a major role and provides inherent information about the target pixels. It is difficult to adequately describe the feature information of the ground object with only the single-pixel independent spatial information. In this case, the well-trained pixel-based classification models may not be robust to resist the salt-and-pepper noise and fragmentation, which was clearly shown in Figure 4a. The patch-based method produces image objects (patches) for model input. Spatial relationships and semantic links between pixels can be explored with the patch-based method, resulting in more homogeneous images. Prior studies have demonstrated the importance of patch-based methods [19,41]. Further, two-dimensional and three-dimensional convolutional neural networks have advantages in rice area extraction due to the use of patch-based methods. Zhang and Lin applied the CNN model to extract rice planting areas and the overall accuracy of the CNN is 6.43% and 7.68% higher than that of RF and SVM, respectively [19]. Similar results were obtained in South China with the overall accuracy of CNN and SVM being 93.60% and 91.05%, respectively [56]. CNN is also used in identifying and mapping of other geo-objects [57]. Two-dimensional convolutional neural networks have the highest classification accuracy of 91% for identifying urban built-up areas [41].
The use of convolution is another reason for improving the performance of Conv-2D, Conv-3D_1 and Conv-3D_2. Convolution extracts and creates complex features and multilevel convolution can ensure that the model has sufficient expressive and generalization ability [42,55,56,58]. Chen et al. evaluated the effectiveness of convolution by analyzing the intermediate activation feature maps for comparing the similarity in the same class and the divisibility between different classes [58]. They found that convolutional direction or scenario was closely related with model performance. In both experimental areas of the current work, Conv-1D obtained the lowest F1 score among the CNNs. This is in agreement with the result of Kestur, R., et al. [49]. However, Conv-1D has achieved good performance in land use classification based on single- or multi-temporal hyperspectral datasets [49,58] and vegetation index (NDVI, EVI, etc.) time-series data [37], for example, Chen et al. [58] and Guidici et al. [49] reported that the Conv1D architectures can be applied to hyperspectral data and achieve high classification accuracy. This is mainly because hyperspectral datasets used by Chen et al. [58] and Guidici et al. [49] have inherited sequential interdependence, the spectral features in a single pixel are closely linked, allowing for more beneficial features to be convolved in an integrated machine learning framework. However, the data used in that article have lower sufficient sequence dependencies. Convolution only in the spectral direction may introduce more noise. This could explain why Conv-3D_1 and Conv-3D_2 classifiers gave slightly less accurate results than Conv_2D, since Conv_3D_1 and Conv_3D_2 are convoluting simultaneously in the spectral and spatial directions.

5.1.2. Performance Comparison of Non-CNNs

XGBoost had the best classification performance among all the non-CNN classifiers in Banan District. XGBoost is considered by many scholars or contestants as one of the best algorithms for processing non-perceptual data [50,59]. Recently, XGBoost has obtained considerable success in remote-sensing data analysis [37,50,60]. Zhong used Landsat enhanced vegetation index (EVI) time series data for crop classification, and XGBoost achieved a better result than SVM and RF [37]. Joharestani used XGBoost, RF and MLP to predict the PM 2.5 distribution in Tehran’s urban area and XGBoost obtained the best model performance with R2 of 0.81 [50]. SVM produced the highest F1 score for mapping rice paddies among the four non-CNN classifiers in Zhongxian County. The efficient performance of SVM has been confirmed in some comparative studies [50,57]. Abdi used multi-temporal Sentinel-2 data for land cover and land use classification in complex boreal landscapes based on a variety of non-parametric algorithms. SVM achieved the highest overall accuracy of 0.758, followed by XGBoost of 0.751 [59]. However, SVM had the longest testing time in non-CNN classifiers, reflecting its inherent shortcomings of low operating efficiency when generalized to new samples outside the training dataset in large datasets. In the two local experiments of this paper, RF performed moderately and stably. Rodriguez-Galiano has verified the high performance and robustness of RF in land-use classification based on several criteria: sensitivity to data set size, mapping accuracy and noise [57].

5.2. Comparison of Model Transferability

Convolutional neural networks (except one-dimensional convolution) performed well at local areas, but worst at other sites. Advanced deep neural networks have achieved impressive results on many image classification tasks [19,34,35,48]. However, deep learning models are fragile and unstable, and their generalization ability has always been a challenge. In fact, subtle and imperceptible perturbations of the data samples are sufficient to fool deep learning classifiers and result in incorrect classification [61]. A classic study is to superimpose noise that is not perceivable by human eyes (which does not affect human recognition) on panda photos, but the neural network model identifies it into other categories [62]. In other words, the similarity between the training dataset and the testing dataset could severely affect the performance of deep learning-based methods, and if the new inputs differ from what the net saw at training time, the model might break in absurd ways. In the remote sensing community, studies have confirmed that models trained at one site may not be applicable to other areas due to radiometric variation of source and target datasets [63,64]. Moreover, CNN classifiers are based on patches, which extract high-dimensional features by inputting presegmented sample blocks and then perform model training. The sample blocks contain the spatial context information of the geo-objects, and the differences in the morphology of the geo-objects between the source domain and target domain might be detrimental to the model transfer. Rice paddies in Banan District are relatively narrow and scattered than that of Zhongxian County. Consequently, in this study, massive interference might reduce the transferability of models due to the radiometric variation and morphological difference between source and target datasets.
The tree-based ensemble learning algorithms, RF and XGBoost, performed well in the transfer experiments. Ensemble learning builds and combines multiple learners to complete learning tasks, and often can obtain significantly better generalization capabilities than individual learners (such as C4.5 decision tree algorithms, SVM, MLP, etc.) and reduce the risk of falling into a local optimum. Rodriguez-Galiano’s research shows that the RF outperforms the decision tree in noise immunity [57]. The basic classifiers of RF and XGBoost are both decision trees, which subdivide the input data according to the condition judgment and have stronger anti-interference ability than SVM and MLP.

5.3. Deficiencies and Prospects

Conv-2D classifiers in this article gave satisfactory classification accuracy, however there might be numerous mixed pixels in the result maps. The future work is considering combining deep learning and spectral unmixing for subpixel-level land use/cover classification. Moreover, the potential of CNNs cannot be fully exploited because of the inconsistent size between the original image and the ever-decreasing feature image. Some deep or ensemble learning architectures that perform well in image semantic segmentation, such as U-Net [62], PSP-Net [65], and HRNet [66], and extremely randomized trees [67,68], could provide alternatives to extract geo-objects. Besides, how to improve the transferability of CNN models is worthy of in-depth research. Some studies have attempted to increase the transferability and generalization capabilities of models through transfer learning or data augmentation techniques. Data augmentation resamples the original input, including geometric and/or radiosity transformations to expand the sample space and improve the generalization ability of the models [63]. Transfer learning uses the knowledge learned from the source data domain to solve a new or related task in the target domain, which is expected to solve the problem of training reliable models without sufficient labeled data. A common used idea of transfer learning is to obtain the feature representations, which are close to the source data domain for model training in the target domain [69].

6. Conclusions

We developed deep-learning CNNs with various architectures and four popular non-CNN models to extract rice growing areas using multi-temporal sentinel-2 data in Banan District and Zhongxian County. The most satisfactory classification result was achieved by a two-dimensional convolutional neural network (Conv-2D) in both study areas. According to F1 scores and overall accuracy, Conv-2D was significantly superior to popular Non-CNN classifiers such as random forest, XGBoost and support vector machine as well as one-dimensional convolutional neural network (Conv-1D). The proposed method gave overall accuracy and F1 scores of higher than 95% and 0.90, respectively. Moreover, among the non-CNN classifiers, SVM and XGBoost received the best result in Banan District and Zhongxian County, respectively. Besides, two experiments were designed to study the transferable performance of eight models. Almost all CNN classifiers had low transferability, while RF and XGBoost models achieved acceptable transferability in both experiments.

Supplementary Materials

The following are available online at https://www.mdpi.com/2072-4292/12/10/1620/s1, Figure S1: Array slicing algorithm; Figure S2: An example of pure pixel extraction procedure; Code based on python and GDAL; Figure S3: Training samples and validation samples of Banan District; Figure S4: Training samples and validation samples of Zhongxian County.

Author Contributions

Methodology, W.Z., H.L. and W.W.; Resources, L.Z.; Software, W.Z.; Supervision, H.L.; Writing—original draft, W.Z.; Writing—review and editing, H.L., W.W. and J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Scarrow, R. Food security Ghana rice revolution falters. Nat. Plants 2017, 3, 17111. [Google Scholar] [CrossRef] [PubMed]
  2. Godfray, H.C.J.; Beddington, J.R.; Crute, I.R.; Haddad, L.; Lawrence, D.; Muir, J.F.; Pretty, J.; Robinson, S.; Thomas, S.M.; Toulmin, C. Food security: The challenge of feeding 9 billion people. Science 2010, 327, 812–818. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Sutrisno, D.; Ambarwulan, W.; Turmudi, I.N.; Suryanta, J.; Windiastuti, R.; Kardono, P. Cellular automata markov method, an approach for rice self-sufficiency projection. J. Ecol. Eng. 2019, 20, 117–125. [Google Scholar] [CrossRef]
  4. Wang, Y.S. The challenges and strategies of food security under rapid urbanization in China. Sustainability 2019, 11, 542. [Google Scholar] [CrossRef] [Green Version]
  5. Gandhi, V.P.; Zhou, Z.Y. Food demand and the food security challenge with rapid economic growth in the emerging economies of India and China. Food Res. Int. 2014, 63, 108–124. [Google Scholar] [CrossRef]
  6. Bouman, B.A.M.; Humphreys, E.; Tuong, T.P.; Barker, R. Rice and water. Adv. Agron. 2007, 92, 187–237. [Google Scholar]
  7. Chen, H.; Zhu, Q.A.; Peng, C.H.; Wu, N.; Wang, Y.F.; Fang, X.Q.; Jiang, H.; Xiang, W.H.; Chang, J.; Deng, X.W.; et al. Methane emissions from rice paddies natural wetlands, lakes in China: Synthesis new estimate. Glob. Chang. Biol. 2013, 19, 19–32. [Google Scholar] [CrossRef]
  8. Kirby, R.P. Remote-sensing and image interpretation—Lillesand, Tm, Kiefer, Rw. Scot. Geogr. Mag. 1980, 96, 124–125. [Google Scholar]
  9. Gong, P.; Wang, J.; Yu, L.; Zhao, Y.C.; Zhao, Y.Y.; Liang, L.; Niu, Z.G.; Huang, X.M.; Fu, H.H.; Liu, S.; et al. Finer resolution observation and monitoring of global land cover: First mapping results with Landsat TM and ETM+ data. Int. J. Remote Sens. 2013, 34, 2607–2654. [Google Scholar] [CrossRef] [Green Version]
  10. Sun, P.J.; Zhang, J.S.; Zhu, X.F.; Pan, Y.Z.; Liu, H.L. A highly efficient temporal-spatial probability synthesized model from multi-temporal remote sensing for paddy rice identification. Eur. J. Remote Sens. 2017, 50, 98–110. [Google Scholar] [CrossRef] [Green Version]
  11. Motohka, T.; Nasahara, K.N.; Miyata, A.; Mano, M.; Tsuchida, S. Evaluation of optical satellite remote sensing for rice paddy phenology in monsoon Asia using a continuous in situ dataset. Int. J. Remote Sens. 2009, 30, 4343–4357. [Google Scholar] [CrossRef]
  12. Zhang, G.L.; Xiao, X.M.; Dong, J.W.; Kou, W.L.; Jin, C.; Qin, Y.W.; Zhou, Y.T.; Wang, J.; Menarguez, M.A.; Biradar, C. Mapping paddy rice planting areas through time series analysis of MODIS land surface temperature and vegetation index data. ISPRS J. Photogramm. Remote Sens. 2015, 106, 157–171. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Peng, D.L.; Huete, A.R.; Huang, J.F.; Wang, F.M.; Sun, H.S. Detection and estimation of mixed paddy rice cropping patterns with MODIS data. Int. J. Appl. Earth Obs. Geoinf. 2011, 13, 13–23. [Google Scholar] [CrossRef]
  14. Sakamoto, T.; Van Nguyen, N.; Ohno, H.; Ishitsuka, N.; Yokozawa, M. Spatio-temporal distribution of rice phenology and cropping systems in the Mekong Delta with special reference to the seasonal water flow of the Mekong and Bassac rivers. Remote Sens. Environ. 2006, 100, 1–16. [Google Scholar] [CrossRef]
  15. Xiao, X.; Boles, S.; Frolking, S.; Salas, W.; Moore, B.; Li, C.; He, L.; Zhao, R. Observation of flooding and rice transplanting of paddy rice fields at the site to landscape scales in China using VEGETATION sensor data. Int. J. Remote Sens. 2002, 23, 3009–3022. [Google Scholar] [CrossRef]
  16. Xiao, X.M.; Boles, S.; Liu, J.Y.; Zhuang, D.F.; Frolking, S.; Li, C.S.; Salas, W.; Moore, B. Mapping paddy rice agriculture in southern China using multi-temporal MODIS images. Remote Sens. Environ. 2005, 95, 480–492. [Google Scholar] [CrossRef]
  17. Xu, H.Q. Modification of normalised difference water index (NDWI) to enhance open water features in remotely sensed imagery. Int. J. Remote Sens. 2006, 27, 3025–3033. [Google Scholar] [CrossRef]
  18. Xiao, X.M.; Boles, S.; Frolking, S.; Li, C.S.; Babu, J.Y.; Salas, W.; Moore, B. Mapping paddy rice agriculture in South and Southeast Asia using multi-temporal MODIS images. Remote Sens. Environ. 2006, 100, 95–113. [Google Scholar] [CrossRef]
  19. Zhang, M.; Lin, H.; Wang, G.X.; Sun, H.; Fu, J. Mapping paddy rice using a Convolutional Neural Network (CNN) with landsat 8 datasets in the Dongting Lake Area, China. Remote Sens. 2018, 10, 1840. [Google Scholar] [CrossRef] [Green Version]
  20. Wang, W.; Yao, X.; Tian, Y.C.; Liu, X.J.; Ni, J.; Cao, W.X.; Zhu, Y. Common spectral bands and optimum vegetation indices for monitoring leaf nitrogen accumulation in rice and wheat. J. Integr. Agric. 2012, 11, 2001–2012. [Google Scholar] [CrossRef]
  21. Qin, Y.W.; Xiao, X.M.; Dong, J.W.; Zhou, Y.T.; Zhu, Z.; Zhang, G.L.; Du, G.M.; Jin, C.; Kou, W.L.; Wang, J.; et al. Mapping paddy rice planting area in cold temperate climate region through analysis of time series Landsat 8 (OLI), Landsat 7 (ETM+) and MODIS imagery. ISPRS J. Photogramm. Remote Sens. 2015, 105, 220–233. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Jin, C.; Xiao, X.M.; Dong, J.W.; Qin, Y.W.; Wang, Z.M. Mapping paddy rice distribution using multi-temporal Landsat imagery in the Sanjiang Plain, northeast China. Front. Earth Sci. 2016, 10, 49–62. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Shiu, Y.S.; Lin, M.L.; Huang, C.H.; Chu, T.H. Mapping paddy rice agriculture in a highly fragmented area using a geographic information system object-based post classification process. J. Appl. Remote Sens. 2012, 6, 63526. [Google Scholar]
  24. Liu, J.T.; Feng, Q.L.; Gong, J.H.; Zhou, J.P.; Liang, J.M.; Li, Y. Winter wheat mapping using a random forest classifier combined with multi-temporal and multi-sensor data. Int. J. Digit Earth 2018, 11, 783–802. [Google Scholar] [CrossRef]
  25. Drusch, M.; Del Bello, U.; Carlier, S.; Colin, O.; Fernandez, V.; Gascon, F.; Hoersch, B.; Isola, C.; Laberinti, P.; Martimort, P.; et al. Sentinel-2: ESA’s optical high-resolution mission for GMES operational services. Remote Sens. Environ. 2012, 120, 25–36. [Google Scholar] [CrossRef]
  26. Immitzer, M.; Vuolo, F.; Atzberger, C. First experience with sentinel-2 data for crop and tree species classifications in central Europe. Remote Sens. 2016, 8, 166. [Google Scholar] [CrossRef]
  27. Du, Y.; Zhang, Y.H.; Ling, F.; Wang, Q.M.; Li, W.B.; Li, X.D. Water bodies’ mapping from sentinel-2 imagery with modified normalized difference water index at 10-m spatial resolution produced by sharpening the SWIR band. Remote Sens. 2016, 8, 354. [Google Scholar] [CrossRef] [Green Version]
  28. Schwalbert, R.A.; Amado, T.J.C.; Nieto, L.; Varela, S.; Corassa, G.M.; Horbe, T.A.N.; Rice, C.W.; Peralta, N.R.; Ciampitti, I.A. Forecasting maize yield at field scale based on high-resolution satellite imagery. Biosyst. Eng. 2018, 171, 179–192. [Google Scholar] [CrossRef]
  29. Mishra, V.N.; Prasad, R.; Kumar, P.; Srivastava, P.K.; Rai, P.K. Knowledge-based decision tree approach for mapping spatial distribution of rice crop using C-band synthetic aperture radar-derived information. J. Appl. Remote Sens. 2017, 11, 046003. [Google Scholar] [CrossRef]
  30. Onojeghuo, A.O.; Blackburn, G.A.; Wang, Q.M.; Atkinson, P.M.; Kindred, D.; Miao, Y.X. Mapping paddy rice fields by applying machine learning algorithms to multi-temporal Sentinel-1A and Landsat data. Int. J. Remote Sens. 2018, 39, 1042–1067. [Google Scholar] [CrossRef] [Green Version]
  31. Karkee, M.; Steward, B.L.; Tang, L.; Aziz, S.A. Quantifying sub-pixel signature of paddy rice field using an artificial neural network. Comput. Electron. Agric. 2009, 65, 65–76. [Google Scholar] [CrossRef]
  32. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  33. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  34. Hu, W.; Huang, Y.Y.; Wei, L.; Zhang, F.; Li, H.C. Deep convolutional neural networks for hyperspectral image classification. J. Sens. 2015, 2015, 1–12. [Google Scholar] [CrossRef] [Green Version]
  35. Yue, J.; Zhao, W.Z.; Mao, S.J.; Liu, H. Spectral-spatial classification of hyperspectral images using deep convolutional neural networks. Remote Sens. Lett. 2015, 6, 468–477. [Google Scholar] [CrossRef]
  36. Chen, Y.; Fan, R.S.; Yang, X.C.; Wang, J.X.; Latif, A. Extraction of urban water bodies from high-resolution remote-sensing imagery using deep learning. Water 2018, 10, 585. [Google Scholar] [CrossRef] [Green Version]
  37. Zhong, L.H.; Hu, L.N.; Zhou, H. Deep learning based multi-temporal crop classification. Remote Sens. Environ. 2019, 221, 430–443. [Google Scholar] [CrossRef]
  38. Vedrenne, M.; Borge, R.; Lumbreras, J.; Rodriguez, M.E.; de la Paz, D.; Perez, J.; de Andres, J.M.; Quaassdorff, C. A comprehensive approach for the evaluation and comparison of emission inventories in Madrid. Atmos. Environ. 2016, 145, 29–44. [Google Scholar] [CrossRef]
  39. Vakalopoulou, M.; Karantzalos, K.; Komodakis, N.; Paragios, N. Building detection in very high resolution multispectral data with deep learning features. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 1873–1876. [Google Scholar]
  40. Zhao, W.; Du, S.; Wang, Q.; Emery, W.J. Contextually guided very-high-resolution imagery classification with semantic segments. ISPRS J. Photogramm. Remote Sens. 2017, 132, 48–60. [Google Scholar] [CrossRef]
  41. Zhang, T.; Tang, H. A comprehensive evaluation of approaches for built-up area extraction from landsat oli images using massive samples. Remote Sens. 2019, 11, 2. [Google Scholar] [CrossRef] [Green Version]
  42. Zhang, L.; Zhang, L.; Du, B. Deep learning for remote sensing data: A technical tutorial on the state of the art. IEEE Geosci. Remote Sen. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
  43. Ball, J.E.; Anderson, D.T.; Chan, C.S. Comprehensive survey of deep learning in remote sensing: Theories, tools, and challenges for the community. J. Appl. Remote Sens. 2017, 11, 042609. [Google Scholar] [CrossRef] [Green Version]
  44. Kussul, N.; Lavreniuk, M.; Skakun, S.; Shelestov, A. Deep learning classification of land cover and crop types using remote sensing data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 778–782. [Google Scholar] [CrossRef]
  45. Qi, J.; Chehbouni, A.; Huete, A.R.; Kerr, Y.H.; Sorooshian, S. A modified soil adjusted vegetation index. Remote Sens. Environ. 1994, 48, 119–126. [Google Scholar] [CrossRef]
  46. Liu, W.; Henneberry, S.R.; Ni, J.; Radmehr, R.; Wei, C. Socio-cultural roots of rural settlement dispersion in Sichuan Basin: The perspective of Chinese lineage. Land Use Policy 2019, 88, 104162. [Google Scholar] [CrossRef]
  47. Yu, G.; Yu, Q.; Hu, L.; Zhang, S.; Fu, T.; Zhou, X.; He, X.; Liu, Y.A.; Wang, S.; Jia, H. Ecosystem health assessment based on analysis of a land use database. Appl. Geogr. 2013, 44, 154–164. [Google Scholar] [CrossRef]
  48. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  49. Guidici, D.; Clark, M.L. One-Dimensional convolutional neural network land-cover classification of multi-seasonal hyperspectral imagery in the San Francisco Bay Area, California. Remote Sens. 2017, 9, 629. [Google Scholar] [CrossRef] [Green Version]
  50. Wei, J.; Huang, W.; Li, Z.; Xue, W.; Peng, Y.; Sun, L.; Cribb, M. Estimating 1-km-resolution PM2. 5 concentrations across China using the space-time random forest approach. Remote Sens. Environ. 2019, 231. [Google Scholar] [CrossRef]
  51. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  52. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 22–27 August 2016; pp. 785–794. [Google Scholar]
  53. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  54. Benediktsson, J.A.; Swain, P.H.; Ersoy, O.K. Neural network approaches versus statistical methods in classification of multisource remote sensing data. IEEE Trans. Geosci. Remote Sens. 1990, 28, 540–552. [Google Scholar] [CrossRef]
  55. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Kestur, R.; Farooq, S.; Abdal, R.; Mehraj, E.; Narasipura, O.S.; Mudigere, M. UFCN: A fully convolutional neural network for road extraction in RGB imagery acquired by remote sensing from an unmanned aerial vehicle. J. Appl. Remote Sens. 2018, 12, 016020. [Google Scholar] [CrossRef]
  57. Rodriguez-Galiano, V.F.; Ghimire, B.; Rogan, J.; Chica-Olmo, M.; Rigol-Sanchez, J.P. An assessment of the effectiveness of a random forest classifier for land-cover classification. ISPRS J. Photogramm. Remote Sens. 2012, 67, 93–104. [Google Scholar] [CrossRef]
  58. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  59. Abdi, A.M. Land cover and land use classification performance of machine learning algorithms in a boreal landscape using Sentinel-2 data. GISci. Remote Sens. 2020, 57, 1–20. [Google Scholar] [CrossRef] [Green Version]
  60. Khatami, R.; Mountrakis, G.; Stehman, S.V. A meta-analysis of remote sensing research on supervised pixel-based land-cover image classification processes: General guidelines for practitioners and future research. Remote Sens. Environ. 2016, 177, 89–100. [Google Scholar] [CrossRef] [Green Version]
  61. Nguyen, A.; Yosinski, J.; Clune, J. Deep neural networks are easily fooled: High confidence predictions for unrecognizable images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 427–436. [Google Scholar]
  62. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer (MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  63. Ji, S.; Wei, S.; Lu, M. A scale robust convolutional neural network for automatic building extraction from aerial and satellite imagery. Int. J. Remote Sens. 2019, 40, 3308–3322. [Google Scholar] [CrossRef]
  64. Ling, F.; Boyd, D.; Ge, Y.; Foody, G.M.; Li, X.; Wang, L.; Zhang, Y.; Shi, L.; Shang, C.; Li, X. Measuring river wetted width from remotely sensed imagery at the subpixel scale with a deep convolutional neural network. Water Resour. Res. 2019, 55, 5631–5649. [Google Scholar] [CrossRef]
  65. Fang, H.; Lafarge, F. Pyramid scene parsing network in 3D: Improving semantic segmentation of point clouds with multi-scale contextual information. ISPRS J. Photogramm. Remote Sens. 2019, 154, 246–258. [Google Scholar] [CrossRef] [Green Version]
  66. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–21 June 2019; pp. 5693–5703. [Google Scholar]
  67. Wei, J.; Li, Z.; Guo, J.; Sun, L.; Huang, W.; Xue, W.; Fan, T.; Cribb, M. Satellite-derived 1-km-resolution PM1 concentrations from 2014 to 2018 across China. Environ. Sci. Technol. 2019, 53, 13265–13274. [Google Scholar] [CrossRef] [PubMed]
  68. Wei, J.; Li, Z.; Cribb, M.; Huang, W.; Xue, W.; Sun, L.; Guo, J.; Peng, Y.; Li, J.; Lyapustin, A.; et al. Improved 1 km resolution PM2.5 estimates across China using enhanced space-time extremely randomized trees. Atmos. Chem. Phys. 2020, 20, 3273–3289. [Google Scholar] [CrossRef] [Green Version]
  69. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? In Proceedings of the International Conference on Neural Information Processing Systems (ICONIP), Kuching, Malaysia, 3–6 November 2014; pp. 3320–3328. [Google Scholar]
Figure 1. Maps of location, training samples and validation samples of Banan District (a) and Zhongxian County (b). “Non-rice val “and “Rice val” represent validation samples and “Non-rice train” and “Rice train” denote training samples (more detailed information please refer to supplementary (i.e., Figures S3 and S4)).
Figure 1. Maps of location, training samples and validation samples of Banan District (a) and Zhongxian County (b). “Non-rice val “and “Rice val” represent validation samples and “Non-rice train” and “Rice train” denote training samples (more detailed information please refer to supplementary (i.e., Figures S3 and S4)).
Remotesensing 12 01620 g001
Figure 2. Network structure of Conv1D (a), Conv2D (b), Conv3D_1 (c) and Conv3D_2 (d). The input layer, fully connected layer and output layer of the four classifiers remain unchanged. Convolution kernels and pooling kernels have different dimensions, which vary according to the convolution dimensions. The parameters of convolutional neural networks (CNNs) are referenced at “https://keras.io/”.
Figure 2. Network structure of Conv1D (a), Conv2D (b), Conv3D_1 (c) and Conv3D_2 (d). The input layer, fully connected layer and output layer of the four classifiers remain unchanged. Convolution kernels and pooling kernels have different dimensions, which vary according to the convolution dimensions. The parameters of convolutional neural networks (CNNs) are referenced at “https://keras.io/”.
Remotesensing 12 01620 g002
Figure 3. Paddy-rice map of the study area generated by Conv-2D in Banan District (a) and Zhongxian County (b) the two subregions with 5 km × 5 km were used to compare the detailed information created by the models).
Figure 3. Paddy-rice map of the study area generated by Conv-2D in Banan District (a) and Zhongxian County (b) the two subregions with 5 km × 5 km were used to compare the detailed information created by the models).
Remotesensing 12 01620 g003
Figure 4. Subregion classification results ((a) Subregion A in Banan District and (b) Subregion B in Zhongxian County) based on different methods, Land Use Map (cyan-blue pixels represent the referenced rice pixels) is land-use vector data in 2017 for comparison, RGB is Sentinel-2 true color composite image (green pixels represent the overlap between result maps and the reference map, and red pixels and yellow pixels represent under-classification pixels and over-classification pixels, respectively); transfer represents the transferable results of the relatively reasonable classifiers (refer to Section 4.2).
Figure 4. Subregion classification results ((a) Subregion A in Banan District and (b) Subregion B in Zhongxian County) based on different methods, Land Use Map (cyan-blue pixels represent the referenced rice pixels) is land-use vector data in 2017 for comparison, RGB is Sentinel-2 true color composite image (green pixels represent the overlap between result maps and the reference map, and red pixels and yellow pixels represent under-classification pixels and over-classification pixels, respectively); transfer represents the transferable results of the relatively reasonable classifiers (refer to Section 4.2).
Remotesensing 12 01620 g004aRemotesensing 12 01620 g004b
Table 1. The division of pure pixel samples in Banan District and Zhongxian County.
Table 1. The division of pure pixel samples in Banan District and Zhongxian County.
Banan DistrictZhongxian County
TrainingValidationTrainingValidation
Rice15593861600386
Non-rice4800118948001180
Total6359157564001566
Table 2. Parameters of non-CNN classifiers.
Table 2. Parameters of non-CNN classifiers.
AlgorithmAbbreviationHyperparameterCandidate ValuesNumber of CombinationsOptimal Parameter Value
Random ForestRFn_estimators{(100,210),10}70180
max_features2,3,4,5,6,7,87
Support Vector MachineSVMC10−2,10−1, 1.0, 10, 102, 1037510
gamma0.01,0.1, 1, 2, 8,1010
Extreme Gradient BoostingXGBoostreg_alpha0.001, 0.01, 0.1360.1
colsample_bytree0.5, 0.6, 0.7, 0.8, 0.90.5
subsample0.5, 0.6, 0.7, 0.8, 0.90.7
Multilayer PerceptionMLPlearning_rate_init10−5, 10−4, 10−3, 10−2, 10−1350.001
batch_size{(20, 120), 20}80
Table 3. Confusion matrix.
Table 3. Confusion matrix.
Prediction
RiceNon-RiceSum
Ground TruthRiceTrue Positive (TP)False Negative (FN)Actual Positive (TP + FN)
Non-RiceFalse Positive (FP)True Negative (TN)Actual Negative (FN + TP)
SumPredicted Positive (TP + FP)Predicted Negative (FN + TN)TN+ TP+ FN+ FP
Table 4. Classifiers local accuracy evaluation in Banan District (a) and Zhongxian County (b); Boxes with black and grey colors represent the maximum and minimum values of the evaluation indicators of CNNs or non-CNNs, respectively).
Table 4. Classifiers local accuracy evaluation in Banan District (a) and Zhongxian County (b); Boxes with black and grey colors represent the maximum and minimum values of the evaluation indicators of CNNs or non-CNNs, respectively).
(a) Banan District Local Classification Results
Classifier TypeRecallPrecisionOverall AccuracyF1 ScoreTraining Time (s)Testing Time (s)
CNNsConv-1D0.72800.78270.88380.754457680
Conv-2D0.82640.88610.93140.85521271030
Conv-3D_10.76680.87090.91170.8067121014,948
Conv-3D_20.75910.89060.91770.81964434378
Mean0.77010.85760.91120.8090
Non-CNNsRF0.68650.86320.89600.76480.7662759
XGBoost0.71500.84400.89730.77420.4933113
SVM0.66580.86240.89160.75151.30561963
MLP0.44820.94020.85710.6070203.48380
Mean0.62680.87750.88550.7244
Overall mean0.69950.86750.89840.7667
(b) Zhongxian County local classification results
Classifier TypeRecallPrecisionOverall AccuracyF1 scoreTraining Time (s)Testing Time (s)
CNNsConv-1D0.95850.5929 0.8340 0.7451127755
Conv-2D0.79530.88990.92530.83992302320
Conv-3D_10.82380.82810.91440.82605517878
Conv-3D_20.84720.81750.91570.83213333626
Mean0.85620.78210.89730.8108
Non-CNNsRF0.69170.80420.88250.74370.7285759
XGBoost0.7228 0.78150.88190.75100.5123113
SVM0.7098 0.80350.88570.75381.30561963
MLP0.75390.72930.87040.7414156.4380
Mean0.71960.77960.88010.7475
Overall mean0.78790.78090.88870.7791
Table 5. Classifiers transfer accuracy evaluation in Experiment A and Experiment B.
Table 5. Classifiers transfer accuracy evaluation in Experiment A and Experiment B.
(a) Experiment A: Classifiers that Trained in Banan District Transferred to Zhongxian County
Classifier TypeRecallPrecisionOverall AccuracyF1 Score
CNNsConv-1D0.98450.40640.64180.5753
Conv-2D0.96370.48560.73950.6458
Conv-3D_10.99740.26630.32180.4203
Conv-3D_20.98700.39400.62260.5632
Non-CNNsRF0.87560.53910.78480.6673
XGBoost0.81090.64270.84230.7171
SVM0.0000-0.7535-
MLP0.99740.28600.38570.4446
(b) Experiment B: Classifiers that Trained in Zhongxian County Transferred to Banan District
Classifier TypeRecallPrecisionOverall AccuracyF1 Score
CNNsConv-1D0.22280.65650.78000.3327
Conv-2D0.97410.26840.33990.4208
Conv-3D_11.00000.25390.27680.4050
Conv-3D_20.96890.29800.43050.4558
Non-CNNsRF0.50780.89090.86350.6469
XGBoost0.54660.86830.86800.6709
SVM0.0000-0.7538-
MLP0.10881.00000.78060.1963

Share and Cite

MDPI and ACS Style

Zhang, W.; Liu, H.; Wu, W.; Zhan, L.; Wei, J. Mapping Rice Paddy Based on Machine Learning with Sentinel-2 Multi-Temporal Data: Model Comparison and Transferability. Remote Sens. 2020, 12, 1620. https://doi.org/10.3390/rs12101620

AMA Style

Zhang W, Liu H, Wu W, Zhan L, Wei J. Mapping Rice Paddy Based on Machine Learning with Sentinel-2 Multi-Temporal Data: Model Comparison and Transferability. Remote Sensing. 2020; 12(10):1620. https://doi.org/10.3390/rs12101620

Chicago/Turabian Style

Zhang, Weichun, Hongbin Liu, Wei Wu, Linqing Zhan, and Jing Wei. 2020. "Mapping Rice Paddy Based on Machine Learning with Sentinel-2 Multi-Temporal Data: Model Comparison and Transferability" Remote Sensing 12, no. 10: 1620. https://doi.org/10.3390/rs12101620

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop