Open Access
2 May 2022 Mapping large area tea plantations using progressive random forest and Google Earth Engine
Le’an Qu, Manchun Li, Zhenjie Chen, Wangbing Liu, Junjun Zhi, Lechun Zhang
Author Affiliations +
Abstract

A timely and accurate understanding of the spatial distribution of tea plantations is beneficial for agricultural management and regional sustainable development. However, obtaining detailed distribution data on large-area tea plantations remains challenging owing to limitations in computational capabilities, training data, and workflow design. Utilizing the Google Earth Engine, which provides a catalog of multisource data in a cloud-based environment, we developed a methodology to generate a highly accurate tea plantation map, with a 10-m resolution, for Anhui Province, China, by integrating a random forest model with a progressive model. Our major contribution lies in this hybrid approach, which comprises two major components: (1) an optimal classification band combination derived from Sentinel-2 products and the digital elevation model filtered by the J-M distance model and (2) a progressive random forest method introduced for tea plantation classification. The experimental results show that our proposed workflow achieved an average classification accuracy of 89.27% for the entire Anhui Province. In addition, this approach is semiautomatic and can effectively reduce the labor required during the generation of training data compared with traditional classification approaches. These findings demonstrate the potential of integrating machine learning and progressive models to produce high-precision remote sensing classification maps.

1.

Introduction

Tea, which originated in China, has become one of the three most popular beverages worldwide.1 In China, the global leader in tea production, tea is primarily grown in tropical and subtropical regions, such as Anhui Province,2 where tea cultivation plays a vital role in the agricultural economy and rural development.3 According to the Anhui Provincial Bureau of Statistics, the area of tea plantations in Anhui in 2019 was 1870.58  km2, an increase of 57.38% from 1990.4 Although the expansion of tea plantations has promoted local economic development, it has also caused a series of environmental problems, such as decreased soil fertility and soil erosion.5 Therefore, a timely and accurate understanding of the spatial distribution of tea plantations is conducive to governmental environmental protection and agricultural management.6

Information on the distribution of tea plantations can generally be obtained through manual reporting and remote sensing image classification.7 Utilizing remote sensing data to identify tea plantations saves time and labor costs; hence, it is more popular than manual reporting.8 The biggest difficulty in using remote sensing data is that tea trees are perennial evergreen woody plants, and their spectral characteristics are easily confused with similar woody plants, such as natural forests.9 Thus, it is difficult to achieve accurate tea plantation classifications using only spectral bands.1 Hence, previous studies primarily used hyperspectral and high-resolution images, fusions of multiple types of remote sensing data, and classifications based on phenological features to identify tea plantations.1,2,911 The classification accuracies using the former three methods are very high, but often the data are expensive and the study area is small, which makes it difficult to identify tea plantations over large areas.12 In this situation, phenological feature-based classification using Sentinel-2 data is among best choices for identifying the distribution of tea plantations over large areas.13 This is because Sentinel-2 data have three distinct advantages over other satellite images: worldwide coverage, high temporal and spatial resolution, and a variety of red-edge bands that are sensitive to vegetation.14

Nevertheless, there are some challenges in applying this method to identifying tea plantations over a large area.15 First, determining the optimal combination of bands for classification is difficult.12 Too few classification bands may lead to low classification accuracy, whereas too many bands may lead to low classification efficiency, over-fitting, and local optima.16 Second, the classification of tea plantations over a large area requires many accurate training samples.17 If the number of training samples is small, the classification accuracy is usually low and has a large uncertainty.18 However, obtaining multiple training samples is costly and time-consuming.19 Therefore, it is essential to balance the number of training samples with the required labor. Finally, for large-area time-series classifications, a high storage capacity is necessary to store the remote sensing data.20 Furthermore, strong computing power is required to achieve high classification efficiency.21

Web-based remote-sensing cloud platforms such as Google Earth Engine (GEE) can greatly reduce computing time and enable the classification of large areas using remote sensing time series data.22 GEE contains a variety of remote sensing images from different sensors, which reduces personal data storage requirements.23 In addition, GEE provides high computing power that greatly improves classification efficiency, and it can effectively perform repeated comparison experiments to determine the optimal classification bands.24 Moreover, GEE provides a visual interactive platform that can dynamically add training samples, providing a balance between training sample acquisition and labor.25 In this study, we used the Jeffries–Matusita distance (J-M distance) model to select the optimal classification features and utilized the progressive random forest method to classify tea plantations. These processes were conducted on the GEE platform, allowing our workflow to achieve high classification accuracy for tea plantations over a large study area with minimal time and labor costs.

The goals of this study were to classify Sentinel-2 time series data to obtain a high-precision tea plantation distribution map for Anhui Province during 2020 and determine which classification features and methods effectively perform high-precision tea plantation classification over a large area with complex terrain. Specifically, we addressed the following two questions: (1) how can we select the optimal combination of classification features in GEE, so it can classify tea plantations with high classification accuracy and efficiency? (2) How can a suitable classification method be constructed in GEE to ensure that tea plantations can be identified with high accuracy over a large area?

2.

Study Area and Datasets

2.1.

Study Area

Anhui Province is in eastern China, between 114°54E and 119°37E and 29°41N and 34°38N. It spans 450 km from east to west and 570 km from north to south, with a total area of 140,100  km2, accounting for 1.45% of the total land area of China.26,27 Anhui is in the middle to lower reaches of the Yangtze River and the Huaihe River. It borders Jiangsu to the east, Zhejiang and Jiangxi to the south, Henan and Hubei to the west, and Shandong to the north.28 As of July 2020, there were 16 cities in the administrative regions in Anhui Province: Hefei, Wuhu, Bengbu, Huainan, Ma’anshan, Huaibei, Tongling, Anqing, Huangshan, Fuyang, Suzhou, Chuzhou, Lu’an, Xuancheng, Chizhou, and Bozhou (Fig. 1).

Fig. 1

Location of the study area. Maps showing the (a) location, (b) elevation, (c) slope, (d) annual average temperature, and (e) annual average precipitation in the study area.

JARS_16_2_024509_f001.png

In general, the natural conditions of northern and southern Anhui differ greatly.29 The terrain is generally high in the west and south and low in the east and north. The north is dominated by plains, whereas the south is dominated by hills and mountains. Anhui is in the mid-latitude zone, which belongs to the transitional region between the warm temperate and subtropical zones. The annual average temperature in Anhui is 14°C to 17°C, and the annual average precipitation is 750 to 1750 mm.26 Precipitation is higher in the south and the mountainous areas and is lesser in the north, the plains, and the hills. Rainfall is abundant in summer, accounting for 40% to 60% of the annual precipitation.27 The unique climatic and topographical conditions have made Anhui Province one of the major tea-producing provinces in China.6 In 2019, the tea plantation area reached 187,058 hectares, and the total output of tea was 121,980 tons.4

2.2.

Basic Data and Data Preprocessing

The data used in this study include four types of raster data, specifically, Sentinel-2 MSI data, land use and land cover data, topographic data, and Google Earth CNES/Airbus imageries, and two types of vector data, specifically, administrative boundary data and field survey data. All data were open-access datasets, except for the field survey data. See Table 1 for detailed descriptions of the data used in this study.

Table 1

Data included in this study and their usage.

Data typeData nameSpatial resolutionUsage
RasterSentinel-2 MSIa10 mIdentify tea gardens and other land use types
Google Earth CNES/Airbus imageriesa0.3 mGenerate sample points through visual interpretation
GlobeLand3030 mCross-validate land use and land cover types except for tea plantations through visual interpretation
SRTM V3 DEMa30 mAssist remote sensing image classification
VectorAdministrative boundary1:10,000Determine the extent of the study area
Field survey dataGenerate training and verification sample points
Note: SRTM: Shuttle Radar Topography Mission.

aIt can be called on Google Earth Engine directly.

The Sentinel-2 data were obtained by a satellite cluster composed of two identical satellites, Sentinel-2 A and Sentinel-2 B.30 The revisit periods of the two satellites were 5 days and were simultaneous.31 In this study, Sentinel-2 satellite data were selected as the main remote sensing image data for accurately identifying the distribution of tea plantations. Sentinel-2 satellite data products include three levels: level-0, level-1C, and level-2A. Among them, level-2A images are bottom-of-atmosphere reflectance in cartographic geometry after atmospheric correction and, thus, are ready to use.32 In this study, we used level-2A products; only the identification of clouds and cloud shadows was required prior to use. We used the Fmask 4.0 algorithm by Qiu et al. for cloud detection and achieved good results.33

Topographic features (elevation, slope, and aspect) are important factors that affect tea tree growth; therefore, they can effectively improve the classification accuracy for tea plantations.11 To describe the topographic features, we used a digital elevation model (DEM) with a 30-m resolution generated by the Shuttle Radar Topography Mission (SRTM).34 This DEM is a postprocessed elevation dataset that is widely used because of its high accuracy and extensive coverage.35

High-precision reference data and an appropriate classification system are prerequisites for classifications. In this study, we combined imagery observations and field investigations to determine the reference points. We collected field survey data acquired with a handheld GPS/GNSS receiver in point-positioning mode in Huangshan and Wuhu City in 2020. The imagery observation data for other cities were incrementally added via Google Earth CNES/Airbus during the progressive random forest classification process. According to Giuseppe Pulighe et al., an error of 1  m occurred in visualizing added points.36 Considering that our classification results have an accuracy of 10 m, these points can be used for classification and accuracy verification.

GlobeLand30 was used for qualitative cross-validation with Sentinel-2 MSI classification results through visual interpretation. Then, according to the distribution of the ground objects in the study area, they were reclassified into six LULC (land use/land cover) types: cropland, forest, grassland, water bodies, built-up land, and unused.37

After the data were collected, the inputs were resampled to a spatial resolution of 10 m, which corresponds to the fine resolution of the Sentinel-2 MSI. Finally, all data were projected into the WGS84 coordinate system (EPSG: 4326) for tea plantation classification.

3.

Methods

Figure 2 shows the overall workflow for the classification of tea plantations in the study area in 2020. First, to enhance the separability of the tea plantations from the other LULC types, we used original band features, remote sensing index features, texture features, tasseled cap transformation (TCT) features, and terrain features to generate overall classification features. Then, based on maintaining the separability between the tea plantations and other LULC types, the J-M distance model was applied to obtain the optimal classification features. Finally, we used a progressive random forest classification method for high-precision classification.

Fig. 2

Overall workflow chart for the classification of tea plantations.

JARS_16_2_024509_f002.png

3.1.

Selection and Optimization of Classification Features

3.1.1.

Spectral features of the original bands

The spectral features of the original bands of the remote sensing images record the reflection intensity of different ground objects.38 By analyzing the different spectral characteristics of these objects, it is possible to distinguish vegetation, water bodies, and other types of ground objects.39 The 60-m bands in the Sentinel-2 MSI data are mainly used for atmospheric correction; therefore, they were excluded from this study. The other 10 original bands in the Sentinel-2 MSI data were selected for classification, specifically, blue, green, red, red edge1-4, NIR, and SWIR1-2. To analyze the phenological characteristics of the vegetation, we calculated the maximum, minimum, median, and standard deviation of the digital number of all original bands in the study area in 2020. Tea trees and natural forests show strong spectral similarity; therefore, in addition to the original spectral features, other classification features, including remote sensing indices, red-edge vegetation indices, TCT features, and topographic features, were utilized in this study.

3.1.2.

Features of remote sensing indices

Due to the surface environment, the same objects may have different spectral features, and different objects may have the same spectral features, which directly impacts the classification accuracy.40 To overcome this issue, this study selected five commonly used remote sensing indices to improve classification accuracy. These indices were mainly used to enhance the distinction among other LULC types, rather than tea plantations. Nevertheless, because of the strong spectral similarity between tea trees and forests, eight red-edge vegetation indices were used to further distinguish tea plantations from forests.

The five common remote sensing indices include the normalized difference vegetation index (NDVI), soil adjusted vegetation index (SAVI), normalized difference water index (NDWI), modified normalized difference water index (MNDWI), and normalized difference built-up index (NDBI).4145 The eight red-edge vegetation indices include the normalized difference vegetation index Red-edge1-3 (NDVIre1, NDVIre2, and NDVIre3), normalized difference Red-edge1,2 (NDre1 and NDre2), inverted red-edge chlorophyll index (IRECI), MERIS terrestrial chlorophyll index (MTCI), and red-edge chlorophyll index (CIre).42,46 The equations to calculate these indices are listed in Table 2.

Table 2

Remote sensing indices used in this study and their equations for calculation.

NameCalculation equationNumber
NDVINDVI = (B8 – B4)/(B8 + B4)(1)
SAVISAVI = (B8 – B4) × (1+0.5)/(B8 + B4 + 0.5)(2)
NDWINDWI = (B8 – B11)/(B8 + B11)(3)
MNDWIMNDWI = (B3 – B11)/(B3 + B11)(4)
NDBINDBI = (B11 – B8)/(B11 + B8)(5)
NDVIre1NDVIre1=(B8AB5)/(B8A+B5)(6)
NDVIre2NDVIre2=(B8AB6)/(B8A+B6)(7)
NDVIre3NDVIre3=(B8AB7)/(B8A+B7)(8)
NDre1NDre1=(B6B5)/(B6+B5)(9)
NDre2NDre2=(B7B5)/(B7+B5)(10)
IRECIIRECI = (B7 – B4)/(B5/B6)(11)
MTCIMTCI = (B6 – B5)/(B5 - B4)(12)
CIreCIre=B7/B51(13)
Note: NDVI, SAVI, NDWI, MNDWI, and NDBI represent the normalized difference vegetation index, soil adjusted vegetation index, normalized difference water index, modified normalized difference water index, and normalized difference built-up index, respectively. NDVIre1-3, NDre1,2, IRECI, MTCI, and CIre represent the normalized difference vegetation index Red-edge1-3, normalized difference Red-edge1,2, inverted red-edge chlorophyll index, MERIS terrestrial chlorophyll index, and red-edge chlorophyll index, respectively. B3, B4, B5, B6, B7, B8, B8A, B11, and B12 are blue, green, red, red-edge 1, red-edge 2, red-edge 3, NIR, red-edge 4, SWIR1, and SWIR2, respectively.

3.1.3.

Texture features

Different ground objects have different texture features; therefore, adding texture features helps enable the full capabilities of remote sensing image information.47,48 To facilitate tea harvests and obtain adequate nutrition (e.g., CO2 and soil nutrients), tea trees are planted at certain spatial intervals.1 Due to this unique cultivation method, tea plantations have obvious textural features that differ significantly from natural vegetation such as forests and grass.9 Currently, the most common method of texture analysis is the gray-level co-occurrence matrix (GLCM).49 Because the GLCM has demonstrated good adaptability and strong robustness, it was selected to extract texture information.50

The GLCM method requires an appropriate band when calculating texture features.49 Because the texture features calculated based on Sentinel-2 MSI images are the most effective for distinguishing tea plantations from forests, the selected band should be highly capable of detecting vegetation changes. Considering that the resolution for our classification results was 10 m, we chose the 10-m resolution NDVI to calculate the GLCM matrix. In addition, to manage the interannual variation in NDVI, the median value of the annual NDVI was utilized in calculating the GLCM matrix. This is calculated in GEE by applying the “glcmTexture()” function; however, three parameters must first be set: size, kernel, and average.51 Because the individual tea plantations in the study area were generally small and scattered, the size was set to 3, which means that each neighborhood was a 3×3 area. The remaining parameters used default values. We used 12 texture features with lower correlations, including angular second moment, contrast, correlation, inverse difference moment, entropy, sum average, sum entropy, difference entropy, information measure of correlation 1, information measure of correlation 2, cluster shade, and cluster prominence.

3.1.4.

Tasseled cap transformation features

TCT features refer to those obtained by TCT, which is a result of special principal component analysis.52 Unlike general principal component analysis, the conversion coefficient of TCT is a fixed transformation matrix.53 TCT transforms the original multispectral image into the actual physical meaning of brightness (comprehensive surface reflectance), greenness (integration of the degree of surface vegetation coverage), and wetness (comprehensive surface water conditions) three-dimensional feature space, which can fully reflect surface reflection, vegetation coverage, and moisture information.32 The TCT process can reduce the feature dimensions and enhance remote sensing image information. The specific transformation equation is as follows:

Eq. (1)

Y=CX+a,
where Y represents the image after TCT, C represents the coefficient matrix corresponding to the TCT, X represents the original multispectral image before transformation, and a is a constant added to avoid negative values. The coefficient matrix C is a fixed transformation matrix that depends on the specific sensor. The TCT matrix coefficients of Sentinel-2 images are shown in Table 3. A total of six bands are required for TCT, including blue, green, red, NIR, SWIR1, and SWIR2.

Table 3

TCT matrix coefficients for Sentinel-2 images.

TCT featuresBlueGreenRedNIRSWIR1SWIR2
Brightness0.08220.13600.26110.38950.38820.1366
Greenness−0.1128−0.1680−0.34800.3165−0.4578−0.4064
Wetness0.13630.28020.3072−0.0807−0.4064−0.5602

3.1.5.

Topographic features

Topographic features are widely used to distinguish ground objects.54 Tea trees have extremely strict environmental requirements for growth; thus, the terrain in a tea tree plantation should be neither too flat nor too steep.9 In flat areas, rainfall will accumulate at the bottom of the tea trees and cause the roots to rot. In steep areas, it is difficult to maintain essential soil nutrients and moisture for tea tree growth, and tea farmers have difficulty picking tea. Therefore, tea trees can only grow in less steep mountainous or hilly areas. Therefore, terrain features can effectively improve the recognition accuracy for tea plantations and other objects, owing to the different terrain requirements for each object.2 For example, the objects in plains are typically cropland, built-up land, and water bodies, whereas those in mountainous and hilly areas are mainly forests and grasslands with only a small percentage of cropland and built-up land.

Three topographic features were considered in this study: elevation, slope, and aspect. Elevation directly affects the temperature in mountainous areas, thereby indirectly affecting the distribution of vegetation.11 The slope expresses the degree of steepness of a surface, which affects not only the surface energy exchange rate but also human activities and ultimately the spatial distribution of ground objects.55 The aspect, which is the facing direction of the slope, indirectly affects the growth of vegetation and ultimately affects its spatial distribution.5 In GEE, these topographic features can be directly calculated through the “ee.Algorithms.Terrain()” function.

3.1.6.

Optimization of classification features

Considering a phenological perspective, the maximum, minimum, median, and standard deviation were calculated for the original spectral features, remote sensing indices, and red-edge vegetation indices, resulting in a total of 110 classification features. Notably, strong correlations can exist among these features, resulting in low classification accuracy and efficiency.56 Therefore, this does not mean that more features are better, and appropriate classification features must be selected.

The J-M distance method is relatively simple and intuitive for directly measuring the distinguishability of different classes; therefore, we utilize it for feature optimization.57 Essentially, the J-M distance uses the distance between samples of different classes to measure their separability, thereby determining distinguishable features based on the training samples.58 The larger the J-M distance is, the higher the separability between classes is, and the easier classification it is to them. The calculation method for J-M distance is shown as

Eq. (2)

J-M(ci,cj)=(p(x|ci)p(x|cj)2dx,
where x is a time series, ci and cj are the two classes to be compared, and p(x|ci) and p(x|cj) are the conditional probability density functions for the time series x. When only considering the separability between two classes, Eq. (27) can be simplified to

Eq. (3)

J-M=2(1eBij),

Eq. (4)

Bij=18(MiMj)T(Vi+Vj2)1(MiMj)+12ln[|Vi+Vj2|/|Vi||Vj|],
where Bij is the Bhattacharyya distance of classes i and j; Vi and Vj are the sample covariance matrices of i and j, respectively; and Mi and Mj are the mean vectors of the corresponding samples. The values of the J-M distance range from 0 to 2. The larger the J-M distance is, the stronger the separability between the two classes is, with 2 corresponding to the largest separation. If JMij1, there is no overlap between classes i and j; if JMij<1, overlap exists between classes i and j.

According to Bruzzone et al.,59 when performing multiclass cases, the most important classification features for classification accuracy are determined by the classes with the lowest discrimination. In our classification process, tea plantations and natural forests were frequently misclassified, whereas combinations of other LULC types were easy to classify. The distinguishability between tea plantations and natural forests directly impacted the overall accuracy (OA) of classification results; therefore, the J-M distance between tea plantations and natural forests was calculated to optimize the classification features.

3.2.

Progressive Random Forest Classification Approach

The progressive random forest classification method used in this study combines prior knowledge and human intervention regarding the use of training samples for machine learning. Progressive random forest classification is similar to classical supervised classification; the goal of both is to obtain the highest possible accuracy through human–computer interaction. The difference is that supervised classification improves classification accuracy by marking a large number of samples before classification, whereas progressive random forest classification uses a strategy based on classification accuracy. After a classification accuracy assessment, which sorts the probability of correct classification for each pixel from low to high, some of the pixels with probability values below a specific value (80% was used in this study) are manually added to the training sample. From the perspective of practical application, progressive random forest classification assumes that the lower accuracy samples are always located in areas that are difficult to discriminate, which can quickly and continuously promote the performance of the classifier. The overall performance of the progressive random forest classification model is controllable and does not fall into a local optimum.

The progressive random forest classification model reflects the human learning process, progressing from easy to difficult. The model is trained using a small number of training samples based on classification accuracy, and complex samples are gradually added to improve the classification accuracy. In our study, a typical tea plantation distribution area was first selected, and a small number of training samples were used for classification. Subsequently, the validation samples were used to calculate the OA and the confidence of each pixel. If the OA was high, the classification area was expanded. If the OA was low, low-confidence pixels were added to the training sample and reclassified. After reclassification, if the OA improved, the newly added training samples were retained; if the OA decreased, the newly added training samples were removed and reselected. The above steps were repeated, expanding the classification area by gradually increasing the number of training samples; thus, the classification accuracy was gradually improved until the entire study area achieved a satisfactory classification accuracy.

As this study aimed to classify tea plantations over a large area, it was preferable to add low-precision tea plantation sample points before other types when applying the progressive random forest method. This ensured that the classification result for each pixel point within a tea plantation has high confidence. When applying the random forest model in GEE, two parameters must be set: numberOfTrees and minLeafPopulation. After repeated verification, we set numberOfTrees to 100 and minLeafPopulation to 10.

3.3.

Postprocessing and Accuracy Assessment

There were a small number of blank values in the classification result, which directly impacted the classification accuracy. There are two main reasons that blank values are generated. One reason is that some pixels have classification features outside the normal range, resulting in empty classification results. For these pixels, we rechecked the values for all classification features, eliminated outliers, and then reclassified them. The other reason is that some pixels are recognized as clouds by the cloud recognition method. For these, a multivalue within the 3×3 neighborhood of these pixels was assigned.

Tea trees prefer shade; hence, many tea plantations are in the shadows of hills. Thus, some of these plantations were easily misclassified as natural forests; therefore, we also used the 3×3 neighboring pixels to eliminate the influence of the hill shadows.

After postprocessing, we used a 10-fold cross-validation to assess the classification accuracy. We calculated the classification error matrix based on the classification results, which were quantified using the OA. The OA is the sum of the diagonal elements divided by the sum of all of the elements of the confusion matrix. Then, we calculated the producer’s accuracy (PA), user’s accuracy (UA), and F1-score. The F1-score is particularly useful for class-level accuracy assessment because it assigns equal importance to the PA and UA. Finally, the above process was repeated 10 times, and the average OA, PA, UA, and F1-score were calculated. The PA, UA, and F1-score are calculated for each class as follows:

Eq. (5)

PAi=j=1rniinij,

Eq. (6)

UAi=j=1rniinji,

Eq. (7)

(F1)i=2×PAi×UAiPAi+UAi,
where r is the number of classes and nij is the element of the confusion matrix in the i’th column and j’th row, that is, the count of elements of class j classified as class i. PAi, UAi, and (F1)i stand for the PA, UA, and F1-score for class i, respectively.

4.

Results

4.1.

Optimized Classification Features

In this study, 110 classification features were optimized to ensure classification efficiency and achieve a higher classification accuracy. Both the topographic and TCT features had only three features and weak correlations; hence, they were excluded from the feature optimization process. Therefore, the rest of the classification features, including the original bands, remote sensing indices, and texture features, were optimized using the J-M distance. The feature optimization process was conducted in the following two steps. First, 1000 tea plantation samples and 1000 natural forest samples were randomly selected from the training samples. Then, the J-M distance for each feature was calculated (see Table 8 for details), and the features with J-M distance values lower than a certain standard were eliminated.

The J-M distance can determine the discrimination between different classification bands but not the number of optimal features for classification. Here, we selected the optimal feature bands based on Lorenzo Bruzzone et al.59 As shown in Table 8, the maximum, minimum, and median values of the J-M distance of the original bands were 0.483, 0.085, and 0.229, respectively. Thus, the tea plantations and natural forests were difficult to distinguish. The original bands with J-M distance values greater than 0.2 were selected. Similarly, the characteristics of tea plantations and natural forests were also very similar in the remote sensing indices. Those with J-M distances greater than 0.2 were also selected. However, the distinguishability between tea plantations and natural forests was very strong for the red-edge vegetation indices, which are more sensitive concerning vegetation. Therefore, the red-edge vegetation indices with J-M distances greater than 0.4 were selected. A brief description of the optimized classification features, including the number and J-M distance values, is shown in Table 4 (for details, see Table 9). The total number of classification features was reduced from 116 to 64, a 45% decrease, which dramatically improved the classification efficiency. The overall J-M distance values varied from 1.414 for the original classification features to 1.412 for the optimized classification features, which is a difference of 0.14%; therefore, a comparable result was obtained after the feature optimization process. From the perspective of balancing the classification efficiency and accuracy, the feature optimization process performed excellently.

Table 4

Optimization results for the classification features.

Before optimizationAfter optimization
NumberJ-MNumberJ-M
Original bands401.030270.870
Remote sensing indices200.983150.895
Red-edge indices321.39971.320
Texture features121.19891.183
TCT features30.52830.528
Topographic features30.37030.370
Total1161.414641.412

4.2.

Classification Results for the Tea Plantations and Other Land Cover Types

Figures 3 and 4 show the results of the tea plantation classification using the progressive random forest classification model from a regional and local perspective, respectively. As shown in Fig. 3, tea plantations were scattered throughout in Anhui Province. Tea plantations were mainly distributed in the southern and western study area, which predominantly have a mountainous terrain. In the middle of the study area, there were a small number of tea plantations in hilly terrain. There were few tea plantations in the northern study area or along the Yangtze River; these places are plains that are not suitable for the growth of tea trees. The city with the most tea plantations was Huangshan City, which was consistent with the statistical data.

Fig. 3

Map of the tea plantations in the study area in 2020.

JARS_16_2_024509_f003.png

Fig. 4

Comparison of the classification results in typical areas. (a), (c), (e) Sentinel-2 MSI true-color images of a hilly region, hill-to-mountain transition region, and mountain region, respectively, and (b), (d), and (f) the corresponding classification results.

JARS_16_2_024509_f004.png

Figure 4 shows the classification results for three typical regions. The tea plantations identified using the progressive classification model were generally consistent with reality. In hilly areas, tea plantations were well identified, and the cropland and forest classifications were accurate [Fig. 4(b)]. In the transitional regions between hills and mountains, the overall tea plantation classifications were also relatively accurate; however, a small amount of forest was misclassified as tea plantations [Fig. 4(d)]. The classification results for forests and water bodies were more accurate; however, some cropland was misclassified as built-up land. In mountainous areas, the classification results for tea plantations and forests were accurate; however, some cropland was again misclassified as built-up land [Fig. 4(f)]. In general, tea plantations were accurately classified in the main distribution areas.

4.3.

Classification Accuracy Analysis

To test the classification accuracy for tea plantations, the OA, UA, PA, and F1-score were calculated quantitatively using the confusion matrix. In the quantitative analysis, a 10-fold cross-validation was used to determine the stability of the classification model. The calculation results are presented in Table 5.

Table 5

Classification accuracy for results in the study area.

UA (%)PA (%)F1-score
Tea plantations82.04 ± 0.0386.33 ± 0.020.8413
Cropland88.67 ± 0.0287.67 ± 0.010.8817
Forest95.50 ± 0.0294.74 ± 0.010.9512
Grassland76.08 ± 0.0464.95 ± 0.100.7007
Water bodies90.15 ± 0.0191.97 ± 0.010.9105
Built-up land85.60 ± 0.0287.58 ± 0.010.8658
Unused land58.49 ± 0.0646.56 ± 0.110.5185
OA = 89.27 ± 0.03%, Kappa = 0.8831
Note: UA, PA, and OA represent the user’s accuracy, producer’s accuracy, and overall accuracy, respectively. The values of the UA, PA, and OA are the mean ± standard deviation.

The average OA reached 89%. The classification accuracy for cropland, forest, and water bodies was greater than 85%, and that for tea plantations reached 82%. With the spatial resolution and the large area of Anhui Province, this result is generally satisfactory. However, the classification accuracies for grassland and unused land were poor, which was mainly caused by their small distribution areas within the study area and the small number of sample points used during classification. Nevertheless, these low accuracies had little influence on the identification of tea plantations because of their small areas and low probability of misclassification as tea plantations.

5.

Discussion

5.1.

Impacts of Different Feature Selection Schemes

To compare the impacts of different classification features on the classification accuracy for tea plantations, this study designed 10 different classification feature combination schemes. Because the red-edge bands had the largest J-M distance, comparisons with and without these bands were carried out to analyze whether they have prominent effects when designing the feature combination method. Due to the large study area and the number of comparative experiments, only Huangshan City, which has the most tea plantations, was selected for comparative experiments. The experimental results are presented in Table 6.

Table 6

Classification accuracies of different classification feature combinations in Huangshan city.

UA (%)PA (%)OA (%)F1-score
Scene1 (original bands+ remote sensing indices)86.1096.0383.990.9079
Scene2 (original bands+ remote sensing indices+ red-edge indices)87.4596.185.390.9157
Scene1 + texture features87.995.3285.10.9146
Scene2 +texture features88.6195.1785.90.9177
Scene1 + TCT features87.4795.0884.580.9112
Scene2 + TCT features88.4995.0285.650.9164
Scene1 + topographic features86.8995.8184.630.9113
Scene2 + topographic features87.896.185.820.9176
All89.1295.2586.500.9208
Optimized features88.6795.3386.410.9188

Table 6 shows that the combination of the original bands and the remote sensing indices obtained a high OA value of 83.99% because the constructed classification feature space could extract the phenological features of each class. The red-edge indices had the most significant impacts on improving the classification accuracies for tea plantations due to their high sensibility for vegetation. The topographical features had the second largest positive effect on classification accuracy. Tea plantations have strict environmental requirements; therefore, elevation, slope, and aspect all have direct effects on the growth of tea trees. The TCT features also showed positive effects because they comprehensively reflect the brightness, greenness, and wetness of ground objects. For example, the water in the roots of tea plantations and natural forests differs, resulting in a difference in TCT features. Tea trees are planted at certain spatial intervals to facilitate both growth and harvesting; thus, their greenness is generally lower than that of natural forests. The smallest improvement in the accuracy was from the texture features. This was mainly because the spatial resolution of the Sentinel-2 data was 10 m, which does not fully reflect the textural features of the tea plantations, resulting in a minor improvement.

When using all features, the UA for tea plantations was 89.12%, the PA was 95.25%, and the F1-score was 0.9208. In addition, these accuracy indices obtained using the optimized features were 88.67%, 95.33%, and 0.9188, respectively. Although the UA is slightly decreased when using the optimized features, the PA improved, and the F1-score, which measures the comprehensive classification accuracy, dropped by 0.002. The OA using all features was 86.50%, whereas that using the optimized features was 86.41%, which is a decrease of 0.09%. Although the use of optimized features resulted in a slight loss of accuracy, the number of classified features was dramatically reduced from 116 to 64. A reduction in the number of features can significantly improve the classification efficiency. When using the progressive random forest model, it is necessary to continuously add training samples and then reclassify them. Therefore, compared with the slight loss of accuracy, the improvement in the classification efficiency is more important.

5.2.

Improvement in Classification Accuracy by Progressive Random Forest

In this study, we designed a progressive random forest model that combines random forest and progressive models. The progressive random forest model improves classification accuracy by gradually increasing the number of training samples on GEE. These newly added training samples were based on the probable classification result for each pixel. A total of 10 iterative classifications were performed, and the results are shown in Fig. 5.

Fig. 5

The accuracy of the progressive random forest classification with the number of samples used during the classification process.

JARS_16_2_024509_f005.png

Figure 5 shows that, in the process of iterative classification, the OA, UA, and PA of the tea plantations constantly improved. This shows that the proposed progressive random forest model was effective. For the first iteration, the samples were derived from the results of our field surveys in Huangshan and Wuhu. Huangshan and Wuhu are typical mountainous and hilly tea-producing cities in the study area, respectively. For the first iteration, the samples were concentrated in mountainous and hilly areas, resulting in low OA, UA, and PA for the tea plantations. Then, the visually selected training samples for other cities were gradually added according to the accuracies of the classification results. With additional training samples, the OA, UA, and PA greatly improved. On the 10th iteration, the OA and PA exceeded 85%, which met our objectives; hence, we stopped the classification process.

The first visual samples were from the areas surrounding Huangshan and Wuhu; subsequently, the area was gradually expanded. The results of the visually selected additions and land survey samples are shown in Fig. 6. The figure shows that there were more samples in the southern study area but few samples in the northern region. This is because tea plantations and natural forests were initially misclassified; thus, we added samples of tea plantations and natural forests. With the increase in natural forest samples, tea plantations were mainly misclassified as croplands. Therefore, we added samples of cropland in the middle of the study area. When the UA and PA of the tea plantations exceeded 80%, we added samples from other classes to improve the OA.

Fig. 6

The results of the field survey and the addition of visually selected training samples.

JARS_16_2_024509_f006.png

5.3.

Advantages and Disadvantages

Because the spectral bands of tea plantations and natural forests are highly similar, obtaining high-precision classification of tea plantations over a large area remains challenging. With ready-made remote-sensing cloud platforms (e.g., GEE) and the use of nonparametric classifiers, the main challenge becomes selecting suitable classification bands and training samples.

The high classification accuracy and F1-score for tea plantations and other main land use classes demonstrate the practicality of the progressive random forest method. The effectiveness of this method depends on accessing sufficient computing power. In this study, this was not a problem because the GEE provided sufficient computing power.

In this study, a new method for the classification of sentinel-2 MSI data, which was combined with a random forest and progressive model running on GEE, was proposed. This is advantageous for the following reasons: (1) it retains the advantages of the random forest and the progressive model, which can effectively and robustly run high-dimensional data for a large area. (2) It can classify sparse objects (such as the tea plantations in this study) with a high classification accuracy. (3) The entire classification process is carried out on GEE to ensure highly efficient overall workflow.

We used classification and regression trees (CART), support vector machine (SVM), and deep neural networks (DNN) to verify the generality and reliability of our conclusions. In the comparative experiments, the samples used for the CART, SVM, and DNN models were the same as for the RF models. Default parameters were set for CART and SVM in GEE. The Keras library was utilized for the DNN hyperparameter optimization in Google Colab, and the key parameters of DNN were set as follows: 64–128–256 for the hidden layers, 10 for the epochs, ReLU (the rectified linear units) for the activation function, Adam for the optimizer, and 20% for the dropout ratio. The OA values of the comparative experiment classifications are shown in Table 7. In general, when a small number of samples was used, a low OA value was obtained. In addition, the OA values of the compared classification models increased with the increase in the number of samples, indicating that increasing the number of samples could significantly improve classification accuracy, which was similar to the RF model. However, the contributions of increasing samples to the OA were found to be different among the compared classification models. The OA values of the CART model were moderately improved with the increase of samples, but the highest OA of the CART model (79.52%) was significantly lower than that of the RF model (89.27%). The SVM model only needs a small number of samples to obtain high classification accuracy, but it easily overfits and its highest OA (83.89%) was also lower than that of the RF model. Increasing the number of samples can dramatically improve the classification accuracy of the DNN model, which needs a large number of samples to obtain high classification accuracy. Although the highest OA of DNN (75.96%) was significantly lower than that of the RF model, it is worth noting that the classification accuracy value of the DNN model continuously improved as the samples increased, which means that it may be close to or exceed the highest classification accuracy of the RF model if the number of samples is more than 6193. For balancing the classification accuracy and the limit number of samples, the combination of the progressive model and the RF model is the optimal choice.

Table 7

Overall accuracies of different comparative experiment models in Huangshan city.

Number of samplesCART (%)SVM (%)DNN (%)
112844.0565.3710.19
209850.1970.8022.36
298055.0576.5630.50
385659.3979.6039.23
456362.2483.0846.77
514367.5283.8954.38
553272.7683.0862.56
586477.2883.1869.77
602779.5283.0873.83
619376.9083.6075.96

Although our progressive random forest model has many advantages, it also has some disadvantages. First, because the GEE visualization interactive platform only provides current remote sensing images, the training data added by manual visual inspection only utilizes the current data, and the generation of sample points using historical data requires other platforms. Second, progressive random forest is semiautomated. Manual assignment is required each time that new samples are added to the training sample dataset. In future studies, we will consider introducing models (such as LandTrender) to generate long-term sample data. We are also considering adding other existing land use/land cover products to reduce manual assignment and improve the automation of the classification workflow.

6.

Conclusions

A methodology was proposed for the classification of tea plantations in Anhui Province in 2020. The J-M model distance was applied to the original bands, remote sensing indices, texture features, TCT features, and terrain features to select the optimal combination of classification features. A progressive random forest model was used to classify 3554 scenes from Sentinel-2 MSI images. Using these methods, we obtained a map of tea plantations in Anhui Province in 2020 and analyzed its classification accuracy.

The progressive random forest classification method used in this study combines the advantages of the random forest and progressive models. First, remote sensing classification for a large area was carried out quickly by utilizing random forest. Moreover, owing to the advantages of the progressive model, the OA of classification remained high even with a small number of training samples. Furthermore, the progressive random forest classification model accurately identified the spatial distribution of tea plantations.

By comparing the J-M distance, the number of classification features was reduced dramatically, which improves the classification efficiency. Moreover, high classification accuracy was obtained using the reduced classification features.

In addition, when using only the original bands and remote sensing indices, good classification results were still obtained. This is because the calculations of statistical values, such as the maximum, minimum, and standard deviation, reflect the phenological characteristics of the vegetation. Furthermore, the red-edge indices are the most effective for the identification of tea plantations, followed by topographical features. The TCT and texture features have minimal positive effects on improving the classification accuracy for tea plantations.

In future studies, we will explore the effects of applying other machine learning models to the classification process along with the progressive model to obtain a more precise map of tea plantations. Meanwhile, we will consider extending our research workflow using Landsat time-series data to obtain long-term spatial distributions and changes in tea plantations.

7.

Appendix A

7.1.

J-M Distance Results of Classification Features

Table 8 shows the details of the calculated J-M distance for each feature. The low value of the J-M distance represents low separability and the features with J-M distance values lower than a certain standard (i.e., 0.2) were further eliminated. Table 9 shows the details of the optimized classification features, including the number and the J-M distance values. The total number of classification features was reduced from 116 to 64, a 45% decrease, which dramatically improved the classification efficiency.

Table 8

J-M distance results of classification features.

NameJ-MNameJ-MNameJ-MNameJ-M
B2-max0.085B2-min0.351B2-median0.361B2-std0.124
B3-Max0.093B3-Min0.341B2-median0.400B3-Std0.125
B4-Max0.095B4-Min0.347B3-median0.400B4-Std0.116
B5-Max0.116B5-Min0.311B4-median0.450B5-Std0.114
B6-Max0.171B6-Min0.316B5-median0.440B6-Std0.150
B7-Max0.207B7-Min0.321B6-median0.453B7-Std0.148
B8-Max0.213B8-Min0.306B7-median0.446B8-Std0.104
B8A-Max0.229B8A-Min0.309B8-median0.455B8A-Std0.120
B11-Max0.451B11-Min0.264B8A-median0.483B11-Std0.387
B12-Max0.350B12-Min0.263B11-median0.473B12-Std0.347
NDVI-Max0.279NDVI-Min0.171B12-median0.291NDVI-Std0.297
SAVI-Max0.279SAVI-Min0.170NDVI-median0.292SAVI-Std0.296
NDWI-Max0.206NDWI-Min0.307SAVI-median0.374NDWI-Std0.281
MNDWI-Max0.141MNDWI-Min0.253MNDWI-median0.452MNDWI-Std0.238
NDBI-Max0.216NDBI-Min0.049NDBI-median0.238NDBI-Std0.163
NDVIre1-Max0.164NDVIre1-Min0.160NDVIre1-median0.195NDVIre1-Std0.265
NDVIre2-Max0.283NDVIre2-Min0.234NDVIre2-median0.296NDVIre2-Std0.294
NDVIre3-Max0.636NDVIre3-Min0.438NDVIre3-median0.226NDVIre3-Std0.568
NDre1-Max0.221NDre1-Min0.158NDre1-median0.243NDre1-Std0.293
NDre2-Max0.139NDre2-Min0.174NDre2-median0.174NDre2-Std0.272
IRECI-Max0.256IRECI-Min0.243IRECI-median0.296IRECI-Std0.133
MTCI-Max0.920MTCI-Min0.286MTCI-median0.153MTCI-Std0.583
CIre-Max1.09CIre-Min0.163CIre-median0.152CIre-Std0.947
ASM0.411CONTRAST0.428CORR0.171IDM0.421
ENT0.370SAVG0.325SENT0.468DENT0.466
IMCORR10.432IMCORR20.584SHADE0.476PROM0.589
Note: Max, Min, median, and Std represent the maximum, minimum, median, and standard deviation, respectively.

Table 9

Optimization results of classification features.

TypeBefore optimizationAfter optimization
Original featuresNumberJ-MOptimized featuresNumberJ-M
Original bandsThe maximum, minimum, median,
and standard deviation of the
original bands 2-8, 8A, 11, 12
401.03B7-Max, B8-Max, B8A-Max, B11-Max, B12-Max, B2-Min, B3-Min, B4-Min, B5-Min,B6-Min, B7-Min, B8-Min, B8A-Min, B11-Min, B12-Min, B2-median, B3-median, B4-median, B5-median, B6-median, B7-median, B8-median, B8A-median, B11-median, B12-median, B11-Std, B12-Std270.870
Remote sensing indicesThe maximum, minimum, median,
and standard deviation of NDVI,
SAVI, NDWI, MNDWI, and NDBI
200.983NDVI-Max, SAVI-Max, NDWI-Max, NDBI-Max, NDWI-Min, MNDWI-Min, NDVI-median, SAVI-median, NDWI-median, NDBI-median, MNDWI-median, NDVI-Std, SAVI-Std, NDWI-Std, MNDWI-Std150.895
Red-edge indicesThe maximum, minimum, median,
and standard deviation of NDVIre1, NDVIre2,
NDVIre3, NDre1, NDre2,
IRECI, MTCI, and CIre
321.399CIre-Max, CIre-Std, MTCI-Max, NDVIre3-Max, MTCI-Std, NDVIre3-Std, NDVIre3-Min71.320
Texture featuresASM,CONTRAST,CORR,IDM,ENT,
SAVG,SENT,DENT,IMCORR1,
IMCORR2,SHADE,PROM
121.198ASM, CONTRAST, IDM, SENT, DENT, IMCORR1, IMCORR2, SHADE, PROM91.183
TCT featuresBrightness, greenness, wetness30.528Brightness, greenness, wetness30.528
Topographic featuresElevation, slope, aspect30.370Elevation, slope, aspect30.370
ALL1161.414641.412

7.2.

Optimization Results of Classification Features

Acknowledgments

This research was funded by the National Key Research and Development Program of China, Grant No. 2017YFB0504205; the National Natural Science Foundation of China, Grant No. 42171396; and the Natural Science Research Project of Higher Education in Anhui Province, Grant No. KJ2020A0089. The authors declare that they have no conflicts of interest regarding the publication of this paper.

References

1. 

W. Xu et al., “The pruning phenological phase-based method for extracting tea plantations by field hyperspectral data and Landsat time series imagery,” Geocarto Int., 1 –21 (2020). https://doi.org/10.1080/10106049.2020.1801859 Google Scholar

2. 

Y. Chen and S. Tian, “Feature-level fusion between Gaofen-5 and Sentinel-1A data for tea plantation mapping,” Forests, 11 (12), 1357 (2020). https://doi.org/10.3390/f11121357 FOPEA4 Google Scholar

3. 

W. Xu et al., “Quantifying spatial-temporal changes of tea plantations in complex landscapes through integrative analyses of optical and microwave imagery,” Int. J. Appl. Earth Obs. Geoinf., 73 697 –711 (2018). https://doi.org/10.1016/j.jag.2018.08.010 Google Scholar

4. 

, “Anhui Statistical Yearbook 2020,” http://www.ahtjj.gov.cn/tjj/web/tjnj_view.jsp Google Scholar

5. 

B. Wang et al., “Mapping tea plantations from multi-seasonal landsat-8 OLI imageries using a random forest classifier,” J. Indian Soc. Remote Sens., 47 (8), 1315 –1329 (2019). https://doi.org/10.1007/s12524-019-01014-5 Google Scholar

6. 

Y.-C. Wang et al., “Diverse Colletotrichum species cause anthracnose of tea plants (Camellia sinensis (L.) O. Kuntze) in China,” Sci. Rep., 6 (1), 35287 (2016). https://doi.org/10.1038/srep35287 SRCEC3 2045-2322 Google Scholar

7. 

S. Su, “Economic benefit and ecological cost of enlarging tea cultivation in subtropical China: characterizing the trade-off for policy implications,” Land Use Policy, 66 183 –195 (2017). https://doi.org/10.1016/j.landusepol.2017.04.044 Google Scholar

8. 

P. Phan et al., “Using multi-temporal MODIS NDVI data to monitor tea status and forecast yield: a case study at Tanuyen, Laichau, Vietnam,” Remote Sens., 12 (11), 1814 (2020). https://doi.org/10.3390/rs12111814 Google Scholar

9. 

Z. Tang, M. Li and X. Wang, “Mapping tea plantations from VHR images using OBIA and convolutional neural networks,” Remote Sens., 12 (18), 2935 (2020). https://doi.org/10.3390/rs12182935 Google Scholar

10. 

R. R. Nidamanuri, “Hyperspectral discrimination of tea plant varieties using machine learning, and spectral matching methods,” Remote Sens. Appl. Soc. Environ., 19 100350 (2020). https://doi.org/10.1016/j.rsase.2020.100350 Google Scholar

11. 

N. Li et al., “Mapping the spatial distribution of tea plantations using high-spatiotemporal-resolution imagery in northern Zhejiang, China,” Forests, 10 (10), 856 (2019). https://doi.org/10.3390/f10100856 FOPEA4 Google Scholar

12. 

P. Mishra et al., “Fusing spectral and textural information in near-infrared hyperspectral imaging to improve green tea classification modelling,” J. Food Eng., 249 40 –47 (2019). https://doi.org/10.1016/j.jfoodeng.2019.01.009 JFOEDH 0260-8774 Google Scholar

13. 

J. Zhu et al., “An improved multi-temporal and multi-feature tea plantation identification method using Sentinel-2 imagery,” Sensors, 19 (9), 2087 (2019). https://doi.org/10.3390/s19092087 SNSRES 0746-9462 Google Scholar

14. 

L. Valderrama-Landeros et al., “Extrapolating canopy phenology information using Sentinel-2 data and the Google Earth Engine platform to identify the optimal dates for remotely sensed image acquisition of semiarid mangroves,” J. Environ. Manage., 279 111617 (2021). https://doi.org/10.1016/j.jenvman.2020.111617 Google Scholar

15. 

R. Sonobe, T. Sano and H. Horie, “Using spectral reflectance to estimate leaf chlorophyll content of tea with shading treatments,” Biosyst. Eng., 175 168 –182 (2018). https://doi.org/10.1016/j.biosystemseng.2018.09.018 Google Scholar

16. 

H.-J. Chu et al., “Integration of full-waveform LiDAR and hyperspectral data to enhance tea and areca classification,” GISci. Remote Sens., 53 (4), 542 –559 (2016). https://doi.org/10.1080/15481603.2016.1177249 Google Scholar

17. 

X. Liu et al., “Large-scale crop mapping from multisource remote sensing images in Google Earth Engine,” IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 13 414 –427 (2020). https://doi.org/10.1109/JSTARS.2019.2963539 Google Scholar

18. 

J. M. A. Duncan et al., “Observing climate impacts on tea yield in Assam, India,” Appl. Geogr., 77 64 –71 (2016). https://doi.org/10.1016/j.apgeog.2016.10.004 Google Scholar

19. 

Y.-C. Chuang and Y.-S. Shiu, “A comparative analysis of machine learning with worldview-2 pan-sharpened imagery for tea crop mapping,” Sensors, 16 (5), 594 (2016). https://doi.org/10.3390/s16050594 SNSRES 0746-9462 Google Scholar

20. 

M. Calderón-Loor, M. Hadjikakou and B. A. Bryan, “High-resolution wall-to-wall land-cover mapping and land change assessment for Australia from 1985 to 2015,” Remote Sens. Environ., 252 112148 (2021). https://doi.org/10.1016/j.rse.2020.112148 Google Scholar

21. 

L. Liu et al., “Mapping cropping intensity in China using time series Landsat and Sentinel-2 images and Google Earth Engine,” Remote Sens. Environ., 239 111624 (2020). https://doi.org/10.1016/j.rse.2019.111624 Google Scholar

22. 

N. Gorelick et al., “Google Earth Engine: planetary-scale geospatial analysis for everyone,” Remote Sens. Environ., 202 18 –27 (2017). https://doi.org/10.1016/j.rse.2017.06.031 Google Scholar

23. 

M. N. Jahromi et al., “Google Earth Engine and its application in forest sciences,” Spatial Modeling in Forest Resources Management, 629 –649 Springer, Cham (2021). Google Scholar

24. 

L. Kumar and O. Mutanga, “Google Earth Engine applications,” Remote Sens., 11 (5), 591 (2019). https://doi.org/10.3390/rs11050591 Google Scholar

25. 

L. Chu et al., “Assessing the link between human modification and changes in land surface temperature in Hainan, China using image archives from Google Earth Engine,” Remote Sens., 12 (5), 888 (2020). https://doi.org/10.3390/rs12050888 Google Scholar

26. 

M.-S. Zhao et al., “Spatial-temporal change of soil organic carbon in Anhui Province of East China,” Geoderma Regional, 26 e00415 (2021). https://doi.org/10.1016/j.geodrs.2021.e00415 Google Scholar

27. 

Z.-G. Cao et al., “Spatio-temporal pattern of schistosomiasis in Anhui Province, East China: potential effect of the Yangtze River-Huaihe river water transfer project,” Parasitol. Int., 67 (5), 538 –546 (2018). https://doi.org/10.1016/j.parint.2018.05.007 Google Scholar

28. 

Y. Chen et al., “Spatiotemporal characteristics of winter wheat waterlogging in the middle and lower reaches of the Yangtze River, China,” Adv. Meteorol., 9 1 –11 (2018). https://doi.org/10.1155/2018/3542103 Google Scholar

29. 

Q.-Q. Zhu et al., “Epidemiological characteristics of pulmonary tuberculosis in Anhui Province, Eastern China from 2013 to 2018,” PLoS One, 15 (8), e0237311 (2020). https://doi.org/10.1371/journal.pone.0237311 POLNCL 1932-6203 Google Scholar

30. 

P. Griffiths et al., “Towards national-scale characterization of grassland use intensity from integrated Sentinel-2 and Landsat time series,” Remote Sens. Environ., 238 111124 (2020). https://doi.org/10.1016/j.rse.2019.03.017 Google Scholar

31. 

C. Granero-Belinchon et al., “Phenological dynamics characterization of alignment trees with Sentinel-2 imagery: a vegetation indices time series reconstruction methodology adapted to urban areas,” Remote Sens., 12 (4), 639 (2020). https://doi.org/10.3390/rs12040639 Google Scholar

32. 

P. Macintyre, A. van Niekerk and L. Mucina, “Efficacy of multi-season Sentinel-2 imagery for compositional vegetation classification,” Int. J. Appl. Earth Obs. Geoinf., 85 101980 (2020). https://doi.org/10.1016/j.jag.2019.101980 Google Scholar

33. 

S. Qiu, Z. Zhu and B. He, “Fmask 4.0: improved cloud and cloud shadow detection in Landsats 4–8 and Sentinel-2 imagery,” Remote Sens. Environ., 231 111205 (2019). https://doi.org/10.1016/j.rse.2019.05.024 Google Scholar

34. 

P. Li et al., “A new quality validation of global digital elevation models freely available in China,” Surv. Rev., 48 (351), 409 –420 (2016). https://doi.org/10.1179/1752270615Y.0000000039 Google Scholar

35. 

A. Mouratidis and D. Ampatzidis, “European digital elevation model validation against extensive global navigation satellite systems data and comparison with SRTM DEM and ASTER GDEM in Central Macedonia (Greece),” ISPRS Int. J. Geo-Inf., 8 (3), 108 (2019). https://doi.org/10.3390/ijgi8030108 Google Scholar

36. 

G. Pulighe, V. Baiocchi and F. Lupia, “Horizontal accuracy assessment of very high resolution Google Earth images in the city of Rome, Italy,” Int. J. Digital Earth, 9 (4), 342 –362 (2016). https://doi.org/10.1080/17538947.2015.1031716 Google Scholar

37. 

J. J. Arsanjani, A. Tayyebi and E. Vaz, “GlobeLand30 as an alternative fine-scale global land cover map: challenges, possibilities, and implications for developing countries,” Habitat Int., 55 25 –31 (2016). https://doi.org/10.1016/j.habitatint.2016.02.003 Google Scholar

38. 

V. Upadhyay and A. Kumar, “Hyperspectral remote sensing of forests: technological advancements, opportunities and challenges,” Earth Sci. Inf., 11 (4), 487 –524 (2018). https://doi.org/10.1007/s12145-018-0345-7 Google Scholar

39. 

J. M. Peña-Barragán et al., “Object-based crop identification using multiple vegetation indices, textural features and crop phenology,” Remote Sens. Environ., 115 (6), 1301 –1316 (2011). https://doi.org/10.1016/j.rse.2011.01.009 Google Scholar

40. 

P. Hurskainen et al., “Auxiliary datasets improve accuracy of object-based land use/land cover classification in heterogeneous savanna landscapes,” Remote Sens. Environ., 233 111354 (2019). https://doi.org/10.1016/j.rse.2019.111354 Google Scholar

41. 

P. D’Odorico et al., “Experimental evaluation of Sentinel-2 spectral response functions for NDVI time-series continuity,” IEEE Trans. Geosci. Remote Sens., 51 (3), 1336 –1348 (2013). https://doi.org/10.1109/TGRS.2012.2235447 IGRSD2 0196-2892 Google Scholar

42. 

W. J. Frampton et al., “Evaluating the capabilities of Sentinel-2 for quantitative estimation of biophysical variables in vegetation,” ISPRS J. Photogramm. Remote Sens., 82 83 –92 (2013). https://doi.org/10.1016/j.isprsjprs.2013.04.007 IRSEE9 0924-2716 Google Scholar

43. 

X. Yang et al., “Mapping of urban surface water bodies from Sentinel-2 MSI imagery at 10 m resolution via NDWI-based image sharpening,” Remote Sens., 9 (6), 596 (2017). https://doi.org/10.3390/rs9060596 Google Scholar

44. 

Y. Du et al., “Water bodies’ mapping from Sentinel-2 imagery with modified normalized difference water index at 10-m spatial resolution produced by sharpening the SWIR band,” Remote Sens., 8 (4), 354 (2016). https://doi.org/10.3390/rs8040354 Google Scholar

45. 

M. Piragnolo et al, “Comparison of vegetation indices from RPAS and sentinel-2 imagery for detecting permanent pastures,” Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci., XLII-3 (3), 1381 –1387 (2018). https://doi.org/10.5194/isprs-archives-XLII-3-1381-2018 1682-1750 Google Scholar

46. 

C. Li et al., “Estimating apple tree canopy chlorophyll content based on Sentinel-2A remote sensing imaging,” Sci. Rep., 8 (1), 1 –10 (2018). https://doi.org/10.1038/s41598-018-21963-0 SRCEC3 2045-2322 Google Scholar

47. 

S. Filin, “Surface clustering from airborne laser scanning data,” Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci., 34 (3/A), 119 –124 (2002). 1682-1750 Google Scholar

48. 

Q. Feng, J. Liu and J. Gong, “UAV remote sensing for urban vegetation mapping using random forest and texture analysis,” Remote Sens., 7 (1), 1074 –1094 (2015). https://doi.org/10.3390/rs70101074 Google Scholar

49. 

P. Mohanaiah, P. Sathyanarayana and L. GuruKumar, “Image texture feature extraction using GLCM approach,” Int. J. Sci. Res. Publ., 3 (5), 1 –5 (2013). Google Scholar

50. 

D. Gadkari, Image Quality Analysis Using GLCM, University of Central Florida Orlando, Florida (2004). Google Scholar

51. 

A. Tassi and M. Vizzari, “Object-oriented LULC classification in Google Earth Engine combining SNIC, GLCM, and machine learning algorithms,” Remote Sens., 12 (22), 3776 (2020). https://doi.org/10.3390/rs12223776 Google Scholar

52. 

T. Shi and H. Xu, “Derivation of tasseled cap transformation coefficients for Sentinel-2 MSI at-sensor reflectance data,” IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 12 (10), 4038 –4048 (2019). https://doi.org/10.1109/JSTARS.2019.2938388 Google Scholar

53. 

J. Lastovicka et al., “Sentinel-2 data in an evaluation of the impact of the disturbances on forest vegetation,” Remote Sens., 12 (12), 1914 (2020). https://doi.org/10.3390/rs12121914 Google Scholar

54. 

Y. Chen et al., “Hierarchical object oriented classification using very high resolution imagery and LIDAR data over urban areas,” Adv. Space Res., 43 (7), 1101 –1110 (2009). https://doi.org/10.1016/j.asr.2008.11.008 ASRSDW 0273-1177 Google Scholar

55. 

J. Zhang et al., “Machine learning-based spectral library for crop classification and status monitoring,” Agronomy, 9 (9), 496 (2019). https://doi.org/10.3390/agronomy9090496 AGRYAV 0065-4663 Google Scholar

56. 

F. Löw et al., “Impact of feature selection on the accuracy and spatial uncertainty of per-field crop classification using support vector machines,” ISPRS J. Photogramm. Remote Sens., 85 102 –119 (2013). https://doi.org/10.1016/j.isprsjprs.2013.08.007 IRSEE9 0924-2716 Google Scholar

57. 

Y. Wang, Q. Qi and Y. Liu, “Unsupervised segmentation evaluation using area-weighted variance and Jeffries-Matusita distance for remote sensing images,” Remote Sens., 10 (8), 1193 (2018). https://doi.org/10.3390/rs10081193 Google Scholar

58. 

R. Sen, S. Goswami and B. Chakraborty, “Jeffries-Matusita distance as a tool for feature selection,” in Int. Conf. Data Sci. and Eng., 15 –20 (2019). https://doi.org/10.1109/ICDSE47409.2019.8971800 Google Scholar

59. 

L. Bruzzone, F. Roli and S. B. Serpico, “An extension of the Jeffreys-Matusita distance to multiclass cases for feature selection,” IEEE Trans. Geosci. Remote Sens., 33 (6), 1318 –1321 (1995). https://doi.org/10.1109/36.477187 IGRSD2 0196-2892 Google Scholar

Biography

Le’an Qu received his ME degree in geographic information science from Chang’an University, Xi’an, China, in 2009. He is currently working toward his PhD in geography from Nanjing University, Nanjing, China. His research interests include machine learning and remote sensing applications.

Manchun Li received his PhD in cartography from Nanjing University, Nanjing, China, in 1992. He is currently a professor at Jiangsu Provincial Key Laboratory of Geographic Information Science and Technology, Nanjing University. His research interests include GIS and remote sensing applications.

Zhenjie Chen received his PhD in cartography and geographic information science from Nanjing University, Nanjing, China, in 2006. He is currently an associate professor at Jiangsu Provincial Key Laboratory of Geographic Information Science and Technology, Nanjing University. His research interests include GIS design and remote sensing applications.

Wangbing Liu received his MS degree in agricultural remote sensing and information technology from Zhejiang University, Hangzhou, China, in 2013. He is currently an engineer with Anhui Province Institute of Land Surveying and Planning, Key Laboratory of JiangHuai Arable Land Resources Protection and Eco-restoration, Ministry of Natural Resources. His research interests include remote sensing applications.

Junjun Zhi received his PhD in agricultural remote sensing and information technology from Zhejiang University, Hangzhou, China, in 2014. He is currently an associate professor at the School of Geography and Tourism, Anhui Normal University. His research interests include machine learning and agricultural remote sensing applications.

Lechun Zhang is currently working toward his BSc degree in GIS from Anhui Normal University, Wuhu, China. His research interests include GIS design and applications.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Le’an Qu, Manchun Li, Zhenjie Chen, Wangbing Liu, Junjun Zhi, and Lechun Zhang "Mapping large area tea plantations using progressive random forest and Google Earth Engine," Journal of Applied Remote Sensing 16(2), 024509 (2 May 2022). https://doi.org/10.1117/1.JRS.16.024509
Received: 24 November 2021; Accepted: 12 April 2022; Published: 2 May 2022
Lens.org Logo
CITATIONS
Cited by 3 scholarly publications.
Advertisement
Advertisement
KEYWORDS
Remote sensing

Image classification

Vegetation

Statistical modeling

Neodymium

Process modeling

Data modeling

Back to Top