Next Article in Journal
The Influence of Multistakeholder Value Cognition and Risk Attitudes on Sustainable Interior Landscape Design Decisions
Next Article in Special Issue
Urban and Peri-Urban Vegetation Monitoring Using Satellite MODIS NDVI Time Series, Singular Spectrum Analysis, and Fisher–Shannon Statistical Method
Previous Article in Journal
Sustainable and Environmental Dyeing with MAUT Method Comparative Selection of the Dyeing Recipe
Previous Article in Special Issue
Impacts of Land Use and Land Cover Changes on Land Surface Temperature over Cachar Region, Northeast India—A Case Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning in Forest Tree Species Classification Using Sentinel-2 on Google Earth Engine: A Case Study of Qingyuan County

1
College of Mathematics and Computer Science, Zhejiang A&F University, Hangzhou 311300, China
2
Zhejiang Provincial Key Laboratory of Forestry Intelligent Monitoring and Information Technology, Hangzhou 311300, China
3
School of Information Engineering, Xinjiang Institute of Technology, Aksu 843100, China
4
College of Economics and Management, Zhejiang A&F University, Hangzhou 311300, China
5
Research Academy for Rural Revitalization of Zhejiang Province, Zhejiang A&F University, Hangzhou 311300, China
6
Institute of Ecological Civilization, Zhejiang A&F University, Hangzhou 311300, China
7
Financial Big Data Research Institute, Sunyard Technology Co., Ltd., Hangzhou 310053, China
8
College of Computer Science and Technology, Zhejiang University, Hangzhou 310063, China
9
Zhejiang Shuren University, Hangzhou 310015, China
*
Authors to whom correspondence should be addressed.
Sustainability 2023, 15(3), 2741; https://doi.org/10.3390/su15032741
Submission received: 28 December 2022 / Revised: 13 January 2023 / Accepted: 31 January 2023 / Published: 2 February 2023

Abstract

:
Forest tree species information plays an important role in ecology and forest management, and deep learning has been used widely for remote sensing image classification in recent years. However, forest tree species classification using remote sensing images is still a difficult task. Since there is no benchmark dataset for forest tree species, a forest tree species dataset (FTSD) was built in this paper to fill the gap based on the Sentinel-2 images. The FTSD contained nine kinds of forest tree species in Qingyuan County with 8,815 images, each with a resolution of 64 × 64 pixels. The images were produced by combining forest management inventory data and Sentinel-2 images, which were acquired with less than 20% clouds from 1 April to 31 October, including the years 2017, 2018, 2019, 2020, and 2021. Then, the images were preprocessed and downloaded from Google Earth Engine (GEE). Four different band combinations were compared in the paper. Moreover, a Principal Component Analysis (PCA) and Normalized Difference Vegetation Index (NDVI) were also calculated using the GEE. Deep learning algorithms including DenseNet, EfficientNet, MobileNet, ResNet, and ShuffleNet were trained and validated in the FTSD. RGB images with red, green, and blue (PC1, PC2, and NDVI) obtained the highest validation accuracy in four band combinations. ResNet obtained the highest validation accuracy in all algorithms after 500 epochs were trained in the FTSD, which reached 84.91%. As a famous and widely used remote sensing classification satellite imagery dataset, NWPU RESISC-45 was also trained and validated in the paper. ResNet achieved a high validation accuracy of 87.90% after training 100 epochs in NWPU RESISC-45. The paper shows in forest tree species classification based on remote sensing images and deep learning that (1) PCA and NDVI can be combined to improve the accuracy of classification; (2) ResNet is more suitable than other deep learning algorithms including DenseNet, EfficientNet, MobileNet, and ShuffleNet in remote sensing classification; and (3) being too shallow or deep in ResNet does not perform better in the FTSD, that is, 50 layers are better than 34 and 101 layers.

1. Introduction

Forest tree species play a virtual role in forest management and forest resource surveys [1,2,3], which have been adopted as an indicator for evaluating forest carbon. Accurate classification of forest tree species is conducive to mastering the overall situation and ecology quality of forest resources in a country or region.
Ground survey of forest resources is time consuming and inefficient. Furthermore, some places cannot be covered. However, remote sensing images provide more survey area and higher frequency than human field surveys, which have been used for forest tree species classification. High-resolution remote sensing images such as IKONOS or WorldView are used for forest tree species classification [4]. Immitzer et al. [5] adopted the high spatial resolution of 8-band WorldView-2 satellite data to classify and identify 10 tree species in Austrian temperate forests. Fang et al. [6] used multi-temporal WorldView-3 imagery to classify tree species in different taxonomic levels in Washington D.C., USA. Multispectral remote sensing images have been combined with other sensors, which include airborne laser scanning, light detection and ranging [7], and drone-based image point clouds [8] in individual tree species identification. However, the above studies on forest tree species classification based on remote sensing images have only focused on a small area, and cannot be applied to a large area.
Medium-resolution remote sensing images enable the monitoring of larger areas than high-resolution images. The repetition cycle of the Sentinel-2 satellites (2–5 days) is higher than Landsat (8–16 days). Furthermore, Sentinel-2 provides higher spatial resolution (10–20 m) than Landsat (30 m), so more detailed information can be captured from Sentinel-2 [9]. Axelsson et al. [10] used Sentinel-2 images and Bayesian inference for tree species classification and selected images from April to October. Immitzer et al. [11] classified tree species using Sentinel-2 and a random forest classification, and they found that images from April to August obtained the highest accuracy. Persson et al. [12] classified five tree species using four images (April, May, July, and October) of the same year in a random forest classification. Puletti et al. [13] selected Sentinel-2 images of spring, summer, and autumn and used a random forest classification to discriminate between coniferous, broad-leaved, and mixed forests, achieving a maximum overall accuracy of 86.2% using a separate validation dataset. From the above, it can be seen that random forest classification is usually used in tree species classification, but it easily leads to overfitting, so deep learning was used in this paper. The images were selected from April to October in this paper.
Principal Component Analysis (PCA) is a useful and important tool for dimension reduction [14]. Due to remote sensing images having more bands than RGB channels, PCA was used in our paper to retain most information and combine it with RGB images. Zhang et al. [15] used PCA to detect clouds using remote sensing. Li et al. [16] adopted PCA and other methods to calculate the ecology index of the Beijing-Hangzhou Grand Canal. The Normalized Difference Vegetation Index (NDVI) has also been widely applied to classify vegetation units [17]. Therefore, PCA and NDVI were used in the paper for band combination.
As the amount of available data has increased, a convolutional neural network (CNN) has been developed in tree species classification [18]. Li et al. (2016) used a CNN to detect and count oil palm trees in remote sensing images [19]. Mubin et al. [20] developed a CNN to detect juvenile and mature oil palms. Freudenberg et al. [21] applied a UNET neural network to detect large-scale palm tree images. From the above research, it can be seen that CNN has not been used for multi-type forest tree species classification.
This study aims to answer the questions of which band combination is the best combination for RGB visual identification, and which deep learning algorithm is the best estimator in the above band combinations.
The other object of the study was to build a forest tree species dataset as a benchmark of classification, utilizing other applications with forest management inventories, and verify which deep learning was well accomplished in forest tree species identification using Sentinel-2 images. A forest tree species dataset was built in this paper named FTSD, which includes nine kinds of forest tree species with 8815 images. Due to utilizing various band information of Sentinel-2, four different band combinations were compared in the paper, and PCA and NDVI were also used. Eight kinds of different deep learning algorithms were trained and validated, and ResNet50 reached the best validation accuracy in the FTSD and NWPU RESISC-45.

2. Material and Methods

2.1. Overall Workflow

The study consisted of three primary steps (Figure 1): first, the Sentinel-2 images of the study area were processed and downloaded from the GEE platform; secondly, GGC was calculated by *.shp files of the domain tree species in forest management inventories, and sample images were clipped through GGC in ArcGIS 10.4 software; finally, the above images were trained and validated using deep learning algorithms with Pytorch.

2.2. Study Area

The study area of Qingyuan County is located in Zhejiang Province, which is in southeastern China. The geographical location is 118°50′–119°30′ E and 27°25′–27°51′ N, as shown in Figure 2. The county has a total area of 1898 km2, which spans 49 km from north to south and 67 km from east to west.
Qingyuan County has a subtropical monsoon climate with four distinct seasons throughout the year, which is warm in winter and cool in summer. The average temperature is 17.4 °C, the precipitation is 1760 mm, and the frost-free period is 245 days annually. Qingyuan County is rich in natural resources, and its forest land is 1.6233 × 105 hectares in 1.8976 × 105 hectares of total land, with the forest land accounting for 85.5% of the county.

2.3. Sentinel-2

The Sentinel-2 images were downloaded from GEE (https://earthengine.google.com, accessed on 15 May 2022), which are released in the Level-1C format. Radiometric, atmospheric, and cloud corrections but no other processing is required in Level-1C format product [22]. The bands of Sentinel-2 are shown in Table 1.
Level-1C products utilize a Digital Elevation Model (DEM) to project the images into cartographic geometry. In addition to per-pixel radiometric measurements in Top of Atmosphere (TOA) reflectance, the products also include parameters for converting the measurements into radiances. Level-1C products are resampled with constant Ground Sampling Distances (GSDs) of 10, 20, and 60 m depending on the native resolution of the different spectral bands. To generate images with an equal spatial resolution in all bands, images were re-sampled to 10 m [23] in GEE before being downloaded in the paper. Attributed to the Sentinel-2 images being sensitive to weather conditions, all images were acquired with less than 20% clouds from 1 April to 31 October, including the years 2017, 2018, 2019, 2020, and 2021.
To compare classification accuracy, four different band combinations were used in our paper, which are shown in Table 2.
As a standard remote sensing product for discriminating and interpreting mapped vegetation units, NDVI is often used in classification [17]. NDVI is calculated by the following equation:
N D V I = N I R R e d N I R + R e d
where NIR is the near-infrared band and Red is the red band, which is Band 8 and Band 4 in Sentinel-2 images. PC1, PC2, and PC3 are the first, second, and third most important principal components after PCA manipulation.

2.4. Forest Management Inventories

Forest management inventories produce stand-level estimates to guide management decisions using data obtained with sampling intensities, which include the most detailed and up-to-date information on trees. The inventories of the study area were updated in 2020, and geographical information on dominant tree species was used in this paper. GGC was calculated, which was stored as latitude and longitude together with attribute information.

2.5. Dataset Description

2.5.1. Forest Tree Species Dataset (FTSD)

The forest tree species dataset (FTSD) was used in our research to classify forest tree species using Sentinel-2 data. To ensure clipped Sentinel-2 images in the dominant tree species zone, a minimum area of 8 hectares was acquired and a sample balance of forest tree species was considered. Therefore, 9 kinds of forest tree species with different areas were selected in the inventory data as Table 3. The dataset contains 9 dominant tree species and 8815 RGB images with 64 × 64 pixels. The spatial resolution of images is 10 m per pixel.

2.5.2. NWPU RESISC-45

RESISC-45 (https://www.tensorflow.org/datasets/catalog/resisc45, accessed on 15 May 2022) is a famous satellite imagery database and is widely used in remote sensing classification [24]. It includes 45 scene classes and 31,500 RGB images. Each class includes 700 images with a size of 256 × 256 pixels. The spatial resolution of the dataset ranges from 20 cm to more than 30 m, and it covers a wide range of countries [25]. A forest class is in the dataset, but there is no subdivision category of forest in the dataset. Therefore, the dataset cannot be provided for forest tree species classification.

2.6. Methods

The main steps of the study are as follows: first, the geographical geometric center (GGC) was calculated for the nine kinds of forest tree species, based on forest management inventories and acquiring an area larger than 8 hectares to ensure the clipping image in the same tree species; similarly, after some pre-processing, Sentinel-2 images were downloaded from the GEE platform in the study area; then, the downloaded images were clipped with GGCs of different kinds of forest tree species; and finally, the clipped images with different deep learning algorithms were trained, validated, and compared to find the best validation accuracy algorithm.

3. Methodology

3.1. Data Preprocess

After downloading Sentinel-2 images from GEE, the center mass of the geographical location was calculated in ArcGIS 10.4. The main process is described as follows. Firstly, the .shp files of tree species were selected, which are described in Table 1, then the corresponding features were changed to points as the center of geographical location and images were clipped using a square with 64 × 64 pixels. An example of the above steps is described in Figure 3.
Visual inspection was taken for removing some low-quality images (Figure 4) from the dataset, and the quantity of tree species images was shown in Table 1 in the images column.

3.2. Transfer Learning

Due to few samples, the transfer learning method has been widely used in image classification and other fields [26]. Because higher layers of representation are crucial to classification [27], transfer learning was used in this study for obtaining deeper layers. After being trained in a big dataset, the parameters of the deep learning model are preserved, which will improve the accuracy and efficiency of forest tree species identification.
ImageNet is usually used for training models as transfer learning [28]. By dropping the final layer parameters of the trained model and fixing other features, then training and learning complex features can be achieved from the Sentinel-2 images. A fully connected layer, a batch normal layer, and a dropout layer were sequentially combined, then connected to a dense net as the final layer to classify forest tree species.
Eight deep learning models were used to train in both datasets using transfer learning to compare them with each other. Five kinds of popular deep learning models of CNN, which include DenseNet, EfficientNet, MobileNet, ResNet, and ShuffleNet, were used for forest tree species classification by transfer learning.

3.2.1. DenseNet

DenseNet was introduced with a densely connected convolutional network architecture [29]. For retaining more information between layers in the feed-forward network, all layers are connected directly with each other. Feature maps of all preceding layers are used as inputs in all layers. The problem of vanishing gradient has been alleviated in DenseNet, which has a small number of parameters. DenseNet121 was used in our study, which contains 121 layers. Pre-trained weights of the model were downloaded from PyTorch.

3.2.2. EfficientNet

The EfficientNet model consists of a group of 8 CNN models, and the Swish activation function was used while the Rectifier Linear Unit (ReLU) activation function was adopted in other CNN models [30]. EfficientNet also aims to achieve better accuracy with fewer parameters. EfficientNet v2 trains up to 11x faster while being up to 6.8× smaller compared to EfficientNet and other CNN models, which is a smaller and faster CNN model in image classification [31]. Therefore, EfficientNet v2 was used for forest tree species classification.

3.2.3. MobileNet

MobileNet v2 was first proposed by Sandler [32] in 2018, which was based on MobileNet v1. The model reduces parameters by using deep separable convolution, obtains deeper layers by proposing a reverse residual structure, and reduces the loss of low-dimensional features by using a linear bottleneck structure. Therefore, MobileNet v2 was used for forest tree species classification.

3.2.4. ResNet

ResNet was first introduced by He et al. [33] in 2015, and the model won 1st place with an error rate of 3.57% in the ImageNet Large Scale Visual Recognition Challenge. A network-in-network architecture is adopted in ResNet, which relies on many stacked residual units. These residual units are used to construct the network. The residual units are composed of convolution and average pooling layers [34]. With a further updating of the residual module to use identity mappings to update ResNet, a higher accuracy will be obtained [33]. Pre-trained weights with 50,101,152 layers of ResNet were downloaded from Pytorch and used in our paper.

3.2.5. ShuffleNet

ShuffleNet v2 is an upgraded version of ShuffleNet v1 [35]. To minimize the memory access cost in the convolution operation, the model keeps the same channels of input and output layers in the convolution and uses group convolution [36]. To enhance parallel operation ability, the model minimizes branch structure to improve the network operation speed by reducing element-level operations. Due to the above advantages, ShuffleNet v2 was also used for forest tree species classification.
In comparison with the above-mentioned CNNs, ResNet aims to solve the problem of vanishing gradients that occur in very deep networks by introducing residual connections, which allows the network to learn identity mappings as well as more complex mappings. This allows for much deeper networks, with over 100 layers, to be trained effectively.

4. Results

4.1. Training

To calculate the best parameters, Stochastic Gradient Descent (SGD) was used for training models, which makes it run faster and converge easily when training models [33]. SGD is an optimization algorithm used for finding the minimum of a function. SGD is particularly useful when the function to be minimized is very large or has many local minima, because it is computationally efficient and only requires one training example at a time to update the parameters. SGD is widely used in machine learning and deep learning, particularly for training neural networks. The batch normalization technique and ReLU activation function [37] were adopted in all models. The batch size was set as 16, which was limited to GPU memory. A learning rate of 0.001 was set up in all networks. For a common evaluation framework, a standard training and validation split was completed using the 80% and 20% ratios, respectively.
Data augmentations, which include up and down flip and random crop of 64 × 64 pixel images, were performed to all networks to relieve the lack of quantity of remote sensing images.

4.2. Results of the Experiments

Deep learning algorithms including DenseNet121, EfficientNetv2_s, MobileNet_v3_large, ResNet34, ResNet50 v2, ResNet101 v2, and ShuffleNetv2_x1_0 were trained and validated in the FTSD. In our paper, 80% of the images were for training, and the other 20% images were for validation. All models were trained and validated for 100 epochs in different band combinations.
The accuracy metric is usually adopted for the evaluation of the models and used to find the appropriateness of state-of-the-art deep learning [26]. The accuracy formula is as follows:
a c c u r a c y = T P + T N T P + T N + F P + F N
where TP is True Positive, TN is True Negative, FP is False Positive, and FN is False Negative.
The best validation accuracies of different bands after 100 epochs of training are shown in Table 4.
From Table 4, PC1, PC2, and NDVI grouped as red, green, and blue channels obtained the best accuracy in all deep learning algorithms.
The accuracy of the eight different deep learning algorithms with PCANDVI in the FTSD is shown in Figure 5.
Figure 5 shows that the performance of ResNet is better than other deep learning algorithms in the FTSD, and ResNet50 is better than ResNet34 and ResNet101. This is consistent with the results of Maxim Neumann’s paper [24]. Eight different deep learning algorithms were also trained and validated after 100 epochs of training in RESISC-45, and accuracy is shown in Figure 6. ResNet50 obtained the highest validation accuracy with 87.90%.
The confusion matrix is used to visualize the performance of a classifier and is usually used to assess remote sensing classification [38]. The confusion matrix depicts the proper and mistaken classification for each category [39]. The confusion matrix of ResNet50 in the FTSD is shown in Figure 7.
Classification mistakes were focused on the following aspects which were concluded from Figure 7: (1) the same genus of tree species led to wrong classification, such as that nine Pinus massoniana Lamb. were misclassified as Pinus taiwanensis; (2) intersection and overlap between species in the dataset took place, especially in mixed classes, as nine of the mixed conifers were mistaken as mixed broadleaf-conifers.
The precision formula is as follows:
precision = T P T P + F P
The recall was calculated as the following equation.
a c c u r a c y = T P T P + F N
Precision and recall of classification in the FTSD are shown as follows in Table 5. The precisions of the mixed broadleaf and Pinus taiwanensis were 0.789 and 0.756, which shows that the algorithm mistakenly classified other forest tree species as these species. The recall of mixed conifer was the lowest accuracy at 0.787, which depicts that mixed conifer species cannot be identified as a conifer. Bambusoideae was identified well in the FTSD, and the precision and recall of Bambusoideae were 0.958 and 0.968, respectively.

5. Discussion

More data will help to improve the accuracy of deep learning, and Sentinel-2 has 13 spectral bands. However, almost all deep learning methods in computer vision use RGB images. Therefore, our study transformed images from all bands to RGB images by PCA. PCA is usually used to find the best features [40] and retains most information from all bands. Table 4 shows that the band combinations of PCA (PC1, PC2, and PC3) obtained higher validation accuracy than 843 (Band 8, Band 4, and Band 3) and 432 (Band 4, Band 3, and Band 2) when using ResNet or ShuffleNet, but did not perform well in other algorithms. Therefore, detaining most information from all bands is just one of the important factors but is not enough to identify forest tree species. As one of the most widely used indices to assess vegetation from remote sensing imagery [41], NDVI was added to our study for improving accuracy. Table 4 shows that the combination of PCANDVI (PC1, PC2, and NDVI) obtained higher accuracy than other combinations (PCA, 843, and 432) in all deep learning classifications.
Many models showed a trend toward improved classification accuracy as epochs increased [42]. In deep learning and other machine learning methods, excessive training led to no help in improving validation classification accuracy because the testing data are overfitted to the training data. Overfitting was found after 200 epochs of training in Figure 5 and 50 epochs depicted in Figure 6. ResNet is a network-in-network architecture that relies on many stacked residual units [33]. Because of the usage of global average pooling rather than fully connected layers, it can obtain more accuracy by updating the residual module to use identity mappings. ResNet usually shows better results than other CNNs, which is also demonstrated in Figure 5 and Figure 6. The deeper the layers of the network, the higher accuracy the model achieves in the same dataset [43]. However, this was not the same as the result of our study. Figure 5 and Figure 6 show that ResNet34 and ResNet50 performed better than ResNet101, which has deeper layers of the network.
Forest tree species and their distribution are key factors in a regional ecology [44]. To broaden the use of forest management inventory, dominant forest tree species information was used in the study. The dominant tree species were labeled as the only class in each stand. However, different tree species are on the same stand, which becomes a big challenge for forest tree species classification. With more heterogeneity, the more difficult the task of classification is. Therefore, intersection and overlap between species in the dataset, especially in mixed classes, were easily misclassified in this study, which is shown in Figure 7 and nine mixed conifers were mistaken as mixed broadleaf-conifers. Another challenge of forest tree species classification is that some tree species belong to the same genus. Owing to the many similarities between them, it is hard to classify them. For example, nine Pinus massoniana Lamb. were misclassified as Pinus taiwanensis, which is depicted in Figure 7.

6. Conclusions and Future Work

Forest management inventories not only produce stand-level estimates to guide management decisions but also can be used for forest tree species classification. Primary tree species information in forest management inventories was used to clip Sentinel-2 images in our paper. Therefore, another application of forest management inventories was found. In our study, PCA and NDVI were extracted from images, and RGB image classification with PC1, PC2, and NDVI reached the best performance band combinations. This paper shows that the combination of PC1, PC2, and NDVI is suitable for remote sensing classification, and it promotes classification accuracy.
Deep learning classification methods, such as DenseNet, EfficientNet, MobileNet, ResNet, and ShuffleNet, were used to train and validate. The result showed that the combination with red, green, and blue (PC1, PC2, and NDVI) adopted with ResNet50 obtained the highest accuracy rate of 84.91% after 500 epochs of training in the FTSD; ResNet50 also reached the highest accuracy rate of 87.90% after 100 epochs of training in RESISC-45.
In this study, the Sentinel-2 images were combined with a deep learning method and used to classify forest tree species in the research area. More remote sensing images of forest tree species will be collected in future work as a standard recognized benchmark dataset and evaluation frameworks, and higher spatial resolution remote sensing images will be used for comparison.

Author Contributions

Conceptualization, T.H. and H.Z.; methodology, C.X. and L.X.; software, T.H.; validation, T.H. and K.Z.; formal analysis, J.H.; investigation, X.X.; resources, X.L.; data curation, K.Z. and Q.W.; writing—original draft preparation, T.H.; writing—review and editing, T.H., H.Z., C.X., X.X. and L.X.; visualization, X.X.; supervision, H.Z.; project administration, X.X.; funding acquisition, C.X., L.X. and J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National Nature Science Foundation of China (No.31971493), the Zhejiang Provincial Natural Science Foundation of China (No. LGG22D010001 and No. LGF22C160002), Humanities and Social Sciences in Colleges and Universities of Zhejiang Province (No. 2021QN062), Zhejiang Education Department Foundation of China (No. Y202147381), and the research development fund project of Zhejiang A&F University (No. 2021LFR040 and No. 2020FR066).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

RESISC-45 presented in this study is available at https://www.tensorflow.org/datasets/catalog/resisc45, accessed on 15 May 2022 and the FTSD can be obtained by contacting the corresponding authors.

Acknowledgments

The authors would like to thank the editors and anonymous reviewers for their valuable comments, which are significant for improving this manuscript. This work was supported in part by the Zhejiang Provincial Key Laboratory of Forestry Intelligent Monitoring and Information Technology.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gaem, P.H.; Andrade, A.; Mazine, F.F.; Vicentini, A. Tree species delimitation in tropical forest inventories: Perspectives from a taxonomically challenging case study. For. Ecol. Manag. 2022, 505, 119900. [Google Scholar] [CrossRef]
  2. Liu, J.; Wang, X.; Wang, T. Classification of tree species and stock volume estimation in ground forest images using Deep Learning. Comput. Electron. Agric. 2019, 166, 105012. [Google Scholar] [CrossRef]
  3. Lindberg, E.; Holmgren, J.; Olsson, H. Classification of tree species classes in a hemi-boreal forest from multispectral airborne laser scanning data using a mini raster cell method. Int. J. Appl. Earth Obs. Geoinf. 2021, 100, 102334. [Google Scholar] [CrossRef]
  4. Immitzer, M.; Böck, S.; Einzmann, K.; Vuolo, F.; Pinnel, N.; Wallner, A.; Atzberger, C. Fractional cover mapping of spruce and pine at 1 ha resolution combining very high and medium spatial resolution satellite imagery. Remote Sens. Environ. 2017, 204, 690–703. [Google Scholar] [CrossRef]
  5. Immitzer, M.; Atzberger, C.; Koukal, T. Tree Species Classification with Random Forest Using Very High Spatial Resolution 8-Band WorldView-2 Satellite Data. Remote Sens. 2012, 4, 2661–2693. [Google Scholar] [CrossRef]
  6. Fang, F.; McNeil, B.E.; Warner, T.A.; Maxwell, A.E.; Dahle, G.A.; Eutsler, E.; Li, J. Discriminating tree species at different taxonomic levels using multi-temporal WorldView-3 imagery in Washington D.C., USA. Remote Sens. Environ. 2020, 246, 111811. [Google Scholar] [CrossRef]
  7. Sačkov, I.; Sedliak, M.; Kulla, L.; Bucha, T. Inventory of Close-to-Nature Forests Based on the Combination of Airborne LiDAR Data and Aerial Multispectral Images Using a Single-Tree Approach. Forests 2017, 8, 467. [Google Scholar] [CrossRef]
  8. Kotivuori, E.; Kukkonen, M.; Mehtätalo, L.; Maltamo, M.; Korhonen, L.; Packalen, P. Forest inventories for small areas using drone imagery without in-situ field measurements. Remote Sens. Environ. 2020, 237, 111404. [Google Scholar] [CrossRef]
  9. Yin, H.; Khamzina, A.; Pflugmacher, D.; Martius, C. Forest cover mapping in post-Soviet Central Asia using multi-resolution remote sensing imagery. Sci. Rep. 2017, 7, 1375. [Google Scholar] [CrossRef]
  10. Axelsson, A.; Lindberg, E.; Reese, H.; Olsson, H. Tree species classification using Sentinel-2 imagery and Bayesian inference. Int. J. Appl. Earth Obs. Geoinf. 2021, 100, 102318. [Google Scholar] [CrossRef]
  11. Immitzer, M.; Neuwirth, M.; Bck, S.; Brenner, H.; Atzberger, C. Optimal Input Features for Tree Species Classification in Central Europe Based on Multi-Temporal Sentinel-2 Data. Remote Sens. 2019, 11, 2599. [Google Scholar] [CrossRef]
  12. Persson, M.; Lindberg, E.; Reese, H. Tree species classification with multi-temporal Sentinel-2 data. Remote Sens. 2018, 10, 1794. [Google Scholar] [CrossRef]
  13. Puletti, N.; Chianucci, F.; Castaldi, C. Use of Sentinel-2 for forest classification in Mediterranean environments. Ann. Silvic. Res 2018, 42, 32–38. [Google Scholar]
  14. Wang, S.H.; Huang, S.Y. Perturbation theory for cross data matrix-based PCA. J. Multivar. Anal. 2022, 190, 104960. [Google Scholar] [CrossRef]
  15. Zhang, H.; Huang, Q.; Zhai, H.; Zhang, L. Multi-temporal cloud detection based on robust PCA for optical remote sensing imagery. Comput. Electron. Agric. 2021, 188, 106342. [Google Scholar] [CrossRef]
  16. Li, Y.; Wu, L.; Han, Q.; Wang, X.; Zou, T.; Fan, C. Estimation of remote sensing based ecological index along the Grand Canal based on PCA-AHP-TOPSIS methodology. Ecol. Indic. 2021, 122, 107214. [Google Scholar] [CrossRef]
  17. He, T.; Sun, Y.J.; Xu, J.D.; Wang, X.J.; Hu, C.R. Enhanced land use/cover classification using support vector machines and fuzzy k-means clustering algorithms. J. Appl. Remote Sens. 2014, 8, 083636. [Google Scholar] [CrossRef]
  18. Yuan, Q.; Shen, H.; Li, T.; Li, Z.; Li, S.; Jiang, Y.; Xu, H.; Tan, W.; Yang, Q.; Wang, J.; et al. Deep learning in environmental remote sensing: Achievements and challenges. Remote Sens. Environ. 2020, 241, 111716. [Google Scholar] [CrossRef]
  19. Weijia, L.; Haohuan, F.; Le, Y.; Arthur, C. Deep Learning Based Oil Palm Tree Detection and Counting for High-Resolution Remote Sensing Images. Remote Sens. 2016, 9, 22. [Google Scholar]
  20. Mubin, N.A.; Nadarajoo, E.; Shafri, H.Z.M.; Hamedianfar, A. Young and mature oil palm tree detection and counting using convolutional neural network deep learning method. Int. J. Remote. Sens. 2019, 40, 7500–7515. [Google Scholar] [CrossRef]
  21. Freudenberg, M.; Nölke, N.; Agostini, A.; Urban, K.; Wörgötter, F.; Kleinn, C. Large Scale Palm Tree Detection in High Resolution Satellite Images Using U-Net. Remote Sens. 2019, 11, 312. [Google Scholar] [CrossRef]
  22. Hagolle, O.; Huc, M.; Pascual, D.V.; Dedieu, G. A multi-temporal method for cloud detection, applied to FORMOSAT-2, VENµS, LANDSAT and SENTINEL-2 images. Remote Sens. Environ. 2010, 114, 1747–1755. [Google Scholar] [CrossRef] [Green Version]
  23. Rabiei, S.; Jalilvand, E.; Tajrishy, M. A Method to Estimate Surface Soil Moisture and Map the Irrigated Cropland Area Using Sentinel-1 and Sentinel-2 Data. Sustainability 2021, 13, 11355. [Google Scholar] [CrossRef]
  24. Neumann, M.; Pinto, A.S.; Zhai, X.; Houlsby, N. In-domain representation learning for remote sensing. arXiv 2019, arXiv:1911.06721. [Google Scholar]
  25. Gong, C.; Han, J.; Lu, X. Remote Sensing Image Scene Classification: Benchmark and State of the Art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar]
  26. He, T.; Mu, S.; Zhou, H.; Hu, J. Wood species identification based on an ensemble of deep convolution neural networks. Wood Res. 2021, 66, 1–14. [Google Scholar]
  27. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef]
  28. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  29. Huang, G.; Liu, Z.; Laurens, V.; Weinberger, K.Q. Densely Connected Convolutional Networks; IEEE Computer Society: Washington, DC, USA, 2016. [Google Scholar]
  30. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  31. Tan, M.; Le, Q. Efficientnetv2: Smaller models and faster training. arXiv 2021, arXiv:2104.00298. [Google Scholar]
  32. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  33. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  34. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  35. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design; Springer: Cham, Switzerland, 2018. [Google Scholar]
  36. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2012; p. 25. [Google Scholar]
  37. Rifai, S.; Glorot, X.; Bengio, Y.; Vincent, P. Adding noise to the input of a model trained with a regularized objective. arXiv 2011, arXiv:1104.3250. [Google Scholar]
  38. Lewis, H.; Brown, M. A generalized confusion matrix for assessing area estimates from remotely sensed data. Int. J. Remote Sens. 2001, 22, 3223–3235. [Google Scholar] [CrossRef]
  39. Ye, Y.; Zhou, H.; Yu, H.; Hu, R.; Zhang, G.; Hu, J.; He, T. An Improved EfficientNetV2 Model Based on Visual Attention Mechanism: Application to Identification of Cassava Disease. Comput. Intell. Neurosci. 2022, 2022, 1569911. [Google Scholar] [CrossRef] [PubMed]
  40. Prabhakar, T.; Kora, P.; Sudhakar, A. Nearest Neighbor Classification of Remote Sensing Images with the Statistical Features and PCA-Based Features. In Evolution in Signal Processing and Telecommunication Networks; Springer: Berlin/Heidelberg, Germany, 2022; pp. 491–502. [Google Scholar]
  41. Ju, Y.; Bohrer, G. Classification of Wetland Vegetation Based on NDVI Time Series from the HLS Dataset. Remote Sens. 2022, 14, 2107. [Google Scholar] [CrossRef]
  42. Minowa, Y.; Kubota, Y.; Nakatsukasa, S. Verification of a Deep Learning-Based Tree Species Identification Model Using Images of Broadleaf and Coniferous Tree Leaves. Forests 2022, 13, 943. [Google Scholar] [CrossRef]
  43. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  44. Illarionova, S.; Trekin, A.; Ignatiev, V.; Oseledets, I. Tree Species Mapping on Sentinel-2 Satellite Imagery with Weakly Supervised Classification and Object-Wise Sampling. Forests 2021, 12, 1413. [Google Scholar] [CrossRef]
Figure 1. The overall workflow of the research.
Figure 1. The overall workflow of the research.
Sustainability 15 02741 g001
Figure 2. Location map of the study area showing the overview map of China and Zhejiang Province.
Figure 2. Location map of the study area showing the overview map of China and Zhejiang Province.
Sustainability 15 02741 g002
Figure 3. An example of clipping Sentinel-2 images by using GGC of Pinus massoniana Lamb.
Figure 3. An example of clipping Sentinel-2 images by using GGC of Pinus massoniana Lamb.
Sustainability 15 02741 g003
Figure 4. Examples for removing low-quality images from FTSD: (a) near the residential areas; (b) clouds covered; (c) large naked area.
Figure 4. Examples for removing low-quality images from FTSD: (a) near the residential areas; (b) clouds covered; (c) large naked area.
Sustainability 15 02741 g004
Figure 5. Validation accuracy of eight deep learning algorithms with PCANDVI in FTSD.
Figure 5. Validation accuracy of eight deep learning algorithms with PCANDVI in FTSD.
Sustainability 15 02741 g005
Figure 6. Validation accuracy of eight deep learning algorithms in RESISC-45.
Figure 6. Validation accuracy of eight deep learning algorithms in RESISC-45.
Sustainability 15 02741 g006
Figure 7. The confusion matrix of ResNet50 in FTSD.
Figure 7. The confusion matrix of ResNet50 in FTSD.
Sustainability 15 02741 g007
Table 1. The Parameters of Sentinel-2.
Table 1. The Parameters of Sentinel-2.
Sentinel-2 BandsCentral Wavelength (nm)Resolution (m)
Band 1: coastal aerosol0.44360
Band 2: blue0.49010
Band 3: green0.56010
Band 4: red0.66510
Band 5: vegetation red edge0.70520
Band 6: vegetation red edge0.74020
Band 7: vegetation red edge0.78320
Band 8: NIR0.84210
Band 8A: vegetation red edge0.86520
Band 9: water vapor0.94560
Band 10: SWIR-cirrus1.37560
Band 11: SWIR1.61020
Band 12: SWIR2.19020
Table 2. Different band combinations.
Table 2. Different band combinations.
CombinationRedGreenBlue
432Band 4Band 3Band 2
843Band 8Band 4Band 3
PCAPC1PC2PC3
PCANDVIPC1PC2NDVI
Table 3. Tree species selected in inventory.
Table 3. Tree species selected in inventory.
Dominant Tree SpeciesArea (Hectares)SamplesImages
Bambusoideae>13223935
Cunninghamia lanceolata>132251000
Mixed broadleaf>12205945
Mixed broadleaf-conifer>162371080
Mixed conifer>122271035
Other hardwood>13216985
Pinus massoniana Lamb.>8208880
Pinus taiwanensis>10217930
Quercus spp.>162121025
Table 4. The best validation accuracy of different models in different band combinations after 100 epochs of training.
Table 4. The best validation accuracy of different models in different band combinations after 100 epochs of training.
ModelsBest Validation Accuracy (%)
432843PCAPCANDVI
DenseNet12125.8726.5526.0933.18
EfficientNetv2_s25.7525.3525.3041.69
MobileNet_v3_large21.5021.9521.1636.07
RegNety_400mf23.0321.1521.7223.65
ResNet3449.2964.0468.1771.64
ResNet5040.5047.9958.2564.89
ResNet10135.2843.2851.0553.26
ShuffleNet_v2_x1_040.4438.2351.2858.88
Table 5. Precision and recall of ResNet50 in FTSD.
Table 5. Precision and recall of ResNet50 in FTSD.
Dominant Tree SpeciesPrecisionRecall
Bambusoideae0.9580.968
Cunninghamia lanceolata0.9290.845
Mixed broadleaf0.7890.852
Mixed broadleaf-conifer0.8030.847
Mixed conifer0.8670.787
Other hardwood0.8590.832
Pinus massoniana Lamb.0.9000.818
Pinus taiwanensis0.7560.866
Quercus spp.0.8220.834
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

He, T.; Zhou, H.; Xu, C.; Hu, J.; Xue, X.; Xu, L.; Lou, X.; Zeng, K.; Wang, Q. Deep Learning in Forest Tree Species Classification Using Sentinel-2 on Google Earth Engine: A Case Study of Qingyuan County. Sustainability 2023, 15, 2741. https://doi.org/10.3390/su15032741

AMA Style

He T, Zhou H, Xu C, Hu J, Xue X, Xu L, Lou X, Zeng K, Wang Q. Deep Learning in Forest Tree Species Classification Using Sentinel-2 on Google Earth Engine: A Case Study of Qingyuan County. Sustainability. 2023; 15(3):2741. https://doi.org/10.3390/su15032741

Chicago/Turabian Style

He, Tao, Houkui Zhou, Caiyao Xu, Junguo Hu, Xingyu Xue, Liuchang Xu, Xiongwei Lou, Kai Zeng, and Qun Wang. 2023. "Deep Learning in Forest Tree Species Classification Using Sentinel-2 on Google Earth Engine: A Case Study of Qingyuan County" Sustainability 15, no. 3: 2741. https://doi.org/10.3390/su15032741

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop