Next Article in Journal
Changes in Community Composition of Tropical Evergreen Forests during Succession in Ta Dung National Park, Central Highlands of Vietnam
Next Article in Special Issue
Using U-Net-Like Deep Convolutional Neural Networks for Precise Tree Recognition in Very High Resolution RGB (Red, Green, Blue) Satellite Images
Previous Article in Journal
Quantitative Genetic Variation in Bark Stripping of Pinus radiata
Previous Article in Special Issue
Application of LiDAR Data for the Modeling of Solar Radiation in Forest Artificial Gaps—A Case Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature-Level Fusion between Gaofen-5 and Sentinel-1A Data for Tea Plantation Mapping

School of Earth Sciences and Resources, China University of Geosciences (Beijing), Beijing 100083, China
*
Author to whom correspondence should be addressed.
Forests 2020, 11(12), 1357; https://doi.org/10.3390/f11121357
Submission received: 3 November 2020 / Revised: 3 December 2020 / Accepted: 16 December 2020 / Published: 18 December 2020
(This article belongs to the Special Issue Remote Sensing Applications in Forests Inventory and Management)

Abstract

:
The accurate mapping of tea plantations is significant for government decision-making and environmental protection of tea-producing regions. Hyperspectral and Synthetic Aperture Radar (SAR) data have recently been widely used in land cover classification, but effective integration of these data for tea plantation mapping requires further study. This study developed a new feature-level image fusion method called LPPSubFus that combines locality preserving projection and subspace fusion (SubFus) to map tea plantations. Based on hyperspectral and SAR data, we first extracted spectral indexes, textures, and backscattering information. Second, this study applied LPPSubFus to tea plantation mapping with different classification algorithms. Finally, we compared the performance of LPPSubFus, SubFus, and pixel-level image fusion in tea plantation mapping. Feature-level image fusion performed better than pixel-level image fusion. An improvement of about 3% was achieved using feature-level image fusion compared to hyperspectral data alone. Regarding feature-level image fusion, LPPSubFus improved the overall accuracy by more than 3% compared to SubFus. In particular, LPPSubFus using neural network algorithms achieved the highest overall accuracy (95%) and over 90% producer and user accuracy for tea plantations and forests. In addition, LPPSubFus was more compatible with different classification algorithms than SubFus. Based on these findings, it is concluded that LPPSubFus has better and more stable performance in tea plantation mapping than pixel-level image fusion and SubFus. This study demonstrates the potential of integrating hyperspectral and SAR data via LPPSubFus for mapping tea plantations. Our work offers a promising tea plantation mapping method and contributes to the understanding of hyperspectral and SAR data fusion.

1. Introduction

The tea plant (Camellia sinensis (L.) O. Kuntze), an evergreen broadleaved perennial shrub, originated in China, and tea has gradually become one of the three most popular beverages in the world [1,2]. Tea is widely distributed globally. China, India, and Sri Lanka are the major tea-producing countries in Asia, and Kenya, Malawi, Rwanda, Tanzania, and Uganda are the major tea-producers in Africa [3]. Tea planting and production play a vital role in the agricultural economy and rural development of these developing countries. Driven by the considerable economic benefits of tea cultivation, the area of tea plantations is continually increasing [4]. According to data of the Food and Agriculture Organization (FAO) of the United Nations, the global area of tea plantations reached 4,193,176 hectares in 2018, and the tea plantation area increased by 77.33% between 2000 and 2018 [5]. In China, a global leader in tea cultivation and production, tea is mainly grown in the red soil hilly areas of the southern provinces, such as Fujian and Zhejiang. In 2018, China’s tea plantations occupied an area of 2,347,726 hectares, increasing 161% since 2000 [5]. Although the expansion of tea plantations can promote local economic development, it triggers a series of environmental problems, reducing soil fertility, and contributing to soil erosion [6]. Therefore, accurately investigating tea plantation distribution is conducive to government decision-making and environmental protection.
Remote sensing is a cost-effective and labor-saving technology for tea plantation mapping. The application of different remote sensing data to mapping tea plantations can be divided into three periods. In the early stage, multispectral data was dominant. Akar and Güngör [7] extracted texture features and the normalized vegetation index based on WorldView-2 data to identify tea plantations in northwestern Turkey with an overall accuracy (OA) of 87.89%. However, the ability of multispectral data to address spectral confusion is often limited, thus, the use of multi-source data subsequently became popular. For instance, Chu et al. [8] fused hyperspectral (HS) data with light detection and ranging (LiDAR) data and used support vector machine (SVM) algorithms to extract the distribution of tea plantations, and achieved an OA of 91%. In recent years, time-series data has been gradually applied to mapping of tea plantations. Li et al. [4] proposed mapping tea plantations based on the phenological characteristics from Vegetation and Environmental New micro Spacecraft (VENμS) images and achieved an OA of 95%. However, tea plantations are widespread in subtropical areas [9], for which it is challenging to obtain high-quality time-series data due to clouds and weather [10,11]. Consequently, multi-source data continue to play a significant role in mapping tea plantations. Because of the high cost of acquiring LiDAR data [12], it is necessary to find new data sources.
HS data are helpful for vegetation identification because of their abundant spectral information [13]. Synthetic Aperture Radar (SAR) data can provide extra information on vegetation structure, such as trunks and branches [9], and SAR data are more easily obtained than LiDAR data. Due to the advantages of HS and SAR data, their synergistic use has been prevalent. Hu et al. [14] proposed an object-based fusion approach for the joint use of HS and SAR data, and the results suggested that synergistic use can achieve better performance than that of a single data source. Jouan and Allard [15] found that backscattering can provide information on vegetation coverage and water content, and the combination of HS and SAR data can improve the description of the land cover. Therefore, the integration of HS and SAR data contributes to information extraction, which is a promising method for tea plantation mapping. However, extracting valuable information about tea from numerous HS bands and effectively combining HS and SAR data for tea plantation mapping requires further study.
Previous studies have mainly used image fusion to integrate HS and SAR data. Image fusion combines two or more images into a new image using a specific algorithm [16,17]. Using information from different sensors can be more accurate than that from a single sensor in land use/land cover classification and vegetation mapping [18,19,20,21,22,23]. However, the ability of image fusion of optical and SAR data to promote classification depends on the fusion method [24]. The most commonly used image fusion methods include pixel- and feature-level image fusion. Feature-level image fusion (FLIF) first extracts series of feature information from the original image and then combines them into a feature map, making more accurate decisions [25]. Compared to pixel-level image fusion (PLIF), FLIF has a lower computing environment and less stringent registration. Moreover, FLIF can achieve a considerable compression of information, contributing to the dimensionality reduction in high-dimensional data. Hu et al. [26] used manifold learning to fuse morphological and texture features from HS and SAR data for land cover classification, and the accuracy was improved by 1%–7% after FLIF. Rasti and Ghamisi [27] proposed a subspace fusion (SubFus) method to integrate morphological features from remote sensing data for land cover classification and obtained 85.32% and 99.25% in OA on the Houston and Trento datasets, respectively. Hence, FLIF is a powerful method to extract information from multi-source data. However, few studies have focused on FLIF in tea plantation mapping, and its ability is still unclear.
Therefore, this study extracted features from HS and SAR data and used FLIF to map tea plantations. The objectives of this study included: (1) seeking new representative features of tea from HS data; (2) developing an FLIF method that integrated HS and SAR data, and then applying it to mapping tea plantations; and (3) comparing the performance of different fusion methods for HS and SAR data in tea plantation mapping.

2. Materials and Methods

2.1. Study Area

This study was carried out in Wuyishan City, Nanping City, Fujian Province, China (Figure 1). Wuyishan city is a famous tea-producing area in China. It has a subtropical monsoon climate, with an annual average temperature of 13 °C, annual rainfall of 2000 mm. The area is suitable for tea growth because of sufficient sunshine, a mild climate, suitable geological background, and abundant rainfall. The tea industry is an important industrial pillar of Wuyishan City. In recent years, the area of tea plantations in Wuyishan City has increased rapidly. Because the tea plantations are mainly concentrated in the southwestern area of Wuyishan City, we selected the region (the red polygon in Figure 1) covering an area of 200 km2 to carry out our tea plantation mapping study. Figure 1a,b show the color composite images of HS and SAR data of the study area under the ArcGIS 10.2 software platform.

2.2. Data Source

2.2.1. Field Survey

Field surveys were conducted in June and December 2019. The field survey results showed that, in addition to the widespread distribution of tea plantations, evergreen forests, croplands, waterbodies, built-up areas, and bareness rocks were distributed throughout the study area (Figure 2). Trees were unevenly scattered in some tea plantations (Figure 2a–c). The winter and summer crops in the study area are shown in Figure 2d,e, and tea plants were mixed with other crops in some flat areas in summer (Figure 2f), making it challenging to analyze the distribution of tea plants using remote sensing images. The built-up area, waterbody, and rock are shown in Figure 2g–i.

2.2.2. Remote Sensing and Auxiliary Data

The HS data used in this study was acquired by a Chinese Gaofen-5 satellite, which covers the spectral ranges from 400 to 2500 nm with 30 m spatial resolution. The SAR data was acquired by Sentinal-1A satellites that were operated at a frequency of 5405 GHz and collected in Interferometric Wide Swath mode, which acquires data with a 250 km swath width at 5 × 20 m spatial resolution, and vertical transmit and horizontal receive (VH) and vertical transmit and vertical receive (VV) polarization. These two datasets were collected in November because other crops in the study area had been harvested by this time, which helped distinguish the tea plantations. The detailed information of HS and SAR data used in this study is provided in Table 1. Auxiliary data included land cover/land use map of the study area and a 30 m digital elevation model (DEM).

2.3. Data Preprocessing

2.3.1. Hyperspectral Data Preprocessing

First, two water vapor absorption regions (25 bands), at 1350–1426 and 1798–1948 nm, in the short wave infrared band, were removed from the HS data. Second, the first four short wave infrared bands were removed for the spectral overlap between the visible near-infrared and short wave infrared band. Subsequently, 301 bands were chosen for radiometric calibration, Fast Line-of-sight Atmospheric Analysis of Spectral Hypercubes atmospheric correction, Savitzky–Golay filtering, and orthographic correction. Regarding Savitzky–Golay filtering, the center point of the filtering core was 8, the derivative order was 0, and the degree of the polynomial was 4. Finally, we performed visual inspection and removed 47 bands due to stripe noise. Hence, 254 bands were outputted to conduct the following study.
Terrain correction was performed for the HS image via the C-correction model based on DEM data to reduce the terrain effect. C-correction is a useful method for normalizing topographic illumination effects [28]. This method introduces an empirically derived parameter to explain the overcorrection of the simpler cosine method in low illumination areas [29].
L H   =   L T cos ( s z ) +   c cos ( i ) +   c
where L H is the corrected image value, L T is the uncorrected image value, s z is the solar zenith angle, i is the solar incident angle concerning the surface normal direction, and c is the C-correction coefficient derived by regressing L T against cos ( i ) and taking the quotient of the intercept and slope.

2.3.2. SAR Data Preprocessing

The Single Look Complex data of VV and VH polarization were first multilooked with a range number of two looks in the azimuth and eight in the range, resulting in a ground-range resolution of 29.52 × 27.87 m. The speckle noise was then removed using the refined Lee filter with a 5 × 5 window. The VV and HH images were then geocoded to the World Geodetic System 1984 datum and Universal Transverse Mercator Zone 50 North coordinate system via Range-Doppler terrain correction using the DEM data. Finally, VV and VH polarized data were resampled to a 30 m spatial resolution and georeferenced with the HS data using 20 ground control points with an error of less than 0.5 pixels. After data preprocessing, we obtained processed HS and SAR data with a size of 587 × 377 pixels and a spatial resolution of 30 m.

2.4. Feature Extraction

This study extracted four sets of features from HS and SAR data, namely, principal component bands, spectral indexes, polarimetric features, and texture information. These features were extracted based on the Environment for Visualizing Images (ENVI) software program.

2.4.1. Principal Component Analysis

Principal component analysis (PCA) is a commonly used dimensionality reduction method. PCA transforms a set of variables that may be correlated into fewer variables that are linearly independent using orthogonal transformation [30]. We performed PCA on the HS data. The first three principal components, which accounted for most of the data variance, were extracted to participate in the classification as spectral variables.

2.4.2. Spectral Index

Spectral characteristics of six land cover types are shown in Figure 3. First, six bands with wavelengths 792, 865, 1006, 1173, 1266, and 1662 nm were selected because of the big difference between tea plantation and forest in these bands. Second, four vegetation indexes that reflect the canopy structure, pigmentation, and water content of vegetation canopy were chosen to improve the spectral separability of tea (Table 2), namely, the Chlorophyll Index 2 [31], Anthocyanin Reflectance Index 2 [32], Vogelmann Red Edge Index 2 [33], and Water Band Index [34].
We further analyzed the spectral reflectance curves to fully use the rich spectral characteristics of HS data. We found that the variation tendencies of tea and forest spectral curves are similar, with the reflectance of tea higher than that of the forest, especially at wavelengths 800–1600 nm. For example, the spectral reflectance values of wavelengths 865 nm are different, whereas spectral reflectance values of 454 nm are the same, and thus the slope of the line between wavelengths 865 and 454 nm are different (the black line in Figure 3). Moreover, the triangle area enclosed by the bands with wavelengths 454, 1006, and 2032 nm is different between tea and forest (the blue polygon in Figure 3). Therefore, constructing new spectral indexes for the slope and area-dependent may improve tea and forest separability. We created three tea vegetation indexes (TVIs) based on the spectral slope between wavelengths 454 and 865 nm, 1266 and 2032 nm, and the triangle area enclosed by wavelengths 454, 1006, and 2032 nm. The equations of the slopes between wavelength 454 and 865, and 1266 and 2032 nm, are TVI 1 and TVI 2 , respectively:
TVI 1   =   ( ρ 865   nm ρ 454   nm ) / Δ λ 1
TVI 2   =   ( ρ 2032   nm ρ 1266   nm ) / Δ λ 2
where ρ is the spectral reflectance value of the corresponding band, Δ λ 1 is the difference in wavelength between 454 and 865 nm, and Δ λ 2 is the difference in wavelength between 1266 and 2032 nm.
The equation of the triangle area enclosed by 454, 1006, and 2032 nm is TVI 3 :
TVI 3   =   ( ρ 1006   nm ρ 454   nm ) × Δ ω 3 { ( ρ 1006   nm ρ 454   nm )   ×   Δ ω 1 + ( ρ 1006   nm ρ 2032   nm )   ×   Δ ω 2 + ( ρ 2032   nm ρ 454   nm )   ×   Δ ω 3 } / 2
where Δ ω 1 is the difference in wavelength between 454 and 1006 nm, Δ ω 2 is the difference in wavelength between 1006 and 2032 nm, and Δ ω 3 is the difference in wavelength between 454 and 2032 nm.
The results of the three TVIs are shown in Figure 4. The tea plantations showed a high value in TVI 1 and TVI 3 images, but a low value in TVI 2 images. The value of other land cover types on the TVI images was opposite to that of the tea plantation. Particularly in TVI 1 and TVI 3 images, the difference between tea plantation and forest was evident (the red polygon in Figure 4). These three TVIs all play a role in increasing the difference between tea plantations and other land cover types, thus improving classification accuracy.

2.4.3. Polarimetric Feature

SAR data have the potential for land cover classification because they are not affected by clouds and weather [10,35,36,37]. In addition, SAR data can reflect the geometric and dielectric properties of scattering, improving classification accuracy [38,39]. In this study, four polarimetric features of Sentinel-1A data were extracted. They were the VH and VV intensity, the difference between VV and VH, and the VV and VH intensity ratio (Figure 5).

2.4.4. Spatial Feature

Image texture was defined as the spatial arrangement and variation frequency of the value levels of an adjacent set of pixels in the local neighborhood of an image [21]. Adding texture information contributes to classification accuracy when mapping tea plantations [7]. Hence, the Grey-Level Co-occurrence Matrix textures (contrast, dissimilarity, homogeneity, mean, variance, and correlation) [40] were then sequentially calculated via the first three principal components and four polarimetric features using a window of 3 × 3 with a direction of 45° and a spatial distance of 1 pixel. Therefore, 18 and 24 features were extracted from the HS and the SAR data, respectively.

2.5. Image Fusion

2.5.1. Subspace Fusion (SubFus)

The SubFus method proposed by Rasti and Ghamisi [27] is an innovative FLIF method to integrate diverse remote sensing data for land cover classification. This method proposes a novel subspace minimization with its derived solution to fuse features, making it applicable to various sensors with different characteristics. Moreover, SubFus can be easily extended to other applications in addition to remote sensing. The principle of SubFus is as follows:
Let H and S be the data from HS and SAR sensors, respectively. The features obtained from HS and SAR data can be formulated as:
H   =   V ( H ) ,   S   =   V ( S )
where matrices H ( n × p 1 ) and S ( n × p 2 ) contain the vectorized features for HS and SAR data, respectively, in their i-th columns. Then, this method assumes that multi-source data can be modeled using the fused features (F) in different subspaces. The extracted features contain redundant information, and therefore the multi-source features should have low-rank representations. Accordingly, the features from H and S can be modeled in lower-dimensional space as:
H   =   FV 1 T   +   N 1 ,   S   =   FV 2 T   +   N 2
N 1 and N 2 are n × p matrices representing the noise and model error, F is an ( n × r ) matrix containing the fused features, and V 1 and V 2 are ( p 1 × r and p 2 × r , respectively) subspace projection matrices. For the standard subspace rank r, there is a lower bound for the rank selection. This method suggests the minimum possible rank, i.e., r min ( n , p 1 , p 2 ) , to achieve the maximum reduction in the feature space.
Then the constrain penalized cost function to estimate the unknown matrices F simultaneously, V 1 , and V 2 as
( F ^ , V ^ 1 , V ^ 2 )   = arg   min F , V 1 , V 2   J ( F , V 1 , V 2 )   s . t .   V 1 T V 1   =   I   and   V 2 T V 2   =   I
where:
J ( F , V 1 , V 2 )   =   1 2 H F V 1 T F 2   +   λ 1 2 S F V 2 T F 2   +   λ 2 F TV
In Equation (8),   λ 1 and λ 2 control the weight between the fidelity of SAR data and the smoothness of the subspace features, respectively, with respect to the fidelity of HS data. The constraints enforce the orthogonality on the subspaces. The alternating direction method of multipliers is used to solve Equation (8). The variable F is divided by:
( M ^ , F ^ , V ^ 1 , V ^ 2 )   =   arg   min M ,   F , V 1 , V 2 1 2 H F V 1 T F 2   +   λ 1 2 S F V 2 T F 2   +   λ 2 M TV s . t .   F   =   M ,   V 1 T V 1   =   I   and   V 2 T V 2   =   I
Then, the penalty method is used to add the constraint F   =   M to the minimization problem. Therefore, the augmented minimization problem can be represented as:
( M ^ , F ^ , V ^ 1 , V ^ 2 )   =   arg   min M ,   F , V 1 , V 2 1 2 H F V 1 T F 2   +   λ 1 2 S F V 2 T F 2   +   λ 2 M TV   +   μ 2 M F B F 2 s . t .   V 1 T V 1   =   I   and   V 2 T V 2   =   I
where B is the Lagrangian multiplier. After the variables are split, the cyclic descent algorithm is adopted to solve one unknown matrix at a time, using five iterative steps. The detailed steps are not repeated here.
It should be noted that the right singular vectors of H and S are used to initialize V 1 and V 2 , respectively, and M 0 and B 0 are set to zero in SubFus. However, the singular eigenvector computed from the original data matrix may ignore the essence of the data, which may not be the best representation of the original data. Therefore, a more appropriate method should be explored to initialize V 1 and V 2 .

2.5.2. Combination of Locality Preserving Projection and SubFus

Locality preserving projection (LPP), a manifold learning algorithm, was introduced to improve the SubFus, and we called this new method LPPSubFus. The basic idea of LPP is to maintain the close and far affinity relationship between each sample pair in the projection and retain the local neighborhood structure of the samples in the space when reducing the dimensionality [41]. That is, LPP aims to optimize projection P which converts the original data X to a lower-dimensional representation Y = P T X [42]. The objective of this study is to obtain the projection P that is used to replace the singular vector to initialize V 1 and V 2 . The LPP introduced into the SubFus has two advantages: (1) the data dimensionality can be initially reduced before SubFus, and (2) LPP can better reflect the essence of data than the singular vector. The principle of solving projection P is as follows [43]:
X = [ x 1 , x 2 , , x m ] is set to the original data and lower-dimensional representation of Y = [ y 1 , y 2 , , y m ] is a m   ×   d matrix. A reasonable criterion for deciding Y is to minimize the following objective function:
L = i j ( y i y j ) 2 W i j
The parameter W is obtained by the k-nearest-neighborhood and induces a punishment when the neighbor area x i is far away from x j . The formula of W is as follows:
W i j = { e x i x j 2 σ    x i   and   x j   are   local   neighbors 0      x i   and   x j   are   not   local   neighbors
where σ is a filtering parameter that can be determined empirically.
Suppose p is a transformation vector, that is, Y = p T X , by simple algebraic formulation, the optimization is formulated as follows:
1 2 i j ( y i y j ) 2 W i j = 1 2 i j ( p T x i p T x j ) 2 W i j = i p T x i D i i x i T p i j p T x i W i j x j T p = p T X ( D W ) X T p = p T X L X T p
where D is a diagonal matrix; its entries are the column (or row, because W is symmetric) sum of W. D i i = j W i j . A Laplacian matrix L = D W is constructed. The bigger D i i obtains a more important y i . Therefore, the constraint is imposed as follows:
Y T D Y = 1 p T X D X T p = 1 a r g   min p   p T X L X T p
Subject   to   p T X D X T p = 1
The transformation vector p that minimizes the objective functions is given by the solution of the minimum eigenvalue to the generalized eigenvalues problem:
X L X T p = λ X D X T p
It is easy to show that the matrices X L X T and X D X T are symmetric and positive semidefinite. Let the column vectors p 0 , p 1 , p d 1 be the solutions of Equation (15), ordered according to their eigenvalues, λ 0 , λ 1 , λ d 1 . Thus, the embedding is as follows:
Y   =   P T X ,   P   =   [ p 0 ,   p 1 ,   ,   p d - 1 ] T
The projection vector P H and P S of HS and SAR features can be obtained, respectively. Then, they are used to initialize V 1 and V 2 in SubFus.
In this study, SubFus and LPPSubFus were implemented in the MATLAB software program. Regarding SubFus, λ 1 ,   λ 2 , and μ were set to 0.5, 0.1, and 10, respectively. Then, in LPPSubFus, k was set to 5, and σ was set to 5. The remaining parameter values were the same as those in SubFus.

2.5.3. Pixel-Level Image Fusion

Three PLIF methods were also used in this study to fuse HS and SAR data under the MATLAB software program. The PLIF methods mainly include component substitution, multi-scale decomposition, and the hybrid method [25]. Therefore, we chose PCA [23], discrete wavelet transforms (DWT) [22], and wavelet-principal component analysis (W-PCA) [21] to represent these three types. In this study, six bands of HS data were fused with VV polarized image because the co-polarization SAR data often performs better than cross-polarization in vegetation identification [44,45].

2.6. Image Classification

2.6.1. Classification Algorithm

The machine learning algorithms show good performance in remote sensing applications [24,46,47,48]. This study selected two typical and popular machine learning algorithms, namely SVM and neural network (NN) algorithms, to carry out classifications.
The SVM algorithm aims to define an optimum hyperplane to separate two different classes using training data without any assumptions [49]. Although SVM has limitations in parameter selection and computational requirements, it provides superior performance compared to most other classification algorithms [50]. Many studies have used the SVM method to extract tea plantation distribution and achieved good results [8,9,51]. Moreover, Melgani, and Bruzzone [52] demonstrated that SVM has good performance in HS data classification because it is less sensitive to the Hughes phenomenon. The major parameters of SVM algorithms include kernel function, gamma value, and penalty parameter.
The NN algorithm is a computing model composed of essential nodes, including the input layer, hidden layer, and output layer [53]. This method is competent for parallel computing, automatic learning, and correcting errors [54], and has long been used for remote sensing applications, such as land cover classification [55], wetland vegetation mapping [56], and tree species identification [54]. Many have suggested that NN algorithms are superior to traditional statistical classification algorithms because they learn from training samples without assumptions about data distribution [48]. The major parameters of NN algorithms include training threshold contribution, training rate, training momentum, training root mean square error exit criteria, number of hidden layers, and training iteration times.

2.6.2. Classification Scenario and Sample

Six classification scenarios were developed to compare the performance of different image fusion methods in this study (Table 3). Scenario 1 used only the HS data, scenarios 2–4 used the data fused by PLFI, and scenarios 5 and 6 fused features of HS and SAR data using SubFus and LPPSubFus, respectively. In addition, all corresponding features in each classification scenario were normalized to weaken the impact of different attributes of features.
In this study, the built-up area and waterbody were classified into one category. Therefore, five categories needed to be classified. The training and test samples for the five categories were randomly chosen from the field survey results and land cover/land use map (Table 4). The ratio of training samples to test samples was 1 to 2, and all samples were distributed evenly throughout the study area. The sample number of tea plantations and forests were two times as many as other land cover types, which ensured the accuracy assessment result mainly depended on tea plantations and forests.

2.6.3. Parameter Setting and Accuracy Assessment

The classifications were carried out using the ENVI software program. Regarding the SVM-based classifications, the radial basis function was selected as the kernel function because it is suitable for non-linear data [57]. The gamma value was dependent on the inverse of the number of input variables [9], and the default value (100) of the penalty parameter was used for all classifications. For all NN-based classifications, the logistic function was selected as the activation function, the training threshold contribution was set to 0.9, the different learning rate was set to 0.2, training root mean square error was set to 0.1, and momentum value was set to 0.9. The number of nodes in the hidden layer and the number of training iterations were set to 1 and 1000, respectively. After classifications, an accuracy assessment was also performed, including the traditional confusion matrix, Kappa coefficient, producer accuracy (PA), and user accuracy (UA).

3. Results

3.1. Visual Comparison of Thematic Maps

The classifications of different scenarios using SVM and NN algorithms are shown in Figure 6, where scenarios 5 and 6 yielded the highest OA in SubFus- and LPPSubFus-based classifications (scenario 5: the feature number using SVM and NN algorithms was 27 and 12, respectively; scenario 6: the feature number using SVM and NN algorithms was 28 and 22, respectively). In both SVM- and NN-based classifications, scenario 1 overestimated the tea plantation classification results, and some forests were misclassified as tea plantations (Figure 6a,g). The distribution of tea plantations was reduced in the classifications of scenarios 2–4 (Figure 6b–d,h–j). However, the classifications using scenarios 2–4 retained speckle noise, and these classifications were unsatisfactory. Scenarios 5 and 6 improved the classification with little speckle noise, consistent with the land cover distribution (Figure 6e,f,k,l). Moreover, scenario 6 more accurately classified forests in the central and southwestern parts of the study area than other scenarios.
Some rocks in the upper right part of the study area were incorrectly classified as waterbodies in the classifications of scenarios 2–4 due to topography shadow, whereas the rocks were correctly classified by scenarios 5–6. However, the waterbodies in the classification of scenario 6 using the NN algorithms were not well identified (Figure 6l). By comparison, because the primary cultivated land type in the study area is the paddy field, there was confusion between waterbodies and croplands in the classifications of some scenarios, such as in scenario 4 (Figure 6d,j).
The classifications using only HS data with SVM and NN algorithms had considerable speckle noise. Image fusion improved the noise phenomenon, especially in the FLIF-based classifications. In addition, the classifications using NN algorithms were better than those using SVM algorithms. For example, the distribution of tea plantations and forests was similar in the classifications of scenarios 5 and 6. However, some croplands were mistakenly identified as waterbodies and had evident shadows when using the SVM algorithms (Figure 6e,f,k,l). In terms of visual comparisons, all the scenarios using both SVM and NN algorithms could identify the land cover types well in the study area. Among these, the classifications of scenario 1 had considerable speckle noise, and classifications of scenarios 2–4 were similar. Scenarios 5 and 6 achieved better results than other scenarios.

3.2. Comparison of Classification Accuracy

3.2.1. Comparison of Overall Accuracy

The OA and Kappa coefficient of the six scenarios are summarized in Table 5. In SVM-based classifications, scenario 1 achieved an OA of 89.05%. There was an improvement of OA when using scenarios 2 and 4. However, the OA of the classifications with scenario 3 was 1.31% less than that with scenario 1. Scenario 5 achieved the same OA as scenario 1, and the Kappa coefficient was slightly lower than that of scenario 1. Scenario 6 achieved the highest OA of 92.14% in SVM-based classifications. In NN-based classifications, scenario 1 yielded a 92.14% OA, then the OA of scenarios 2–5 decreased by 1.31%–6.07% compared to scenario 1. Scenario 6 achieved the highest OA of 95% and the highest Kappa coefficient of 0.9355 in NN-based classifications, and the OA with scenario 6 was increased by 2.86% over that with scenario 1.
Figure 7 shows the OA curves of SubFus and LPPSubFus with the feature number. For SubFus-based classifications (Figure 7a), the OA using SVM and NN algorithms had a similar tendency. The OA gradually increased with the increase in the feature number, and finally was close to 90%. For LPPSubFus-based classifications (Figure 7b), the OA of SVM-based classifications changed smoothly and ultimately remained at about 90%. The NN-based classifications had high OA with significant variation later, notably when the feature number exceeded 20. Most of the OA values in NN-based classifications were higher than those in SVM-based classifications. In addition, LPPSubFus-based classifications yielded higher OA than SubFus-based classifications, particularly after the feature number was greater than 16. To summarize, the LPPSubFus achieved better classifications than SubFus, and NN algorithms performed better than SVM algorithms in this study.

3.2.2. Comparison of Producer and User Accuracy

This study aimed to improve the separability of tea plantations via HS and SAR data. The tea and forest in the study area are evergreen broadleaved perennial vegetation [9]. Only if the tea plantation and forest are identified well can accurate tea plantation mapping be achieved. The PA and UA of tea plantations and forests using two classification algorithms under different scenarios are shown in Table 6. Both SVM and NN algorithms achieved high PA of tea plantations, and low PA of forests and low UA of tea plantations, when using scenarios 2–4. The PA and UA of tea plantation and forest only in the classifications of scenario 6 were above 90%, indicating that the tea plantations and forests in this classification were identified well with fewer misclassifications and omissions. Generally, the PA and UA of scenario 6 with NN algorithms were higher than those with SVM algorithms.
Figure 8 shows the PA and UA curves of tea plantations with the feature number using SubFus and LPPSubFus. In the SubFus-based classifications, after the feature number exceeded 18, the PA of classifications using SVM algorithms was slightly higher than that using NN algorithms, with the exception of when the feature number was 22 (Figure 8a,b). However, the UA of NN-based classifications at the later stage was slightly higher than that of SVM-based classifications. In the LPPSubFus-based classifications, there was no significant difference between the PA and UA with SVM and NN algorithms (Figure 8c,d). However, curves of NN-based classifications were sensitive to changes in the feature number with apparent volatility, particularly the UA of tea plantations in the LPPSubFus-based classifications. In general, LPPSubFus-based classifications achieved higher PA or UA than SubFus-based classifications. In particular, the PA of tea plantations using LPPSubFus was close to 100% later, and the UA was around 95% in most cases.
Figure 9 shows the PA and UA curves of forests with the feature number using SubFus and LPPSubFus. The PA and UA of SubFus-based classifications were relatively stable. The PA remained below 90% later, and the UA was close to 100% (Figure 9a,b). Regarding LPPSubFus-based classifications, the UA curves changed steadily (Figure 9d). However, the PA volatility of SVM- and NN-based classifications was more noticeable, especially for the latter (Figure 9c). When the feature number was between 18 and 22, the PA in LPPSubFus-based classifications using NN algorithms approached 100%, and the UA gradually approached 100%. In addition, LPPSubFus identified forest better than SubFus in this study.

4. Discussion

4.1. Improvement under Different Fusion Strategies

4.1.1. Comparison of Pixel- and Feature-Level Image Fusion

The classifications obtained by FLIF were consistent with the real distribution of the tea plantations and forests. In addition, FLIF achieved higher OA than PLIF, with the exception of the classification of SubFus using SVM algorithms, and a 1.31%–8.93% improvement was achieved using FLIF compared to PLIF (Table 5). In particular, LPPSubFus performed well in both SVM- and NN-based classifications, achieving the highest classification accuracy of 95%, and good PA and UA of tea plantations and forests. The above results demonstrate the superior ability of FLIF in mapping tea plantations. In PLIF-based classifications, the classification accuracy using SVM algorithms was not significantly improved compared to the HS data. Even in NN-based classifications, the accuracy dropped considerably. This shows that fusing HS and SAR data cannot guarantee classification accuracy improvement. Therefore, it is crucial to select the appropriate image fusion method for tea plantation mapping. Furthermore, the low OA of scenarios 2–4 in NN-based classifications indicate that FLIF also has the advantage of data dimension reduction and is compatible with more classification algorithms compared to PLIF.
As expected, the FLIF-based classification offered a more generalized visual appearance presentation and a more contiguous depiction of land cover types with fewer speckles than the PLIF-based classification (Figure 6). The FLIF-based classification was similar to that of the object-based classification. However, the performance of the object-based image analysis mainly depends on the complex parameter selection of image segmentation [46]. FLIF can avoid the complex parameter decisions and has an equal effect as object-based image analysis. Therefore, compared with PLIF and object-based image analysis, FLIF is superior in tea plantation mapping.

4.1.2. Comparison of SubFus and LPPSubFus

Based on the SubFus algorithms proposed by Rasti and Ghamisi, this study introduced LPP into SubFus and proposed the LPPSubFus algorithms. The main difference between LPPSubFus and SubFus is that LPPSubFus uses the projection matrix obtained by LPP to initialize variables instead of using singular value decomposition. We designed scenarios 5 and 6 to compare their performance in tea plantation mapping.
First, both SubFus and LPPSubFus effectively solved the noise phenomenon that existed in PLIF-based classifications, and the distribution of land cover obtained was similar. Moreover, the LPPSubFus-based classifications better coincided with our knowledge of the study area than SubFus-based classifications. Second, the classification accuracy results showed that LPPSubFus yielded an improvement in OA of 3.09% and 4.17% compared to SubFus in both SVM- and NN-based classifications (Table 5), respectively. Furthermore, LPPSubFus provided better classifications for tea plantation and forest than SubFus, with both PA and UA more than 90% (Table 6). Third, compared with HS data, LPPSubFus improved the OA in SVM and NN-based classifications, but SubFus did not. The above results indicate that LPPSubFus had outstanding and more stable performance than SubFus in tea plantation mapping. However, LPPSubFus also has shortcomings, such as not identifying waterbodies well in NN-based classifications.

4.2. Benefits of Image Fusion for Tea Plantation Mapping

Due to the development of satellite sensing technology, multi-source remote sensing data applications have become more prevalent [58,59,60,61]. Data from different sensors can provide extra information that makes up for the deficiency of single remote sensing data [25]. In addition, multi-source data are more easily accessible than high-quality time-series data because weather easily influences the optical sensors [35,62]. Among the many different sensors, HS and SAR data have the advantage of low cost and easy access.
The results show that HS data fused SAR data by FLIF improved OA by up to 3.09% compared to using HS data alone (Table 5). However, the fusion of HS and SAR data by PLIF led to decreased classification accuracy generally. SAR data can provide backscattering information in surface roughness, water content, and shape [21]. However, SAR data are easily affected by speckle noise due to the entirely different optical data imaging mechanism, which negatively impacts information extraction [24]. Only by choosing an appropriate image fusion method can useful information be effectively extracted. We found that not all image fusion methods improved classification accuracy for tea plantation mapping. Compared with optical data alone, it is difficult to judge whether the fusion of optical and SAR data can achieve better results in this study. Fortunately, LPPSubFus had stable performance in fusing HS and SAR data in this study, and the classification accuracy was improved under both SVM and NN algorithms. Therefore, it is beneficial to fuse HS and SAR data via LPPSubFus for tea plantation mapping, in terms of both data acquisition and classification accuracy.

4.3. Classification Algorithm for Tea Plantation Mapping

This study used two classification algorithms to compare the performance of different fusion methods in tea plantation mapping. The results show that both SVM and NN algorithms achieved good results with the highest OA of more than 90%. The OA in NN-based classifications using scenario 1 was the same as the highest OA in SVM-based classifications. Moreover, scenario 6 with the NN algorithms achieved the highest OA of 95%, and UA and PA above 90% for tea plantations and forests. These results show that NN algorithms had better performance than SVM in this study.
For scenarios 1, 5, and 6, the OA of the NN-based classifications was higher than that of SVM-based classifications, whereas the OA of scenarios 2–4 was not. We speculated that this might be due to too many features in the classifications of scenarios 2–4, and the NN algorithms had insufficient ability to process high-dimensional data. In addition, NN algorithms were sensitive to the feature number, and the OA volatility in NN-based classifications was more evident than that in SVM-based classifications (Figure 7). SVM algorithms are not susceptible to the “dimensionality disaster” and have a more robust ability to process a large number of features [63]. Therefore, the OA did not change significantly with the feature number. The results indicate the importance of data dimensionality reduction in classification using HS and SAR data. For some classification algorithms, too many features will reduce the data processing speed and classification accuracy. LPPSubFus achieved the highest accuracy in both SVM- and NN-based classifications, thus, it has advantages compared to other fusion methods in terms of classification accuracy and compatibility with different algorithms.

5. Conclusions

The accurate mapping of tea plantations ensures appropriate management of tea-producing region for environmental protection. We proposed a LPPSubFus method to fuse spectral indexes, backscattering, and texture information of HS and SAR data. This study then explored the performance of SubFus, LPPSubFus, and PLIF in tea plantation mapping with SVM and NN algorithms. Through visual comparison and accuracy assessment, the following conclusions can be drawn:
(1)
FLIF can reduce data dimensionality and significantly improve noise phenomenon and classification accuracy. However, PLIF did not contribute to classification accuracy and even led to decreased accuracy in some cases.
(2)
LPPSubFus improved OA by 3% compared to SubFus. In particular, LPPSubFus with NN algorithms achieved the highest OA of 95%, and the PA and UA for tea plantations and forests exceeded 90%, which was concluded to be the best method for tea plantation mapping in this study.
(3)
LPPSubFus is compatible with different classification algorithms and had more stable and superior tea plantation mapping performance than PLIF and SubFus.
In summary, two advantages exist for LPPSubFus for tea plantation mapping. First, LPPSubFus provides a higher classification accuracy than PLIF and has the ability to reduce data dimensionality. Second, compared to SubFus, LPPSubFus has better performance in both SVM and NN algorithms. This study provides a new approach for future tea plantation mapping. The LPPSubFus approach used in this study could also be extended to the use of HS and SAR data in other forest mapping applications.

Author Contributions

Y.C. and S.T. conceived and designed the study; Y.C. conducted all experiments and contributed to the manuscript writing; S.T. contributed to the manuscript revision. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the “Comprehensive Investigation and Evaluation on the Carrying Capacity of Resources and the Environment in Fujian Province” project of the Nanjing Center, China Geological Survey (grant no.DD20190301).

Acknowledgments

The authors thank the European Space Agency for providing Sentinel-1A data and the China Land Satellite Remote Sensing Application Center for providing Gaofen-5 data. The authors would like to express gratitude to Behnood Rasti for releasing the SubFus code. We appreciate the anonymous reviewers for their valuable suggestions and comments, which help improve the quality of this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, X.; He, Y. Discriminating varieties of tea plant based on Vis/NIR spectral characteristics and using artificial neural networks. Biosyst. Eng. 2008, 99, 313–321. [Google Scholar] [CrossRef]
  2. Kumar, A.; Manjunath, K.R.; Meenakshi; Bala, R.; Suda, R.K.; Singh, R.D.; Panigrahy, S. Field hyperspectral data analysis for discriminating spectral behavior of tea plantations under various management practices. Int. J. Appl. Earth Obs. Geoinf. 2013, 23, 352–359. [Google Scholar] [CrossRef]
  3. Wang, Y.; Hao, X.; Wang, L.; Xiao, B.; Wang, X.; Yang, Y. Diverse Colletotrichum species cause anthracnose of tea plants (Camellia sinensis (L.) O. Kuntze) in China. Sci. Rep. 2016, 6, 35287. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Li, N.; Zhang, D.; Li, L.; Zhang, Y. Mapping the spatial distribution of tea plantations using high-spatiotemporal-resolution imagery in northern Zhejiang, China. Forests 2019, 10, 856. [Google Scholar] [CrossRef] [Green Version]
  5. FAOSTAT Home Page. Available online: http://www.fao.org/home/en/ (accessed on 28 November 2020).
  6. Zhang, Z. Shucheng tea plantation optimal ecological zoning based on GIS spatial soil properties. In Proceedings of the 2017 32nd Youth Academic Annual Conference of Chinese Association of Automation, YAC 2017, Hefei, China, 19–21 May 2017; pp. 40–45. [Google Scholar] [CrossRef]
  7. Akar, Ö.; Güngör, O. Integrating multiple texture methods and NDVI to the Random Forest classification algorithm to detect tea and hazelnut plantation areas in northeast Turkey. Int. J. Remote Sens. 2015, 36, 442–464. [Google Scholar] [CrossRef]
  8. Chu, H.-J.; Wang, C.-K.; Kong, S.-J.; Chen, K.-C. Integration of full-waveform LiDAR and hyperspectral data to enhance tea and areca classification. GISci. Remote Sens. 2016, 53, 542–559. [Google Scholar] [CrossRef]
  9. Xu, W.; Qin, Y.; Xiao, X.; Di, G.; Doughty, R.B.; Zhou, Y.; Zou, Z.; Kong, L.; Niu, Q.; Kou, W. Quantifying spatial-temporal changes of tea plantations in complex landscapes through integrative analyses of optical and microwave imagery. Int. J. Appl. Earth Obs. Geoinf. 2018, 73, 697–711. [Google Scholar] [CrossRef]
  10. Evans, T.L.; Costa, M. Landcover classification of the Lower Nhecolândia subregion of the Brazilian Pantanal Wetlands using ALOS/PALSAR, RADARSAT-2 and ENVISAT/ASAR imagery. Remote Sens. Environ. 2013, 128, 118–137. [Google Scholar] [CrossRef]
  11. Costa, M.P.F. Use of SAR satellites for mapping zonation of vegetation communities in the Amazon floodplain. Int. J. Remote Sens. 2004, 25, 1817–1835. [Google Scholar] [CrossRef]
  12. Zhao, D.; Pang, Y.; Liu, L.; Li, Z. Individual tree classification using airborne lidar and hyperspectral data in a natural mixed forest of northeast China. Forests 2020, 11, 303. [Google Scholar] [CrossRef] [Green Version]
  13. Shang, X.; Chisholm, L.A. Classification of Australian native forest species using hyperspectral remote sensing and machine-learning classification algorithms. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2481–2489. [Google Scholar] [CrossRef]
  14. Hu, J.; Ghamisi, P.; Schmitt, A.; Zhu, X.X. Object based fusion of polarimetric SAR and hyperspectral imaging for land use classification. Work. Hyperspectr. Image Signal Process. Evol. Remote Sens. 2016, 1–5. [Google Scholar] [CrossRef]
  15. Jouan, A.; Allard, Y. Land use mapping with evidential fusion of polarimetric synthetic aperture Radar and hyperspectral imagery. Inf. Fusion 2002, 5, 251–267. [Google Scholar] [CrossRef]
  16. Pohl, C.; van Genderen, J. Multisensor image fusion in remote sensing: Concepts, methods and applications. Int. J. Remote Sens. 1998, 19, 823–854. [Google Scholar] [CrossRef] [Green Version]
  17. Klein, L.A. Sensor and Data Fusion Concepts and Applications, 2nd ed.; Society of Photo-Optical Instrumentation Engineers (SPIE): Bellingham, WA, USA, 1999; ISBN 0819432318. [Google Scholar]
  18. Wu, Y.; Zhang, X. Object-Based tree species classification using airborne hyperspectral images and LiDAR data. Forests 2020, 11, 32. [Google Scholar] [CrossRef] [Green Version]
  19. Xi, Y.; Ren, C.; Wang, Z.; Wei, S.; Bai, J.; Zhang, B.; Xiang, H.; Chen, L. Mapping tree species composition using OHS-1 hyperspectral data and deep learning algorithms in Changbai mountains, Northeast China. Forests 2019, 10, 818. [Google Scholar] [CrossRef] [Green Version]
  20. Shi, Y.; Skidmore, A.K.; Wang, T.; Holzwarth, S.; Heiden, U.; Pinnel, N.; Zhu, X.; Heurich, M. Tree species classification using plant functional traits from LiDAR and hyperspectral data. Int. J. Appl. Earth Obs. Geoinf. 2018, 73, 207–219. [Google Scholar] [CrossRef]
  21. Fu, B.; Wang, Y.; Campbell, A.; Li, Y.; Zhang, B.; Yin, S.; Xing, Z.; Jin, X. Comparison of object-based and pixel-based Random Forest algorithm for wetland vegetation mapping using high spatial resolution GF-1 and SAR data. Ecol. Indic. 2017, 73, 105–117. [Google Scholar] [CrossRef]
  22. Wang, X.Y.; Guo, Y.G.; He, J.; Du, L.T. Fusion of HJ1B and ALOS PALSAR data for land cover classification using machine learning methods. Int. J. Appl. Earth Obs. Geoinf. 2016, 52, 192–203. [Google Scholar] [CrossRef]
  23. Richard, J.; Blaschke, T.; Collins, M. Fusion of TerraSAR-x and Landsat ETM + data for protected area mapping in Uganda. Int. J. Appl. Earth Obs. Geoinf. 2015, 38, 99–104. [Google Scholar] [CrossRef]
  24. Zhang, H.; Xu, R. Exploring the optimal integration levels between SAR and optical data for better urban land cover mapping in the Pearl River Delta. Int. J. Appl. Earth Obs. Geoinf. 2018, 64, 87–95. [Google Scholar] [CrossRef]
  25. Kulkarni, S.C.; Rege, P.P. Pixel level fusion techniques for SAR and optical images: A review. Inf. Fusion 2020, 59, 13–29. [Google Scholar] [CrossRef]
  26. Hu, Q.; Sulla-menashe, D.; Xu, B.; Yin, H.; Tang, H.; Yang, P. A phenology-based spectral and temporal feature selection method for crop mapping from satellite time series. Int. J. Appl. Earth Obs. Geoinf. 2019, 80, 218–229. [Google Scholar] [CrossRef]
  27. Rasti, B.; Ghamisi, P. Remote sensing image classification using subspace sensor fusion. Inf. Fusion 2020, 64, 121–130. [Google Scholar] [CrossRef]
  28. Teillet, P.M.; Guindon, B.; Goodenough, D.G. On the slope-aspect correction of multispectral scanner data Seventh International Symposium Machine Processing of remotely sensed data. Can. J. Remote Sens. 1982, 8, 84–106. [Google Scholar] [CrossRef] [Green Version]
  29. Fraser, R.H.; Olthof, I.; Pouliot, D. Monitoring land cover change and ecological integrity in Canada’s national parks. Remote Sens. Environ. 2009, 113, 1397–1409. [Google Scholar] [CrossRef]
  30. Asante-Okyere, S.; Shen, C.; Ziggah, Y.Y.; Rulegeya, M.M.; Zhu, X. Principal component analysis (PCA) based hybrid models for the accurate estimation of reservoir water saturation. Comput. Geosci. 2020, 145, 104555. [Google Scholar] [CrossRef]
  31. Gitelson, A.A.; Gritz, Y.; Merzlyak, M.N. Relationships between leaf chlorophyll content and spectral reflectance and algorithms for non-destructive chlorophyll assessment in higher plant leaves. J. Plant Physiol. 2003, 160, 271–282. [Google Scholar] [CrossRef]
  32. Gitelson, A.A.; Merzlyak, M.N.; Chivkunova, O.B. Optical properties and nondestructive estimation of anthocyanin content in plant leaves. Photochem. Photobiol. 2001, 74, 38–45. [Google Scholar] [CrossRef]
  33. Vogelmann, J.E.; Rock, B.N.; Moss, D.M. Red edge spectral measurements from sugar maple leaves. Int. J. Remote Sens. 1993, 14, 1563–1575. [Google Scholar] [CrossRef]
  34. Kim, Y.; Glenn, D.M.; Park, J.; Ngugi, H.K.; Lehman, B.L. Hyperspectral image analysis for water stress detection of apple trees. Comput. Electron. Agric. 2011, 77, 155–160. [Google Scholar] [CrossRef]
  35. Negri, R.G.; Dutra, L.V.; Freitas, C.D.C.; Lu, D. Exploring the capability of ALOS PALSAR L-band fully polarimetric data for land cover classification in tropical environments. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 5369–5384. [Google Scholar] [CrossRef]
  36. Naidoo, L.; Mathieu, R.; Main, R.; Wessels, K.; Asner, G.P. L-band Synthetic Aperture Radar imagery performs better than optical datasets at retrieving woody fractional cover in deciduous, dry savannahs. Int. J. Appl. Earth Obs. Geoinf. 2016, 52, 54–64. [Google Scholar] [CrossRef]
  37. Li, G.; Lu, D.; Moran, E.; Dutra, L.; Batistella, M. A comparative analysis of ALOS PALSAR L-band and RADARSAT-2 C-band data for land-cover classification in a tropical moist region. ISPRS J. Photogramm. Remote Sens. 2012, 70, 26–38. [Google Scholar] [CrossRef] [Green Version]
  38. Moreira, A.; Prats-Iraola, P.; Younis, M.; Krieger, G.; Hajnsek, I.; Papathanassiou, K.P. A tutorial on synthetic aperture radar. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–43. [Google Scholar] [CrossRef] [Green Version]
  39. Schmitt, A.; Wendleder, A.; Hinz, S. The Kennaugh element framework for multi-scale, multi-polarized, multi-temporal and multi-frequency SAR image preparation. ISPRS J. Photogramm. Remote Sens. 2015, 102, 122–139. [Google Scholar] [CrossRef] [Green Version]
  40. Haralick, R.M.; Dinstein, I.; Shanmugam, K. Textural features for image classification. IEEE Trans. Syst. Man Cybern. 1973, SMC-3, 610–621. [Google Scholar] [CrossRef] [Green Version]
  41. He, X.; Niyogi, P. Locality Preserving Projections; MIT Press: Cambridge, MA, USA, 2004; ISBN 0262201526. [Google Scholar]
  42. Hu, J.; Hong, D.; Wang, Y.; Zhu, X. A comparative review of manifold learning techniques for hyperspectral and polarimetric SAR image fusion. Remote Sens. 2019, 11, 681. [Google Scholar] [CrossRef] [Green Version]
  43. Peng, Y.; Wang, Z.; He, K.; Liu, Y.; Li, Q.; Liu, L.; Zhu, X.; Haajek, P. Feature Extraction of Double Pulse Metal Inert Gas Welding Based on Broadband Mode Decomposition and Locality Preserving Projection. Math. Probl. Eng. 2020, 2020, 7576034. [Google Scholar] [CrossRef]
  44. Evans, D.L.; Farr, T.G.; Ford, J.P.; Thompson, T.W.; Werner, C.L. Multipolarization radar images for geologic mapping and vegetation discrimination. IEEE Trans. Geosci. Remote Sens. 1986, GE-24, 246–257. [Google Scholar] [CrossRef]
  45. Baghdadi, N.; Bernier, M.; Gauthier, R.; Neeson, I. Evaluation of C-band SAR data for wetlands mapping. Int. J. Remote Sens. 2001, 22, 71–88. [Google Scholar] [CrossRef]
  46. Duro, D.C.; Franklin, S.E.; Dubé, M.G. A comparison of pixel-based and object-based image analysis with selected machine learning algorithms for the classification of agricultural landscapes using SPOT-5 HRG imagery. Remote Sens. Environ. 2012, 118, 259–272. [Google Scholar] [CrossRef]
  47. Cai, Y.; Guan, K.; Peng, J.; Wang, S.; Seifert, C.; Wardlow, B.; Li, Z. A high-performance and in-season classification system of field-level crop types using time-series Landsat data and a machine learning approach. Remote Sens. Environ. 2018, 210, 35–47. [Google Scholar] [CrossRef]
  48. Shao, Y.; Lunetta, R.S. Comparison of support vector machine, neural network, and CART algorithms for the land-cover classification using limited training data points. ISPRS J. Photogramm. Remote Sens. 2012, 70, 78–87. [Google Scholar] [CrossRef]
  49. Couellan, N.; Wang, W. Bi-level stochastic gradient for large scale support vector machine. Neurocomputing 2015, 153, 300–308. [Google Scholar] [CrossRef]
  50. Mountrakis, G.; Im, J.; Ogole, C. Support vector machines in remote sensing: A review. ISPRS J. Photogramm. Remote Sens. 2011, 66, 247–259. [Google Scholar] [CrossRef]
  51. Dihkan, M.; Guneroglu, N.; Karsli, F.; Guneroglu, A. Remote sensing of tea plantations using an SVM classifier and pattern-based accuracy assessment technique. Int. J. Remote Sens. 2013, 34, 8549–8565. [Google Scholar] [CrossRef]
  52. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  53. Kimes, D.S.; Nelson, R.F.; Manry, M.T.; Fung, A.K. Review article: Attributes of neural networks for extracting continuous vegetation variables from optical and radar measurements. Int. J. Remote Sens. 1998, 19, 2639–2663. [Google Scholar] [CrossRef]
  54. Xie, Z.; Chen, Y.; Lu, D.; Li, G. Classification of Land Cover, Forest, and Tree Species Classes with ZiYuan-3 Multispectral and Stereo Data. Remote Sens. 2019, 11, 164. [Google Scholar] [CrossRef] [Green Version]
  55. Paola, J.D.; Schowengerdt, R.A. A Detailed Comparison of Backpropagation Neural Network and Maximum-Likelihood Classifiers for Urban Land Use Classification. IEEE Trans. Geosci. Remote Sens. 1995, 33, 981–996. [Google Scholar] [CrossRef]
  56. Zhang, C.; Xie, Z. Combining object-based texture measures with a neural network for vegetation mapping in the Everglades from hyperspectral imagery. Remote Sens. Environ. 2012, 124, 310–320. [Google Scholar] [CrossRef]
  57. Hsu, C.; Chang, C.; Lin, C. A practical guide to support vector classification. BJU Int. 2008, 101, 1396–1400. [Google Scholar]
  58. Zhu, C.; Luo, J.; Shen, Z.; Huang, C. Wetland mapping in the Balqash lake basin using multi-source remote sensing data and topographic features synergic retrieval. Procedia Environ. Sci. 2011, 10, 2718–2724. [Google Scholar] [CrossRef] [Green Version]
  59. Zhou, Y.; Hu, J.; Li, Z.; Li, J.; Zhao, R.; Ding, X. Quantifying glacier mass change and its contribution to lake growths in central Kunlun during 2000–2015 from multi-source remote sensing data. J. Hydrol. 2019, 570, 38–50. [Google Scholar] [CrossRef]
  60. Zhan, T.; Gong, M.; Liu, J.; Zhang, P. Iterative feature mapping network for detecting multiple changes in multi-source remote sensing images. ISPRS J. Photogramm. Remote Sens. 2018, 146, 38–51. [Google Scholar] [CrossRef]
  61. He, Y.; Chen, G.; Potter, C.; Meentemeyer, R.K. Integrating multi-sensor remote sensing and species distribution modeling to map the spread of emerging forest disease and tree mortality. Remote Sens. Environ. 2019, 231, 111238. [Google Scholar] [CrossRef]
  62. Yusoff, N.M.; Muharam, F.M.; Takeuchi, W.; Darmawan, S.; Razak, M.H.A. Phenology and classification of abandoned agricultural land based on ALOS-1 and 2 PALSAR multi-temporal measurements. Int. J. Digit. Earth 2017, 10, 155–174. [Google Scholar] [CrossRef]
  63. Yu, L.; Porwal, A.; Holden, E.J.; Dentith, M.C. Towards automatic lithological classification from remote sensing data using support vector machines. Comput. Geosci. 2012, 45, 229–239. [Google Scholar] [CrossRef]
Figure 1. Location of Wuyishan City, Nanping City, Fujian Province, China, and the location of the study area: (a) the hyperspectral image (from Chinese Gaofen-5 satellite) and (b) the Synthetic Aperture Radar (SAR) image (from Sentinel-1A satellite) of the study area.
Figure 1. Location of Wuyishan City, Nanping City, Fujian Province, China, and the location of the study area: (a) the hyperspectral image (from Chinese Gaofen-5 satellite) and (b) the Synthetic Aperture Radar (SAR) image (from Sentinel-1A satellite) of the study area.
Forests 11 01357 g001
Figure 2. Results of field surveys in the study area: (a) tea plantation in the flat area, (b) tea plantation on hillside, (c) tea plantation and forest, (d) cropland in winter, (e) cropland in summer, (f) mixed planting of tea and crops, (g) waterbody, (h) built-up area, and (i) rock.
Figure 2. Results of field surveys in the study area: (a) tea plantation in the flat area, (b) tea plantation on hillside, (c) tea plantation and forest, (d) cropland in winter, (e) cropland in summer, (f) mixed planting of tea and crops, (g) waterbody, (h) built-up area, and (i) rock.
Forests 11 01357 g002
Figure 3. Spectral characteristic curves of six land cover types in the study area.
Figure 3. Spectral characteristic curves of six land cover types in the study area.
Forests 11 01357 g003
Figure 4. Results of tea vegetation index (TVI). (a) TVI 1 : the image calculated by the slope between wavelength 865 and 454 nm; (b) TVI 2 : the image calculated by the slope between wavelength 2032 and 1266 nm; (c) TVI 3 : the image calculated by the triangle area enclosed by wavelength 454, 1006, and 2032 nm.
Figure 4. Results of tea vegetation index (TVI). (a) TVI 1 : the image calculated by the slope between wavelength 865 and 454 nm; (b) TVI 2 : the image calculated by the slope between wavelength 2032 and 1266 nm; (c) TVI 3 : the image calculated by the triangle area enclosed by wavelength 454, 1006, and 2032 nm.
Forests 11 01357 g004
Figure 5. Sentinel-1A data profiles of six land cover types in the study area. VV: vertical transmit and vertical receive; VH: vertical transmit and horizontal receive.
Figure 5. Sentinel-1A data profiles of six land cover types in the study area. VV: vertical transmit and vertical receive; VH: vertical transmit and horizontal receive.
Forests 11 01357 g005
Figure 6. Classifications of different scenarios using the support vector machine (SVM) and neural network (NN) algorithms: (a) scenario 1 by SVM algorithms, (b) scenario 2 by SVM algorithms, (c) scenario 3 by SVM algorithms, (d) scenario 4 by SVM algorithms, (e) scenario 5 by SVM algorithms, (f) scenario 6 by SVM algorithms, (g) scenario 1 by NN algorithms, (h) scenario 2 by NN algorithms, (i) scenario 3 by NN algorithms, (j) scenario 4 by NN algorithms, (k) scenario 5 by NN algorithms, and (l) scenario 6 by NN algorithms.
Figure 6. Classifications of different scenarios using the support vector machine (SVM) and neural network (NN) algorithms: (a) scenario 1 by SVM algorithms, (b) scenario 2 by SVM algorithms, (c) scenario 3 by SVM algorithms, (d) scenario 4 by SVM algorithms, (e) scenario 5 by SVM algorithms, (f) scenario 6 by SVM algorithms, (g) scenario 1 by NN algorithms, (h) scenario 2 by NN algorithms, (i) scenario 3 by NN algorithms, (j) scenario 4 by NN algorithms, (k) scenario 5 by NN algorithms, and (l) scenario 6 by NN algorithms.
Forests 11 01357 g006
Figure 7. Overall accuracy curves varying with the feature number using support vector machine (SVM) and neural network (NN) algorithms: (a) overall accuracy using subspace fusion (SubFus) and (b) overall accuracy using locality preservation projection and subspace fusion (LPPSubFus).
Figure 7. Overall accuracy curves varying with the feature number using support vector machine (SVM) and neural network (NN) algorithms: (a) overall accuracy using subspace fusion (SubFus) and (b) overall accuracy using locality preservation projection and subspace fusion (LPPSubFus).
Forests 11 01357 g007
Figure 8. Producer and user accuracy curves of tea plantations varying with the feature number using support vector machine (SVM) and neural network (NN) algorithms: (a) producer accuracy using subspace fusion (SubFus), (b) user accuracy using SubFus, (c) producer accuracy using locality preservation projection and subspace fusion (LPPSubFus), and (d) user accuracy using LPPSubFus.
Figure 8. Producer and user accuracy curves of tea plantations varying with the feature number using support vector machine (SVM) and neural network (NN) algorithms: (a) producer accuracy using subspace fusion (SubFus), (b) user accuracy using SubFus, (c) producer accuracy using locality preservation projection and subspace fusion (LPPSubFus), and (d) user accuracy using LPPSubFus.
Forests 11 01357 g008
Figure 9. Producer and user accuracy curves of forests varying with the feature number using support vector machine (SVM) and neural network (NN) algorithms: (a) producer accuracy using subspace fusion (SubFus), (b) user accuracy using SubFus, (c) producer accuracy using locality preservation projection and subspace fusion (LPPSubFus), and (d) user accuracy using LPPSubFus.
Figure 9. Producer and user accuracy curves of forests varying with the feature number using support vector machine (SVM) and neural network (NN) algorithms: (a) producer accuracy using subspace fusion (SubFus), (b) user accuracy using SubFus, (c) producer accuracy using locality preservation projection and subspace fusion (LPPSubFus), and (d) user accuracy using LPPSubFus.
Forests 11 01357 g009
Table 1. Characteristics of optical and synthetic aperture radar (SAR) images in this study.
Table 1. Characteristics of optical and synthetic aperture radar (SAR) images in this study.
Optical ImageSAR Image
SensorGaofen-5SensorSentinel-1A
Spatial resolution30 mSpatial resolution5 m × 20 m
Spectral resolution5–10 nmWavelengthC
Band informationVisible–near-infrared0.39–1.03 μm, 150 bands in totalPolarizationVV and VH
Short wave infrared1.0–2.5 μm, 180 bands in total
Acquisition time19 November 201918 November 2019
VV: vertical transmit and vertical receive; VH: vertical transmit and horizontal receive.
Table 2. Description of spectral indexes used in this study.
Table 2. Description of spectral indexes used in this study.
Name of VariableDescription
Chlorophyll Index 2 ρ 762   nm / ρ 702   nm 1
Anthocyanin Reflectance Index 2 ρ 800   nm ( 1 / ρ 550   nm 1 / ρ 700   nm )
Vogelmann Red Edge Index 2 ( ρ 734   nm ρ 747   nm ) / ( ρ 715   nm + ρ 726   nm )
Water Band Index ρ 900   nm / ρ 970   nm
Table 3. Classification scenarios using the support vector machine and neural network algorithms.
Table 3. Classification scenarios using the support vector machine and neural network algorithms.
ScenarioImage Layer Used for Classification
1Optical: first three PC and six bands, seven spectral indexes, and 18 textures.
2Optical: first three PC bands, six fusion bands using PCA, seven spectral indexes, and 18 textures; SAR: four polarimetric features, DEM data, and 24 textures.
3Optical: first three PC bands, six fusion bands using DWT, seven spectral indexes, and 18 textures; SAR: four polarimetric features, DEM data, and 24 textures.
4Optical: first three PC bands, six fusion bands using W-PCA, seven spectral indexes, and 18 textures; SAR: four polarimetric features, DEM data, and 24 textures.
5Optical: first three PC and six bands, seven spectral indices, and 18 textures; SAR: four polarimetric features, DEM data, and 24 textures. (SubFus)
6Optical: first three PC and six bands of HS data, seven spectral indices, and 18 textures. SAR: four polarimetric features, DEM data, and 24 textures. (LPPSubFus)
PCA: principal component analysis; DWT: discrete wavelet transforms; W-PCA: wavelet-principal component analysis; DEM: digital elevation model; HS: hyperspectral; SAR: Synthetic Aperture Radar; SubFus; subspace fusion; LPPSubFus: locality preserving projection and subspace fusion.
Table 4. Quantities of training and test samples.
Table 4. Quantities of training and test samples.
Sample TypeTrainingTest
PolygonsPixelsPolygonsPixels
Tea plantations3012060240
Forests3012060240
Croplands156030120
Rocks156030120
Built-up area and waterbodies156030120
Total105420210840
Table 5. Overall accuracy and Kappa coefficient with support vector machine (SVM) and neural network (NN) algorithms for different scenarios.
Table 5. Overall accuracy and Kappa coefficient with support vector machine (SVM) and neural network (NN) algorithms for different scenarios.
ScenarioOverall AccuracyKappa CoefficientClassification Algorithm
189.05%0.8593SVM
289.88%0.8698
387.74%0.8423
489.40%0.8636
5 (r = 27)89.05%0.8589
6 (r = 28)92.14%0.8990
192.14%0.8989NN
288.33%0.8501
386.07%0.8213
487.98%0.8458
5 (r = 12)90.83%0.8815
6 (r = 22)95.00%0.9355
Table 6. Producer accuracy (PA) and user accuracy (UA) of tea plantations and forests with support vector machine (SVM) and neural network (NN) algorithms for different scenarios.
Table 6. Producer accuracy (PA) and user accuracy (UA) of tea plantations and forests with support vector machine (SVM) and neural network (NN) algorithms for different scenarios.
ScenarioPA of Tea Plantation (%)UA of Tea Plantation (%)PA of Forest (%)UA of Forest (%)Classification Algorithm
193.7589.2985.4296.70SVM
294.5889.7286.6795.85
392.9285.1180.8394.17
495.4286.4283.3396.62
5 (r = 27)93.7587.5588.3396.80
6 (r = 28)94.5893.8092.5098.67
197.0891.7387.596.77NN
290.4285.4380.8392.38
390.8383.8576.6792.93
491.6784.9477.9294.92
5 (r = 12)92.5086.0590.0092.31
6 (r = 22)96.2594.6796.6796.27
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, Y.; Tian, S. Feature-Level Fusion between Gaofen-5 and Sentinel-1A Data for Tea Plantation Mapping. Forests 2020, 11, 1357. https://doi.org/10.3390/f11121357

AMA Style

Chen Y, Tian S. Feature-Level Fusion between Gaofen-5 and Sentinel-1A Data for Tea Plantation Mapping. Forests. 2020; 11(12):1357. https://doi.org/10.3390/f11121357

Chicago/Turabian Style

Chen, Yujia, and Shufang Tian. 2020. "Feature-Level Fusion between Gaofen-5 and Sentinel-1A Data for Tea Plantation Mapping" Forests 11, no. 12: 1357. https://doi.org/10.3390/f11121357

APA Style

Chen, Y., & Tian, S. (2020). Feature-Level Fusion between Gaofen-5 and Sentinel-1A Data for Tea Plantation Mapping. Forests, 11(12), 1357. https://doi.org/10.3390/f11121357

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop