Next Article in Journal
Estimating Nitrogen from Structural Crop Traits at Field Scale—A Novel Approach Versus Spectral Vegetation Indices
Next Article in Special Issue
WorldView-2 Data for Hierarchical Object-Based Urban Land Cover Classification in Kigali: Integrating Rule-Based Approach with Urban Density and Greenness Indices
Previous Article in Journal
Measuring and Predicting Urban Expansion in the Angkor Region of Cambodia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CNN-Based Land Cover Classification Combining Stratified Segmentation and Fusion of Point Cloud and Very High-Spatial Resolution Remote Sensing Image Data

1
School of Information Engineering China University of Geosciences (Beijing), 29 Xueyuan Road, Haidian, Beijing 100083, China
2
Polytechnic Center for Natural Resources Big-Data, MNR of China, Beijing 100036, China
3
Key Laboratory of Virtual Geographic Environment, Nanjing Normal University, Ministry of Education, Nanjing 210023, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(17), 2065; https://doi.org/10.3390/rs11172065
Submission received: 31 July 2019 / Revised: 27 August 2019 / Accepted: 30 August 2019 / Published: 2 September 2019

Abstract

:
Traditional and convolutional neural network (CNN)-based geographic object-based image analysis (GeOBIA) land-cover classification methods prosper in remote sensing and generate numerous distinguished achievements. However, a bottleneck emerges and hinders further improvements in classification results, due to the insufficiency of information provided by very high-spatial resolution images (VHSRIs). To be specific, the phenomenon of different objects with similar spectrum and the lack of topographic information (heights) are natural drawbacks of VHSRIs. Thus, multisource data steps into people’s sight and shows a promising future. Firstly, for data fusion, this paper proposed a standard normalized digital surface model (StdnDSM) method which was actually a digital elevation model derived from a digital terrain model (DTM) and digital surface model (DSM) to break through the bottleneck by fusing VHSRI and cloud points. It smoothed and improved the fusion of point cloud and VHSRIs and thus performed well in follow-up classification. The fusion data then were utilized to perform multiresolution segmentation (MRS) and worked as training data for the CNN. Moreover, the grey-level co-occurrence matrix (GLCM) was introduced for a stratified MRS. Secondly, for data processing, the stratified MRS was more efficient than unstratified MRS, and its outcome result was theoretically more rational and explainable than traditional global segmentation. Eventually, classes of segmented polygons were determined by majority voting. Compared to pixel-based and traditional object-based classification methods, majority voting strategy has stronger robustness and avoids misclassifications caused by minor misclassified centre points. Experimental analysis results suggested that the proposed method was promising for object-based classification.

1. Introduction

With the development in artificial intelligence (AI) and remote sensing (RS) sensors in recent years, images with very high resolution are demanding more efficient processing means. Therefore, it is becoming increasingly hard to ignore deep learning (DL) models in object classification for RS images [1]. Deep learning which originated from artificial neural network (ANN) is currently widely adopted in diverse domains [2]. Among all DL methods, convolutional neural network is an approved method and has been successfully applied in RS.
Object classification generally consists of land-cover classification and land-use classification [3,4]. For urban areas, objects in remote sensing images are usually classified into diverse land-use categories, while objects are classified as land-cover sorts for non-urban areas. However, in this paper, object classification referred to land-cover classification for the lack of social attribute data. Desirable object classification was based on solid segmentation results. Different segmentation algorithms abound, and these methods can be divided into diverse types from different perspectives [5,6]. According to the nature of segmentation unit, segmentation algorithms are classified as pixel-level and object-level (pixel-based and object-based) segmentation [7,8,9,10]. The term of geographic object-based image analysis (GeOBIA) was first proposed by Blaschke [11]. For different experiment purposes, corresponding segmentation algorithms are selected. A superpixel-based segmentation method has been proposed for classification of very high-spatial resolution images (VHSRIs) [12,13]. Despite obvious advantages compared to pixel-based segmentation algorithms, GeOBIA still requires improvements. To further improve segmentation results, the strategy of stratification was introduced. Zhou et al. and Xu et al. [14] adopted grey-level co-occurrence matrix (GLCM) for a pre-segmentation which roughly segments the image into several homogeneously alike regions before multiresolution segmentation (MRS), which brought new possibilities for fine-MRS. Additionally, an algorithm adopting the voting strategy, which further increases classification efficiency was proposed for convolutional neural network (CNN)-based GeOBIA [15,16,17,18].
Deep learning is a currently adopted technique in GeOBIA-based land cover classification. Deep learning was reinterpreted in 2006 and has been booming ever since then under the unremitting efforts by Hinton [19]. The very first application adopts and realizes DL is from Lecun [20] and has been compared to traditional machine learning methods, including random forest and support vector machine, ever since [21]. However, LeNet fails to recognize images as complex as VHSRIs and is inefficient for land-cover classification in RS. DL consists of supervised learning, semi-supervised and unsupervised learning [22]. Among all models, CNNs (Convolutional Neural Networks) are one of the most frequently adopted deep supervised learning models. Alexnet, which won the Large Scale Visual Recognition Challenge 2012 (ILSVRC2012), successfully adopted rectified linear units (ReLU) as activate function and proposed LRN [23]. The merit of LRN is that it renders Alexnet a strong generalization capacity which enables Alexnet to learn and extract features from complex images including VHSRIs. Meanwhile, the advancement in GPU also encouraged the revival of machine learning. CNN has been applied in numerous domains, especially in computer vision related fields. In the RS field, CNN has been proven as a reliable tool for extraction and classification [24]. Scott et al. [25] further improves land-cover classification accuracy by adopting a deep CNN model. Apart from Alexnet, other newly proposed CNNs including VGGNet (Visual Geometry Group Network), ResNet (Residual Neural Network) and FCN (Fully Convolutional Network) are also successfully applied in GeOBIA [26]. Moreover, VGG which believes that CNNs with deeper architectures generate higher classification accuracy has been applied in VHSRI classification [27]. Apart from CNNs models, traditional machine learning (ML) algorithms and other DL methods perform effectively as well. Hong et al. [28] and Lu et al. [29] introduced the richer convolutional feature (RCF) to road and building edge detection in VHSRIs and overwhelmed traditional methods. Patch-based (CNN) and FCN are two of the most used models currently [30,31]. FCN outputs a result of the exact same size of input through deconvolution (backward learning). However, the structure of FCN is tedious. Even though the combining of CNN and RS images has generated great outcomes, a significant type of information, i.e., height, is missing in classification. Therefore, the introduction of light detection and ranging (LiDAR) becomes inevitable. Point cloud are data generated by LiDAR and are usually utilized as supplementary data in geoscience related fields. Multisource data fusion is commonly seen in RS and its concrete application such as object classification [32]. Multisource data fusion has always been an indispensable topic in RS since the beginning. Data sensed by diverse sensors, such as point cloud (PC), synthetic aperture radar (SAR), points of interests (POI), social sensing data [33,34] and surveyed data, have been applied in RS for deeper analysis in the past two decades [35,36,37]. Different data reflects unique features for diverse objects. POIs refer to objects that attract specific researchers for a certain purpose and is promising for RS image classification. POIs can usually be buildings, bus stops, railway stations, hospitals, etc. Compared to RS imagery, POIs contain social attributes [38]. However, the addition of POIs has its own drawbacks that the information reflected by POIs may not be correct or timely. Similarly, social sensing data also reflects social properties that traditional RS data hardly shows [39]. Surveyed data reflects geometric information and may serve well as auxiliary data. Point cloud are generated by LiDAR laser sensors and mainly reveals elevation of the study area [40,41,42]. PC can be studied separately and be the auxiliary data in RS imagery analysis as well. The combination of PC and RS images demonstrates that the adjunction of elevation information further improves classification results [43,44].
However, although research adopting GeOBIA and DL in VHSRIs classification abound, as well as combining point cloud fusion in remote sensing land cover classification, more efforts are still potentially required. Firstly, due to that the complexity of scale dependence in segmentation of GeOBIA, fine segmentation is the key to GeOBIA classification. Secondly, from the perspective of classification features, how to effectively enhance the feature difference between different object classes or significantly increase the information entropy is the main purpose of data fusion. As a result, the performance of point cloud fusion and the quality of image segmentation should be further improved.
Aiming at these requirements, this paper explored a CNN based land cover classification method combining stratified segmentation and fusion of point cloud and VHSRI data, in which a new fusion named standard normalized digital surface model (StdnDSM) of PC and VHSRIs was first proposed in this paper, then a stratification strategy combining GLCM and MRS was applied to segment the fused data, and finally a finely tuned CNN model was utilized to train samples and classify land-cover objects. Image entropy was introduced to evaluate image quality for StdnDSM. For CNN based GeOBIA classification, the region majority voting strategy was applied to accelerate the procedure and avoid extreme situations that former methods fail to solve. A scene in Helsinki was chosen as the study area and the corresponding data was collected for study.
The remaining parts of this paper were settled as follows. Relevant critical methods and algorithms are well described in Section 2. In Section 3, the proposed StdnDSM and corresponding fused data are compared with former fused and solo data. Meanwhile, the adoption of GLCM in MRS and utilization of PC in segmentation and classification were proved as superior. Pros and cons of the experiment are analysed and discussed in Section 4 and finally in Section 5 conclusions are drawn.

2. Materials and Methods

In this part, four parts are introduced in succession. Firstly, the novel proposed StdnDSM is minutely explained. Secondly, GLCM and MRS along with their relationship are fully demonstrated. Then, the CNN model and classification strategy adopted in experiments are illustrated separately. Eventually, accuracy assessment methods are presented. The workflow for the CNN based land cover classification method combining stratified segmentation and point cloud data fusion is demonstrated in Figure 1.

2.1. StdnDSM for Point Cloud Data Fusion

Before StdnDSM is described in detail, three terms, digital terrain model (DTM), digital surface model (DSM) and normalised DSM (nDSM; also known as digital height model, DHM), ought to be introduced. Moreover, all four digital models here are generated from PC. DTM directly reflects the elevation of the ground. DTM consists of a grid model, triangulated irregular network (TIN) model and contour model, among which grid DTM is most ideal for DL input. Equation (1) shows how DTM is obtained from PC by Kriging interpolation:
Z x 0 = i = 1 n λ i Z x i
where Z x 0 is the required elevation of a certain point and Z x i are the elevation of known points nearby. The Z x 0 is actually the weighted mean of Z x i and λ i is the weight of neighbouring point x i . However, DTM is hardly suitable for data fusion with VHSRIs as segmentation and classification are object-based while DTM fails to emphasize objects. Similarly, DSM (digital surface model) is deficient for fusion or DL.
Thus, nDSM (normalized digital surface model) which is the outcome of subtracting DTM from DSM was proposed to be fused with VHSRIs. Equation (2) presents the generation of nDSM:
nDSM = DSM DTM .
The advantage of nDSM, compared to DTM and DSM, is that nDSM only reflects relative heights. To put it another way, objects such as bare lands and waters read zero in digital number (DN) value while objects including vehicles, trees and buildings are distinctly above zero. Hence, differences among objects in elevation are reflected by nDSM. Yet, nDSM is scarcely perfect for fusion with VHSRIs as the D-values between nDSM and VHSRIs are commonly tremendous. Since nDSM is generated from PC whose accuracy in height reaches centimetre level (airborne-based LiDAR), its DN value floats within thousands. However, for VHSRIs or ordinary RS images, DN values are usually restricted to a certain range, e.g., 0–255 for 8 bit images, 0–65536 for 16 bit images. Therefore, there is an issue of dimensional disunity between nDSM and DN of VHSRI. nDSM may overwhelm VHSRIs if the two are directly fused. Given such circumstances, considering the uniform of dimension, StdnDSM was proposed in this paper to fully and reasonably combine PC and VHSRIs. StdnDSM was obtained by standardly constraining the DN value of nDSM within the same range of DN of VHSRI. In Equation (3), StdnDSM is defined as follows:
StdnDSM = δ D N D N M i n D N M a x D N M i n
where δ is the extent of RS image which is to be fused with StdnDSM. DN is the digital number of a certain pixel, D N M a x is the maximum digital number around the image and D N M i n is the minimum digital number of the image. In Figure 2, original PC data is illustrated along with StdnDSM. The eventual aim for proposing StdnDSM was to convert cloud point data which was non-suitable for DL into an ideal alternative raster image that mathematically surpassed nDSM with richer information.
From the comparison of DTM, DSM, nDSM and StdnDSM shown in Figure 3, it can be vividly seen that StdnDSM and nDSM beat the other two images visually, especially for urban green belts and buildings. Urban green belt and building areas in DTM and DSM visibly appeared to be blurrier compared to nDSM and StdnDSM. The visual effect of DTM was completely ambiguous and basically deficient for any object-level segmentation or classification. As for DSM, essential figures of buildings and woods were distinct while they were mixed in border areas. However, for nDSM and StdnDSM, practically all objects were legible. The reason for such an outcome was that nDSM and StdnDSM removed influences of ground surface and contained only objects heights. The cause behind the excellence of StdnDSM in data fusion was the absence of potential excessive dominance over VHSRIs. That is, in fusion data the influence of elevation reflected by StdnDSM (PC) and spectral information contained in VHSRIs were equally distributed.
As Figure 4 shows, the corresponding fusion data of DTM (a), DSM (b), nDSM (c) and StdnDSM (d) are compared. All fusion data were visually promising for classification while the fusion of StdnDSM and VHSRI remained as the optimal. It was apparent that the fusion data of StdnDSM and VHSRI was spectrally homogenous, which was the credit of confining elevation information consistent with spectral information in the DN value. To put it differently, StdnDSM surpassed DTM, DSM and nDSM in data fusion with RGB images for adding elevation into RS data.

2.2. Stratified Segmentation

As for land-cover classification, objects of diverse sizes including vehicles, buildings and waters were distributed throughout the research area. A uniquely fixed parameter in MRS hardly performs perfectly for all those objects. Put differently, the best segmentation scale for buildings was not the same as water or other objects. Therefore, GLCM based region partition was involved before the fine scale segmentation to improve segmentation efficiency and then heighten classification accuracy. The co-occurrence matrix which was a function of the distance and angular relation between two neighbouring pixels reflected the times a pixel and its specific neighbouring co-occurrence [45,46]. GLCM considers eight directions including 0°, 45°, 90°, 135°, 180°, 225°, 270° and 315° between a pair of neighbouring pixels in specific distance d [47]. In Figure 5, an exhibition of the GLCM working principle is presented. Figure 5a shows the eight viable processing directions. Figure 5b presents a demonstration of a 3 × 3 window as it moves along “1,0” (90°) direction and Figure 5c is the related GLCM calculation result.
GLCM in application acted as texture filters based on the co-occurrence matrix. Those filters, namely mean, variance, homogeneity, contrast, dissimilarity, entropy, second moment, correlation, etc., denote respective texture features [48]. Twenty-three of twenty-eight filters, including contrast, second moment and correlation, were deemed as irrelevant in this experiment and shall not be introduced next. The remaining five filters, i.e., mean, variance, homogeneity, dissimilarity and entropy, are explained and illustrated by Equations (4)–(8). Equation (4) shows how the local mean value of the processing window is obtained.
Mean = i j P i , j / N g 2
where P i , j denotes the value of i th row and j th column in grey-the level co-occurrence matrix while N g refers to the number of grey levels. The variation which shows the extent how pixels in GLCM change is presented as Equation (5).
Var = i j i M e a n 2 p i , j
where p i , j is the i , j th entry in a normalized grey-tone spatial-dependence matrix and equals P i , j / R . R is the number of neighbouring resolution cell pairs for generating the grey-tone spatial-dependence matrix. In Equation (6), homogeneity which is calculated adopting the inverse difference moment algorithm is demonstrated:
Hom = i j 1 1 + i j 2 p i , j .
Filter homogeneity emphases pixels with co-occurrence relation in texture space. On the contrary, filter dissimilarity amplifies the unlikeness between neighbouring pixels in texture and is defined as Equation (7):
Dis = n = 1 N g 1 n i = 1 N g j = 1 N g p i , j i j = n .
The absolute D-value of greyscale, i j = n , is used here to acquire dissimilarity. The last filter adopted in this paper was entropy which reflects the redundancy of texture information and is explained as Equation (8):
Ent = i j p i , j log p i , j .
The outcome of stratification strategy adds texture information in MRS, which improves segmentation in both result and efficiency. However, the final segmentation result was determined by stratification strategy and MRS. The reason for selecting MRS before classification was that MRS performs better in consistency and most importantly efficiency than other segmentation methods. For large research areas, the efficiency of segmentation greatly influences experimental process. The essence of MRS is fractal net evolution approach (FNEA) [49]. In a certain scale, objects diverse in size who seldom share the same finest segmentation resolution synchronously exist. Thus, MRS performs a pixel-level bottom-up local grow segmentation which strictly adheres to maximum homogeneity/minimum heterogeneity criterion to merge spectrally homogeneous pixels. However, not only spectral feature but also the shape is considered as an indispensable part in MRS. In Equation (9), the permissible heterogeneity upper bound H e t e r o which decides the area of segmentation objects is defined.
Hetero = ω c o l o r × h e t e c o l o r + ω s h a p e × h e t e s h a p e
where ω c o l o r and ω s h a p e are respectively the weight of spectral and shape feature while h e t e r o c o l o r and h e t e r o s h a p e denote heterogeneity of objects in spectra and shape. In Equations (10) and (11), h e t e r o c o l o r and h e t e r o s h a p e are further introduced.
h e t e r o c o l o r = x ω x n a h e t e m c h e t e a x + n b h e t e m c h e t e b x
where x means a certain band in the image and ω x refers to the weight of this band in all bands. n a and n b are the number of pixels in object a and object b who are to be merged. h e t e r o a x and h e t e r o b x are respectively the heterogeneity of object a and object b in band x while h e t e r o m c denotes the heterogeneity of potential merged object c.
h e t e r o s h a p e = l n
where l refers to the perimeter of the possible object c while n is the number of pixels in object c.
As is known, the similar objects always cluster in the same local region with similar sizes. Thus, dividing the VHSRI and finely extracting the object in local regions can improve the scale suitability and accuracy. By combining GCLM based region partition and MRS, this paper introduced a stratified MRS method. The stratified MRS method segmented the whole image into several large regions before independently segmenting each region at different scales.

2.3. Convolutional Neural Network and Alexnet

Convolutional neural network which was enlightened by visual perception mechanism of cats is an indispensable subfield of deep learning. Four fundamental ideas behind CNN [50] respectively are local connections, shared weights, pooling and utilization of numerous layers. Diverse CNN algorithms abound in tremendous domains, however there are three basic layers, namely convolutional layer, pooling layer and fully-connected layer, which remain unchanged. Convolutional layers extract local conjunctions from features, and then similar features are merged into one by pooling layers. Eventually, all newly obtained features are combined by fully-connected layers. The natural advantage of CNN is avoiding over-fitting compared to traditional DL algorithms. Moreover, images can be directly input into CNN, which prevents intricate data pre-processing such as data reconstruction. Whereas, due to the black-box mechanism, features extracted are unintelligible and inexplicable, which is the double-edged sword for CNN and brings unreliability and infinite possibilities simultaneously. It is worth noting that the input of images is confined to raster data, which was unsatisfactory for this paper owing to the introduction of LiDAR data.
Alexnet, the champion of ILSVRC2012, is known as a breakthrough in CNN, which was why it was selected in this experiment. Five convolutional layers, three pooling layers and two fully-connected layers are contained in Alexnet. Numerous filters are embodied in convolutional layers to extract features. Rectified linear units (ReLU) is adopted as the activation function to constrain the range of values. In Equation (12), the ReLU activation function is presented.
a i , j , k = m a x I n p i , j , k , 0
where I n p i , j , k is the function input at location i , j on k th channel. An appropriate loss function is demanded for a specific experiment. For this paper, the softmax loss was chosen. Figure 6 depicts the architecture of Alexnet.

2.4. Region-Based Majority Voting

Object-based classification results are based on and considerably influenced by segmentation results. For object-wise CNN, casting CNN training outcomes into segmentation results is unavoidable. RMVCNN is a strategy connecting MRS results and CNN training results. Majority voting strategy consists of two parts, i.e., voters generation and vote. Voters directly engage with CNN, while vote is more postprocessing.
The key thought of majority voting is that each segmented polygon shall contain odd points acting voters which are later labelled by CNN with an attribute. In other words, the final category of a segmented polygon is determined by labelled voters within. The ideal situation is that each polygon has odd voters, which prevents any possibility of a tie vote. Thus, voters generation generates odd points within each segmented polygon. On the basis of the sizes of objects, the number of voters shall accordingly vary. For example, objects, such as vehicles and ships, normally contain dozens of pixels and corresponding segmented polygons are of the small size. A single voter is sufficient for such objects. Whereas, larger objects including buildings require more voters to determine categories. Therefore, all segmented polygons are graded according to size before generating voter(s). To ensure the validity of voter(s), the centre point of polygon and even random points within are required. The number of even random points is decided by the grade of each polygon in size.
Vote refers to rendering segmentation results with voters’ labelled attributes. Each polygon and its voter(s) shall be spatially joined before the vote. The procedure of vote is demonstrated in Figure 7.
There are five voters within the shown polygon, two of which are labelled as road while the other three ones are building. In the light of vote principle, the category of this polygon is then set as building. The adoption of majority voting effectively decreases the chances of misclassification caused by segmentation results. Meanwhile, multiple voters are more persuasive than one centre point to determine the category of an object.

2.5. Accuracy Assessment Methods

The accuracy assessment of this classification experiment is based on two essential indexes, namely overall accuracy (OA) and kappa coefficient. These two indexes can be obtained from the confusion matrix which is illustrated in Table 1.
For Table 1, x t r u e distributed on the diagonal is the number of accurately predicted object A, while y t r u e and z t r u e are respectively that of object B and object C. Unfilled blocks are mis-predicted numbers of each object. Thus, overall accuracy P o and kappa coefficient are acquired as Equations (13) and (14).
P o = x t r u e + y t r u e + z t r u e n
k = P o P e 1 P e
where n is the sum of objects and P e is the theoretical agreement rate which can be calculated as Equation (15).
P e = N a c t × N p r e n 2
where N a c t is the actual number and N p r e is the predicted number of each object. The kappa coefficient is confined to 1 , 1 but is commonly above zero. Moreover, it consists of five levels, i.e., slight (0.0–0.2), fair (0.21–0.4), moderate (0.41–0.6), substantial (0.61–0.8) and almost perfect (0.81–1).
Apart from the accuracy assessment, information entropy (IE) was introduced to assess and compare nDSM and StdnDSM. Referenced from the second law of thermodynamics, the information entropy in Information Science was proposed by C. E. Shannon [51,52] to assess the amount of information contained in a certain information source. Information entropy is actually the numerical statement of information source and is related to the probability of random events which in raster image refers to single pixels. The definition of information entropy is represented in Equation (16):
H X = C i = 1 n p x i log p x i .

2.6. Experiment Description

This section consists of two major parts, Section 3.1 Experiment Description and 3.2 Experiment Result and Comparison. In this section, the study area and data of this experiment, how sampling and stratified MRS were performed and CNN tuning and training are introduced in detail. The classification result of the experiment is shown and compared in Section 3.2. Furthermore, it should be noted that this experiment was performed on a machine learning machine equipped with Window10 OS and two NVIDIA TITAN X 12 Gb GPU. Alexnet was chosen as the fundamental CNN frame for its high performance in object classification and built on Tensorflow 1.7.0. Data processing were performed on professional software including QGIS (mapping and analysing), ArcGIS (mapping), ENVI (nDSM), LiDAR360 (DTM and DSM generating) and eCognition (segmentation).

2.6.1. Image Description and Data Fusion

The study area was a district in Helsinki, the capital city of Finland, which also is one of the biggest port city in Europe. The comprehensiveness of Helsinki ensured that the study area covered nearly all object categories. Meanwhile, VHSRIs and LiDAR data were provided by the National Land Survey of Finland (NLSF) and is demonstrated in Figure 8. The image size is 5995 × 5995.
It should be noted that the coordinate system of both VHSRI and point cloud data are ETRS-TM35FIN. The VHSRI was captured in August 2014 by an airborne sensor and uploaded to NLSF on 9 February 2017. As the image was sensed in summer, vegetations were spectrally easy to be distinguished and objects such as snowfield were avoided. The image resolution was 0.5 m/pixel. The point cloud data was produced on 22 August 2008 by airborne sensor Optech ALTM GEMINI and uploaded to NLSF on 16 December 2016. The elevation precision was 0.15 m while the average point density (resolution) was 43 pis/ m 2 . The inconsistency in time phase brought challenges in data fusion. However, such difficulties could be overcome thanks to the world’s leading urbanization in Helsinki. In other words, although the imaging time of VHSRI and PC are eight years apart, the city appearance of study area barely varies. Therefore, data fusion in this experiment was feasible and worth trying. Nevertheless, the high resolution of both VHSRI and PC ensured the possibility of land cover wise classification.
As the input of Alexnet is confined to raster images, PC needed to be transformed and then fused with VHSRI. Cloud points data was first transformed into DTM and DSM to further generate nDSM. This operation was performed on LiDAR360 software. In the processing, the cell assignment type was set by inverse distance weighting (IDW) and nearest neighbour method was used as void fill algorithm. The proposed StdnDSM which resolves the unequal weight phenomenon of directly fused VHSRI and PC was acquired based on nDSM. Figure 9 compares the StdnDSM and nDSM shown by a single band image.
The difference between those two was hardly visible but was reflected in the final results and information entropy assessment shown following. Assessments on StdnDSM, nDSM and corresponding fusion data were performed, whose outcome shown in Figure 10 suggested that StdnDSM contained greater amount of information compared to nDSM and StdnDSM fusion data beats nDSM fusion data. Moreover, it should be noted that higher IE does not necessarily indicate better performance in classification, but indeed proved the superiority of StdnDSM.
Meanwhile, the advantage of StdnDSM was proven from the data fusion mechanism. Once the StdnDSM was generated, three bands were selected from band R, G, B, and StdnDSM for Alexnet supported only three channels. To choose the ideal band combination, principal components analysis (PCA) was adopted for its ability to select bands with the maximum amount of information. The result of the PCA is shown in Table 2, according to which band R, band G and StdnDSM were selected for PC data fusion.
After the optimum band combination was confirmed, data fusion was performed whose result is demonstrated in Figure 11 of which the RGB channels respectively correspond to red, green and StdnDSM. It should be emphasized that Figure 11 was the input to Alexnet

2.6.2. Sampling for Training and Testing

Sampling is of great significance in semi-supervised classification as the classifier is not capable of rendering an attribute to a to-be classified object independently. In other words, sampling is the procedure where hidden attributes and learning objects are linked. Without proper training samples, CNN fails to distinguish different objects or is likely to be confused and then misclassify objects. After all, human–machine coupling is believed to be the future of AI. Apart from appropriate training samples, good validating samples are also demanded as the final OA is based on them. To ensure the reliability of OA, training samples and validating samples are required to abide by the same sampling criterion and strictly avoid repeated samples. If the same samples are adopted as training samples and validating samples simultaneously, the obtained OA will inflate compared to the genuine OA.
To meet above requirements, training samples and validating samples of this experiment were selected at the same time. To be specific, samples were distributed into training samples and validating samples after the sampling had been done. Therefore, this strategy ensured that there were no points that were concurrently training and validating points. According to the peculiarity of the study area, the categories of land-cover were classified into nine sorts, namely “building”, “dock”, “road”, “tree”, “water”, “ship”, “vehicle”, “shadow” and “bare land”. Each category of samples was selected individually so that different categories could be independently and purposefully fused for contrast experiments. The selected samples are demonstrated in Figure 12 and Table 3. As shown in Figure 12, red points represented training samples and consisted of 25,000 samples, while yellow points denoted validating samples and were of 3811 samples.

2.6.3. Stratified MRS

As mentioned above, the introduction of GLCM is an outcome of stratification strategy. As demonstrated in Figure 13, four homogenous regions respectively contained a type of dominating objects. The adoption of MRS tremendously decreases the number of segmented objects in segmentation results compared to traditional segmentation methods. However, there still remain the predicament that diverse objects share the same segmentation scale, which makes the segmentation result less justified theoretically. It is apparent that the best segmentation scale of vehicles and buildings should not be the same. Nine categories of land-cover objects interpenetrated in the study area. However, the distribution of objects was not completely random. Some similar objects tended to cluster while other objects were hardly possible to exist simultaneously. For example, vehicles could never be within water areas but were always close to roads. In other words, the first law of geography still applies well here. Thus, a region partition on coarse scale before MRS was necessary for a better segmentation result. In this experiment, the GLCM was performed to add texture information to the pre-segmentation and pre-segmented the study area into four large homogenous regions. As mentioned previously, five indexes, namely mean, variance, homogeneity, dissimilarity and entropy were chosen and then fused with band red, green, blue and StdnDSM for pre-segmentation. The reason for selecting those five filters was that they contained more information than other filters visually and statistically. By using GLCM, four homogenous regions respectively containing a type of dominating object were partitioned as demonstrated in Figure 13.
In the north-west region (Region 1), trees were the most distributed objects while water dominated the south-east region (Region 4). According to the sizes of dominating objects in each large homogenous region, different segmentation scales were chosen for the four regions. The best segmentation scales of all four regions are respectively shown in Table 4.
The best segmentation scales of each large homogenous region were the result of trial and error. Adequate trials were performed to guarantee that the best scales adopted were reliable. However, it should be emphasised that the band ratio of Region 4 in MRS was set as 1:1:1:1 which was different from other areas. The reason behind such a decision was that the elevation of waters was less significant in segmentation. The greatest advantage of adopting GLCM for a pre-segmentation was making MRS more theoretically explainable and efficient. In the subsequent section “Classification Results and Overall Accuracy”, a detailed comparison of adopting and non-adopting GLCM MRS will be presented in detail. Within each segmented object (polygon), several odd voters (voting points) were then generated for majority voting. The total amount of voters was 278,836. Thanks to the stratification strategy, the number of voters shrank by hundreds of thousands, which will be presented in Section 3.2.

2.6.4. CNN Training and Classification

As one of the most mature CNNs, the Alexnet was adopted in this experiment. There are several critical parameters that require detail stated. Above all, the learning rate controlling learning process was set as 0.01, as it is commonly adopted by tremendous experts in RS. Then, the number of epochs which refer to the procedure that all data travel through the workflow of CNN for one time was set as 160 as the training accuracies of experimented CNNs tended to be saturated after 150th epochs and more epochs would be simply pointless. Thirdly, the number of batches which evenly divided training samples into certain subsets as the amount of samples was too great to be directly put into CNN was set as 100. To avoid over-fitting, the dropout rate of Alextnet was set as 0.5. Eventually, as this paper considered scale effect in classification, experiments at more than one scale were performed. The scale of training was determined by the window size of CNN. To guarantee the comprehensiveness of this paper, four single scales and ten related scale combinations were independently performed. Four single scales were respectively 15, 25, 35 and 45. Ten scale combinations were 15–25, 15–35, 15–45, 25–35, 25–45, 35–45, 15–25–35, 15–25–45, 15–35–45 and 25–35–45. Results of each training scale will be demonstrated in Section 3.2.
Apart from the parameters discussed above, it should be also noted that the data ratio and training ratio were both set as 0.8. In other words, 20,000 out of 25,000 training samples were randomly input into CNN and 16,000 out of 20,000 training samples were directly taken as training data. The remaining 4000 training samples were used for CNN self-regulation. Once CNN training was done, a tuned network was obtained for further classification.
As explained previously, the fused data was first segmented into polygons, in which voters were generated for rendering attributes. Therefore, the classification of segmented objects was actually determined by the classification of voters. Once voters were input into tuned Alexnet, the network classified by rendering an attribute to each voter according to the features it learned. Then, voters with attributes were utilized to determine the attribute of each segmented and to-be classified objects. As each polygon contained multiple voters, the final attribute of it was democratic and reasonable.
Moreover, the comparative experiment where the B channel was nDSM was performed along with the experiment above. The corresponding results of comparative experiments are to be introduced in detail in Section 3.2 Classification Results and Accuracy.

3. Results

As together segmentation and classification for segmented objects composed the object-based CNN classification, comparisons will be demonstrated from three aspects namely, results, OA and the efficiency. Efficiency is about segmentation, while results and OA are related to classification.

3.1. Efficiency

Improvements in the efficiency of the introduced method were mainly reflected in segmentation. As the stratified MRS was a GEOBIA method, segmentation was of great significance in final classification results. Therefore, a stratified MRS method was proposed to enhance segmentation in both efficiency and fineness. The efficiency of the introduced method is reflected by statistics in Table 5 while the fineness is shown in Figure 14.
As Table 5 shows, stratified MRS of fusion data generated least both polygons and voters among all three methods, which indicated the superiority of both stratification strategy and fusion data (added LiDAR data). The voters generating principle was strictly region oriented. To be specific, polygons sized from 0–50 m2 contained only one voter (centre point), polygons sized from 50–100 m2 contained three voters, polygons sized from 100–500 m2 contained five voters and polygons that were large than 500 m2 contained seven voters. Such a principle was believed as both efficient and rational. It was unnecessary to specify polygons sized from 100–500 m2, as they were basically same or solo similar objects. Moreover, the comparison of unstratified MRS of fusion data and VHSRI greatly showed the advantage of point cloud which reflected elevation information. Polygons generated by directly segmenting fusion data was merely half the number of polygons obtained by segmenting VHSRI and the number of voters decreased by over 43 thousand. It should be noted that the parameters of these two unstratified MRS methods were exactly the same. With different voters generating principles, specific voter numbers of diverse methods may vary. However, the improvement in efficiency of introduced stratified MRS method was solid and nonnegligible.
Figure 14 visually demonstrates the superiorities of the stratified MRS and fusion data. Figure 14a shows the adopted segmentation method, that is stratified MRS of fusion data, whose segmentation result was the finest among those three methods, especially around the borders of large regions mentioned above. Demonstrated regions located at the border of the water region and the mixed region which consisted of almost all types of objects. It was apparent that for Figure 14b,c, segmentation results of unstratified MRS of fusion data, were not as good as the adopted method. To be specific, Figure 14e denoted that for objects of high homogeneity such as roads and waters, the addition of LiDAR surpassed single-source data in efficiency. That is, segmentation result of green lines was obviously less dense than that of red lines. Furthermore, Figure 14f showed that the adopted stratified MRS performed better than direct MRS, as the result of black lines was more efficient than that of green lines.

3.2. Classification Results and Accuracy

In this section, the final results of classification are to be shown and compared. Meanwhile, OAs of all classification results will be compared.
The classification results are shown and compared in Figure 15 and Figure 16.
As shown in Figure 15, the visual effects of four single scale classification results were mainly alike. However, there were still minor differences among them. In Figure 15a (scale 15), some water areas were misclassified into shadows while some shadow areas were misclassified as buildings in Figure 15b (scale 25). In Figure 15c (scale 35) and Figure 15d (scale 45), misclassifications could hardly be seen visually. It should be noted that scale 15 performed best in classifying ships. Moreover, classification results of trees, roads, bare lands and docks were basically similar in all four single scales. The training accuracies of each single scale were respectively 0.876 (scale 15), 0.918 (scale 25), 0.932 (scale 35) and 0.916 (scale 45). Apparently, the training accuracy generally accorded with the classification results shown in Figure 15. However, the final OA and kappa coefficient presented a different result. The OAs of single scale 15 to scale 45 were respectively 0.9016 (scale 15), 0.8719 (scale 25), 0.8856 (scale 35) and 0.8932 (scale 45), while the kappa coefficients were 0.9016 (scale 15), 0.8719 (scale 25), 0.8856 (scale 35) and 0.8932 (scale 45). The finest outcome came from scale 15 while scale 35 owned the highest training accuracy. The cause for such a situation may be the randomness of voters.
The result of multiscale 15–45 which generated the finest classification result among all single scales and their combinations is as demonstrated in Figure 16 along with the original image.
As Figure 16 shows, the classification results of scale 15–45 was basically identical to the ground truth and remained as the optimal among all fourteen scales (or scale combinations). To be specific, the training accuracy of scale 15–45 was 0.9441 and the OA was 0.9095 which was the highest obtained. In Figure 17, the confusion matrices of the top four classification results are demonstrated.
In Table 6 and Figure 18, the training accuracies, OAs and kappa coefficients of all training scales are demonstrated.
As Table 6 and Figure 17 indicate, the trend that training accuracies of single scales were lower than contacted multi-scales is apparent, by which overall accuracies abided. Generally, performances of contacted multi-scales were better than that of single scales by a slight margin. Moreover, a VHSRI was input into the tuned network as well and obtained corresponding training accuracies at four single scales: 0.85 (scale 15), 0.9 (scale 25), 0.92 (scale 35) and 0.9 (scale 45).
The final classification OA of nDSM–VHSRI fusion data were as follow: single-scale 15 was 0.814, single-scale 25 was 0.884, single-scale 35 was 0.886, single-scale 45 was 0.89 and the finest result which was obtained in multi-scale 25–35–45 was 0.904. Despite the excellent result generated by nDSM–VHSRI fusion data, StdnDSM fusion data still beat it in OA.

4. Discussion

Influences of diverse sampling methods and different parameters in CNN on classification results and computing efficiency will be shown and discussed in this part. Furthermore, superiorities and imperfections of adopted fusion data and stratified multiresolution segmentation will be talked over as well here.

4.1. Influences of Sampling Strategy and CNN Parameters

4.1.1. Sampling Strategy

As introduced in Section 2.6.2., this paper adopted the strategy of sampling training and validating points simultaneously. To be specific, samples were randomly divided into training points and validating points after sampling had been done. Such a sampling strategy ensured that training points and validating points abided by the exact same sampling principle. The advantage of this sampling strategy was that no training points would accidentally coincide with validating points and therefore avoided extortionately high overall accuracy. Different sampling strategies were performed and corresponding accuracies of each land-cover categories are compared in Table 7. More specifically, the adopted sampling strategy contained 25,000 training points and 4811 validating points that were sampled at the same time and randomly generated (synchronous sampling), while the contrast sampling strategy consisted of 25,000 training points and 2850 validating points that were asynchronously sampled (asynchronous sampling).
Table 7 shows that accuracies of asynchronous sampling strategy in both individual categories and overall samples generally tended to be numerically higher than that of synchronous sampling strategy. However, greater numbers do not necessarily represent better performance. As training samples and validating samples were asynchronously sampled, the possibility that those two types of samples may have coincided and sampling principles may not be exactly the same. If, however, one of those situations occured, the final accuracies may have varied accordingly. In this paper, asynchronous sampling generated validating samples that coincided with or near training points, which consequently increased final accuracies in error. Therefore, the final outcomes of synchronous sampling may be less pretty than that of asynchronous, but the rationality and reliability of synchronous sampling were superior.

4.1.2. CNN Parameters Setting

Alexnet contains numerous parameters that affect training and final classification results. However, only several parameters, namely, batch_size, epoch_number and data_size, are discussed in detail here. While learning_rate and drop_rate were respectively set as 0.01 and 0.5, as they are commonly adopted empiric values.
Batch_size is the number of divided images put into the network. Batch_size determines the speed of training which indirectly points to the efficiency of the network if training accuracies are the same. In our experiments, batch_size hardly impacted training accuracies of the network. Therefore, a larger batch_size here meant a higher training efficiency. However, the setting of batch_size was not the higher the better, as the increase of batch_size theoretically did lower training accuracies. Batch_size was eventually set as 100, as the training accuracies stabilized at 0.84–0.94, while the training accuracies of batch_size 50, 150 and 200 were unstable and lower.
Epoch_number is the times the workflow of the network runs throughout the training. More specifically, epoch_number determines how many times the CNN will learn the training data and directly decides the value of training accuracy. Generally, training accuracy rises as epoch_number increases till a peak and slightly drops afterwards. Therefore, the final epoch_number was set as 160 after trial and error. Experiments setting epoch_numbers as 50, 100, 150 and 200 were performed as well, yet training results indicated that the maximum training accuracy emerged around 160 epochs.
Data_size is the scale of training or the size of the training sample. In the procedure of the designed experiment, training samples were selected as points, whereas CNN only accepts raster data (small images). Thus, a mutable window centred on sampled training points generated the required data for CNN. As demonstrated in Section 2.6.4, four single sizes (scales) were generated and contacted later. The train and loss curve of the lowest single scale (scale 15), highest single scale (scale 35) and highest multiscale (scale 15–35–45) are demonstrated as follows in Figure 19.
The training accuracy of scale 15 (Figure 19a) was the lowest of all scales and was 0.876. The highest single scale training accuracy was 0.932 from scale 35 (Figure 19b). Figure 19c shows the highest training accuracy of all scales, 0.9508 from scale 15–35–45. In the light of the training accuracies shown in Table 6 and curves shown in Figure 19, it is apparent that training accuracy increased with the rise of training scale and scale combinations obviously overshadow single scale in accuracy.

4.2. Pros and Cons of Point Cloud added Fusion Data and Stratified MRS

As listed in previous sections, superiorities of fusing point clouds (LiDAR data) and VHSRIs (RGB image) by the proposed StdnDSM lie in two main aspects: (1) fusion data contained elevation information which RGB images failed to comprise and (2) StdnDSM evened the ratio of fusion bands and therefore enhanced the amount of information.
The addition of elevation feature encouraged boundary differentiation in both segmentation and training, which were reflected in segmentation quality and training accuracies. The boundary differentiation in segmentation of PC added and non-added data are demonstrated in Figure 20.
As Figure 18 clearly shows that PC added data (StdnDSM fused) surpassed PC non-added data (RGB image) in the segmentation results. Boundaries of buildings and trees where distinct differences in height existed most effectively tell the advantage of PC added data. However, the elevation was not the answer to all problems. Objects like road and bare land were not easily distinguished through StdnDSM and the fusion data was not visual-friendly. Most importantly, the synchrony of point cloud and VHSRIs was hardly guaranteed as such open-sourced data is rare. The data used in this experiment were asynchronous and may be controversial for fusion owing to the inconsistency of moving objects, namely vehicles and ships. However, such objects should not be neglected in classification and most objects, such as buildings, docks, waters and roads, remained basically the same.
Moreover, combining PC and VHSRIs surprisingly mitigated the scale effect in traditional RS image analysis. The possible cause of such a phenomenon was that the addition of elevation defined the boundaries of objects that traditional RGB images failed to specify. However, this explanation is yet to be confirmed and remains a future research subject.
When it comes to stratified MRS, the superiorities of Stratified MRS have been emphasized more than once previously. Above all, the theoretical rationality and computational high efficiency are two unprecedented traits compared to traditional MRS. However, stratified MRS has its own flaws that performing stratified MRS rather complex and cross-platform, while traditional MRS can be easily done on any specialized software. A refined stratified MRS or involving parallel stratified MRS is still a task worth studying in the future.

5. Conclusions

Land-cover classification has always been a critical task in remote sensing. Thanks to the rapid developments in RS sensors, massive images of sub-meter resolution are obtained every day, which poses the challenge to automatically and efficiently conduct those data. Therefore, CNN was applied in image classification and proved as one of the finest classifiers. For land cover classification by using images with spatial resolution of meter level or sub-meter level, the designed methodology contained the novel proposed StdnDSM which added elevation information into CNN and adopted stratified multiresolution segmentation and majority voting who both significantly improved the computation efficiency. Conclusions concerning StdnDSM, stratified multiscale segmentation and RMVCNN are drawn as follows.
The proposed StdnDSM was actually an improved nDSM which rendered a smoother fusion of LiDAR data and VHSRIs. The smooth fusion of two types of data balanced the influences of LiDAR data and VHSRIs in segmentation and classification. Compared to preceding fusion data or single-sourced data, the fusion of StdnDSM and VHSRIs was overall theoretically and practically better. Moreover, as mentioned in Section 4.2., the combination of point cloud and VHSRIs alleviated the influences of scale effect in segmentation and classification. Yet, the mechanism behind such a phenomenon is yet to be studied.
Stratified MRS was actually a segmentation strategy which divided an image into several homogenous regions for further segmentation. The stratified MRS was utilized in land-cover classification for the first time here and generated a promising result. The introduction of stratified MRS significantly heightened the segmentation performance in boundaries and efficiency. Thus, adopting the stratification strategy in segmentation was logical and greatly beneficial to object-based classification.
Majority voting strategy is the indispensable part in object-based classification which connects the results of segmentation and CNN. Compared to currently and widely used centre point-oriented strategy which directly renders the category of the centre point to segmented polygon, majority voting avoids misclassifications caused by misclassifying centre points. Therefore, it had better robustness and higher fault-tolerance.
All in all, fusing LiDAR data and VHSRIs is feasible and effective for both image segmentation and land-cover object classification. However, the asynchrony of capture is yet to be addressed. The adopted Alexnet is of high-performance in image classification, nevertheless, new networks, such as graph neural network (GNN), also show great potential and need further exploration. Moreover, the introduction of stratified MRS in the land-cover classification generated promising results and showed impressive efficiency. However, a convincing assessment criteria system for MSR is not yet established and requires further efforts.

Author Contributions

Conceptualization, K.Z. and D.M.; methodology, K.Z.; software, K.Z. and X.L.; validation, K.Z. and J.F.; formal analysis, K.Z.; investigation, K.Z.; resources, K.Z.; data curation, K.Z. and J.F.; writing—original draft preparation, K.Z.; writing—review and editing, K.Z. and D.M.; visualization, K.Z.; supervision, M.W.; project administration, D.M.; funding acquisition, M.W.

Funding

This research was funded by National Natural Science Foundation of China, grant number 41671369 and 41671341, National Key Research and Development Program, grant number 2017YFB0503600 and ‘’the Fundamental Research Funds for the Central Universities’’.

Acknowledgments

Special thanks to the National Land Survey of Finland (NLSF) for providing data. Last but not least, this paper is not the work of a single person but of a great team including members of Lab Ming of China University of Geosciences (Beijing), and some close friends who have helped a lot.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef] [Green Version]
  2. Minar, M.R.; Naher, J. Recent Advances in Deep Learning: An Overview. arXiv 2018, arXiv:1807.08169. [Google Scholar]
  3. Vanderhoof, M.K.; Lane, C.R. The potential role of very high-resolution imagery to characterise lake, wetland and stream systems across the Prairie Pothole Region, United States. Int. J. Remote Sens. 2019, 40, 5768–5798. [Google Scholar] [CrossRef]
  4. Ming, D.; Zhou, T.; Wang, M.; Tan, T. Land cover classification using random forest with genetic algorithm-based parameter optimization. J. Appl. Remote Sens. 2016, 10, 35021. [Google Scholar] [CrossRef]
  5. Ming, D.; Li, J.; Wang, J.; Zhang, M. Scale parameter selection by spatial statistics for GeOBIA: Using mean-shift based multi-scale segmentation as an example. ISPRS J. Photogramm. Remote Sens. 2015, 106, 28–41. [Google Scholar] [CrossRef]
  6. Chen, Y.; Ming, D.; Zhao, L.; Lv, B.; Zhou, K.; Qing, Y. Review on High Spatial Resolution Remote Sensing Image Segmentation Evaluation. Photogramm. Eng. Remote Sens. 2018, 84, 629–646. [Google Scholar] [CrossRef]
  7. Lv, Z.Y.; Shi, W.Z.; Benediktsson, J.A.; Ning, X.J. Novel Object-Based Filter for Improving Land-Cover Classification of Aerial Imagery with Very High Spatial Resolution. Remote Sens. 2016, 8, 1023. [Google Scholar] [CrossRef]
  8. Cui, G.; Lv, Z.; Li, G.; Benediktsson, J.A.; Lu, Y. Refining Land Cover Classification Maps Based on Dual-Adaptive Majority Voting Strategy for Very High Resolution Remote Sensing Images. Remote Sens. 2018, 10, 1238. [Google Scholar] [CrossRef]
  9. Lv, X.W.; Ming, D.; Chen, Y.Y.; Wang, M. Very high resolution remote sensing image classification with SEEDS-CNN and scale effect analysis for superpixel CNN classification. Int. J. Remote Sens. 2019, 40, 506–531. [Google Scholar] [CrossRef]
  10. Shao, Z.; Fu, H.; Li, D.; Altan, O.; Cheng, T. Remote sensing monitoring of multi-scale watersheds impermeability for urban hydrological evaluation. Remote Sens. Environ. 2019, 232, 111338. [Google Scholar] [CrossRef]
  11. Hay, G.J.; Blaschke, T. Special Issue: Geographic Object-Based Image Analysis (GEOBIA) Foreword. Photogramm. Eng. Remote Sens. 2010, 76, 121–122. [Google Scholar]
  12. Chen, Y.; Ming, D.; Lv, X. Superpixel Based Land Cover Classification of VHR Satellite Image Combining Multi-Scale CNN and Scale Parameter Estimation. Earth Sci. Inform. 2019. [Google Scholar] [CrossRef]
  13. Lv, X.; Ming, D.; Lu, T.; Zhou, K.; Wang, M.; Bao, H. A New Method for Region-Based Majority Voting CNNs for Very High Resolution Image Classification. Remote Sens. 2018, 10, 1946. [Google Scholar] [CrossRef]
  14. Zhou, W.; Ming, D.; Xu, L.; Bao, H.; Wang, M. Stratified Object-Oriented Image Classification Based on Remote Sensing Image Scene Division. J. Spectrosc. 2018, 2018, 1–11. [Google Scholar] [CrossRef] [Green Version]
  15. Xu, L.; Ming, D.P.; Zhou, W.; Bao, H.Q.; Chen, Y.Y.; Ling, X. Farmland Extraction from High Spatial Resolution Remote Sensing Images Based on Stratified Scale Pre-Estimation. Remote Sens. 2019, 11, 108. [Google Scholar] [CrossRef]
  16. Ming, D.; Zhang, X.; Zhou, W.; Wang, M. Cropland Extraction Based on OBIA and Adaptive Scale Pre-estimation. Photogramm. Eng. Remote Sens. 2016, 82, 635–644. [Google Scholar] [CrossRef]
  17. Shao, Z.; Pan, Y.; Diao, C.; Cai, J. Cloud Detection in Remote Sensing Images Based on Multiscale Features-Convolutional Neural Network. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4062–4076. [Google Scholar] [CrossRef]
  18. Ghorbanzadeh, O.; Tiede, D.; Wendt, L.; Sudmanns, L.; Lang, S. Convolutional Neural Network (CNN) for Dwelling Extraction in Refugee/IDP Camps. In Proceedings of the 39th EARSeL Symposium Digital I Earth I Observation, Salzburg, Austria, 1–4 July 2019. [Google Scholar]
  19. Hinton, G.E.; Salakhutdinov, R.R. Reducing the Dimensionality of Data with Neural Networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [Green Version]
  20. Lecun, Y.; Bottou, L.; Haffner, P. Gradient-Based Learning Applied to Document Recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  21. Rezaee, M.; Mahdianpari, M.; Zhang, Y.; Salehi, B. Deep Convolutional Neural Network for Complex Wetland Classification Using Optical Remote Sensing Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 3030–3039. [Google Scholar] [CrossRef]
  22. Khamparia, A.; Singh, K.M. A systematic review on deep learning architectures and applications. Expert Syst. 2019, 36, e12400. [Google Scholar] [CrossRef]
  23. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  24. Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.; Zhang, L.P.; Xu, F.; Fraundorfer, F. Deep Learning in Remote Sensing: A Comprehensive Review and List of Resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef] [Green Version]
  25. Scott, G.J.; England, M.R.; Starms, W.A.; Marcum, R.A.; Davis, C.H. Training Deep Convolutional Neural Networks for Land–Cover Classification of High-Resolution Imagery. IEEE Geosci. Remote Sens. Lett. 2017, 14, 549–553. [Google Scholar] [CrossRef]
  26. Ming, D.; Zhou, W.; Xu, L.; Wang, M.; Ma, Y. Coupling Relationship among Scale Parameter, Segmentation Accuracy, and Classification Accuracy in GeOBIA. Photogramm. Eng. Remote Sens. 2018, 84, 681–693. [Google Scholar] [CrossRef]
  27. Han, S.; Liu, X.; Mao, H.; Pu, J.; Pedram, A.; Horowitz, M.; Dally, B. Deep compression and EIE: Efficient inference engine on compressed deep neural network. In Proceedings of the 2016 IEEE Hot Chips 28 Symposium (HCS), Cupertino, CA, USA, 21–23 August 2016; pp. 1–6. [Google Scholar]
  28. Hong, Z.; Ming, D.; Zhou, K.; Guo, Y.; Lu, T. Road Extraction From a High Spatial Resolution Remote Sensing Image Based on Richer Convolutional Features. IEEE Access 2018, 6, 46988–47000. [Google Scholar] [CrossRef]
  29. Lu, T.; Ming, D.; Lin, X.; Hong, Z.; Bai, X.; Fang, J. Detecting Building Edges from High Spatial Resolution Remote Sensing Imagery Using Richer Convolution Features Network. Remote Sens. 2018, 10, 1496. [Google Scholar] [CrossRef]
  30. Mahdianpari, M.; Salehi, B.; Rezaee, M.; Mohammadimanesh, F.; Zhang, Y. Very Deep Convolutional Neural Networks for Complex Land Cover Mapping Using Multispectral Remote Sensing Imagery. Remote Sens. 2018, 10, 1119. [Google Scholar] [CrossRef]
  31. Maggiori, E.; Tarabalka, Y.; Charpiat, G.; Alliez, P. Convolutional Neural Networks for Large-Scale Remote-Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 645–657. [Google Scholar] [CrossRef]
  32. Ni, H.; Lin, X.; Ning, X.; Zhang, J. Edge Detection and Feature Line Tracing in 3D-Point Clouds by Analyzing Geometric Properties of Neighborhoods. Remote Sens. 2016, 8, 710. [Google Scholar] [CrossRef]
  33. Ghorbanzadeh, O.B.T.; Meena, S.R. Potential of Convolutional Neural Networks for Earthquake-triggered Mass Movement Detection Using Optical and SAR Data. In Proceedings of the Living Planet Symposium, MiCo—Milano Congressi, Milan, Italy, 13–17 May 2019. [Google Scholar]
  34. Xu, X.; Li, W.; Ran, Q.; Du, Q.; Gao, L.; Zhang, B. Multisource Remote Sensing Data Classification Based on Convolutional Neural Network. IEEE Trans. Geosci. Remote Sens. 2018, 56, 937–949. [Google Scholar] [CrossRef]
  35. Ghorbanzadeh, O.; Blaschke, T.; Gholamnia, K.; Meena, S.R.; Tiede, D.; Aryal, J. Evaluation of Different Machine Learning Methods and Deep-Learning Convolutional Neural Networks for Landslide Detection. Remote Sens. 2019, 11, 196. [Google Scholar] [CrossRef]
  36. Serifoglu, C.; Gungor, O.; Yilmaz, V. Performance evaluation of different ground filtering algorithms for uav-based point clouds. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 41, 245–251. [Google Scholar] [CrossRef]
  37. Bittner, K.; Adam, F.; Cui, S.; Korner, M.; Reinartz, P. Building Footprint Extraction From VHR Remote Sensing Images Combined With Normalized DSMs Using Fused Fully Convolutional Networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 2615–2629. [Google Scholar] [CrossRef] [Green Version]
  38. Zhang, X.; Du, S.; Wang, Q. Hierarchical semantic cognition for urban functional zones with VHR satellite images and POI data. ISPRS J. Photogramm. Remote Sens. 2017, 132, 170–184. [Google Scholar] [CrossRef]
  39. Liu, Y.; Liu, X.; Gao, S.; Gong, L.; Kang, C.G.; Zhi, Y.; Chi, G.H.; Shi, L. Social Sensing: A New Approach to Understanding Our Socioeconomic Environments. Ann. Assoc. Am. Geogr. 2015, 105, 512–530. [Google Scholar] [CrossRef]
  40. Yilmaz, V.; Gungor, O.; Yılmaz, V. Determining the optimum image fusion method for better interpretation of the surface of the Earth. Nor. Geogr. Tidsskr. Nor. J. Geogr. 2016, 70, 1–13. [Google Scholar] [CrossRef]
  41. Saralıoğlu, E.; Görmüş, E.T.; Güngör, O. Mineral exploration with hyperspectral image fusion. In Proceedings of the 2016 24th Signal Processing and Communication Application Conference (SIU), Zonguldak, Turkey, 16–19 May 2016; pp. 1281–1284. [Google Scholar]
  42. Nahhas, F.H.; Shafri, H.Z.; Sameen, M.I.; Pradhan, B. Deep Learning Approach for Building Detection Using LiDAR–Orthophoto Fusion. J. Sens. 2018, 2018, 1–12. [Google Scholar] [CrossRef]
  43. Yilmaz, V.; Yilmaz, C.S.; Gungor, O. Genetic algorithm-based synthetic variable ratio image fusion. Geocarto Int. 2019, 1–18. [Google Scholar] [CrossRef]
  44. Pan, X.; Gao, L.; Marinoni, A.; Zhang, B.; Yang, F.; Gamba, P. Semantic Labeling of High Resolution Aerial Imagery and LiDAR Data with Fine Segmentation Network. Remote Sens. 2018, 10, 743. [Google Scholar] [CrossRef]
  45. Kobler, A.; Ogrinc, P. REIN Algorithm and the Influence of Point Ccloud Density on NDSM and DEM Precision in a Submediterranean Forest. In Proceedings of the SPRS Workshop on Laser Scanning 2007 and SilviLaser 2007, Espoo, Finland, 12–14 September 2007. [Google Scholar]
  46. Marceau, D.; Howarth, P.; Dubois, J.; Gratton, D. Evaluation Of The Grey-level Co-occurrence Matrix Method For Land-cover Classification Using Spot Imagery. IEEE Trans. Geosci. Remote Sens. 1990, 28, 513–519. [Google Scholar] [CrossRef]
  47. Haralick, R.M.; Shanmugam, K.; Dinstein, I. Textural Features for Image Classification. IEEE Trans. Syst. Mancybern. 1973, 3, 610–621. [Google Scholar] [CrossRef]
  48. Puissant, A.; Hirsch, J.; Weber, C. The utility of texture analysis to improve per-pixel classification for high to very high spatial resolution imagery. Int. J. Remote Sens. 2005, 26, 733–745. [Google Scholar] [CrossRef]
  49. Baatz, M.; Schäpe, A. Multiresolution Segmentationan Optimization Approach for High Quality Multi-Scale Image Segmentation. Angew. Geogr. Inf. 2000, 12, 12–23. [Google Scholar]
  50. Zarandy, A.; Rekeczky, C.; Szolgay, P.; Chua, L.O.; Akos, Z. Overview of CNN research: 25 years history and the current trends. In Proceedings of the 2015 IEEE International Symposium on Circuits and Systems (ISCAS), Lisbon, Portugal, 24–27 May 2015; pp. 401–404. [Google Scholar]
  51. Seising, R. 60 years “A Mathematical Theory of Communication”-Towards a “Fuzzy Information Theory”. In Proceedings of the Joint 2009 International Fuzzy Systems Association World Congress and 2009 European Society of Fuzzy Logic and Technology Conference 2009, Lisbon, Portugal, 20–24 July 2009; pp. 1332–1337. [Google Scholar]
  52. Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The workflow for the convolutional neural network (CNN)-based land cover classification method combining stratified segmentation and point cloud data fusion.
Figure 1. The workflow for the convolutional neural network (CNN)-based land cover classification method combining stratified segmentation and point cloud data fusion.
Remotesensing 11 02065 g001
Figure 2. The comparison of point cloud (a) and standard normalized digital surface model (StdnDSM) (b). The site was located in Helsinki, Finland.
Figure 2. The comparison of point cloud (a) and standard normalized digital surface model (StdnDSM) (b). The site was located in Helsinki, Finland.
Remotesensing 11 02065 g002
Figure 3. The comparison of (a) digital terrain model (DTM), (b) digital surface model (DSM), (c) normalised DSM (nDSM) and (d) StdnDSM. The site was located in Helsinki, Finland.
Figure 3. The comparison of (a) digital terrain model (DTM), (b) digital surface model (DSM), (c) normalised DSM (nDSM) and (d) StdnDSM. The site was located in Helsinki, Finland.
Remotesensing 11 02065 g003
Figure 4. The comparison of fusion data. (a) DTM, (b) DSM, (c) nDSM and (d) StdnDSM. The site was located in Helsinki, Finland.
Figure 4. The comparison of fusion data. (a) DTM, (b) DSM, (c) nDSM and (d) StdnDSM. The site was located in Helsinki, Finland.
Remotesensing 11 02065 g004
Figure 5. A demonstration of gray-level co-occurrence matrix (GLCM). (a) The eight processing directions of GCLM, (b) the processing window moves and (c) the counted co-occurrence matrix.
Figure 5. A demonstration of gray-level co-occurrence matrix (GLCM). (a) The eight processing directions of GCLM, (b) the processing window moves and (c) the counted co-occurrence matrix.
Remotesensing 11 02065 g005
Figure 6. The architecture of Alexnet. Conv represents the convolutional layer; Pool denotes pooling layer; FulC is the fully-connected layer. Numbers underneath each image indicate corresponding size and dimension.
Figure 6. The architecture of Alexnet. Conv represents the convolutional layer; Pool denotes pooling layer; FulC is the fully-connected layer. Numbers underneath each image indicate corresponding size and dimension.
Remotesensing 11 02065 g006
Figure 7. Demonstration of majority voting.
Figure 7. Demonstration of majority voting.
Remotesensing 11 02065 g007
Figure 8. Experiment area: a scene in Helsinki.
Figure 8. Experiment area: a scene in Helsinki.
Remotesensing 11 02065 g008
Figure 9. The comparison of nDSM (a) and StdnDSM (b).
Figure 9. The comparison of nDSM (a) and StdnDSM (b).
Remotesensing 11 02065 g009
Figure 10. Information entropy demonstration of nDSM, StdnDSM, nDSM fusion and StdnDSM fusion.
Figure 10. Information entropy demonstration of nDSM, StdnDSM, nDSM fusion and StdnDSM fusion.
Remotesensing 11 02065 g010
Figure 11. The training image of Alexnet.
Figure 11. The training image of Alexnet.
Remotesensing 11 02065 g011
Figure 12. Demonstration of (a) training samples and (b) validating samples.
Figure 12. Demonstration of (a) training samples and (b) validating samples.
Remotesensing 11 02065 g012
Figure 13. Large homogenous regions.
Figure 13. Large homogenous regions.
Remotesensing 11 02065 g013
Figure 14. The comparison of stratified and unstratified multiresolution segmentation (MRS) of fusion data.
Figure 14. The comparison of stratified and unstratified multiresolution segmentation (MRS) of fusion data.
Remotesensing 11 02065 g014
Figure 15. The demonstration of four single scale classification results. From (ad) the scale is 15, 25, 35 and 45 respectively.
Figure 15. The demonstration of four single scale classification results. From (ad) the scale is 15, 25, 35 and 45 respectively.
Remotesensing 11 02065 g015
Figure 16. The demonstration of experiment area (a) and final classification result (b).
Figure 16. The demonstration of experiment area (a) and final classification result (b).
Remotesensing 11 02065 g016
Figure 17. The demonstration of confusion matrix. (a) Confusion matrix of multiscale 15–45, (b) 15–25–45, (c) 15–35–45 and (d) 25–45.
Figure 17. The demonstration of confusion matrix. (a) Confusion matrix of multiscale 15–45, (b) 15–25–45, (c) 15–35–45 and (d) 25–45.
Remotesensing 11 02065 g017
Figure 18. Demonstration of classification results.
Figure 18. Demonstration of classification results.
Remotesensing 11 02065 g018
Figure 19. The demonstration of train and loss curve. (a) The train and loss curve of scale 15, (b) scale 35 and (c) multiscale 15–35–45.
Figure 19. The demonstration of train and loss curve. (a) The train and loss curve of scale 15, (b) scale 35 and (c) multiscale 15–35–45.
Remotesensing 11 02065 g019
Figure 20. The comparison of point cloud (PC) added and non-added data in segmentation. Red line (a) represents RGB and black line (b) is PC added.
Figure 20. The comparison of point cloud (PC) added and non-added data in segmentation. Red line (a) represents RGB and black line (b) is PC added.
Remotesensing 11 02065 g020
Table 1. The illustration of the confusion matrix.
Table 1. The illustration of the confusion matrix.
Actual Class
A B C
Predicted Class A X Y 1 Z 1
B X 1 Y Z 2
C X 2 Y 2 Z
Table 2. The result of principal components analysis.
Table 2. The result of principal components analysis.
Correlation Matrix
LayerRGBnDSM
R10.07940.179830.11541
G0.079410.96680.93594
B0.179830.966810.97563
nDSM0.115410.935940.975631
Table 3. Amount of each category in training samples.
Table 3. Amount of each category in training samples.
CategoryBare LandBuildingDockRoadShadowShipTreeVehicleWater
Amount159468718822659294123446147659903
Table 4. The segmentation scales of each homogenous region.
Table 4. The segmentation scales of each homogenous region.
RegionsRegion 1Region 2Region 3Region 4
Parameters
Scale Parameter20202020
Shape0.70.30.50.4
Compactness0.30.70.50.6
Band Ratio1:1:1:71:1:1:71:1:1:71:1:1:1
Table 5. Comparison of stratified and unstratified multiresolution segmentation methods.
Table 5. Comparison of stratified and unstratified multiresolution segmentation methods.
Segmentation MethodsPolygon AmountVoters Amount
Stratified MRS of Fusion Data94,144278,836
Unstratified MRS of Fusion Data106,985297,227
Unstratified MRS of VHSRI202,364340,978
Table 6. The classification results of all scales.
Table 6. The classification results of all scales.
ScaleTraining AccuracyOverall Accuracy
150.8760.9016
250.9180.8719
350.9320.8856
450.9160.8932
15–250.92380.8767
15–350.94460.9
15–450.94410.9095
25–350.9390.889
25–450.94510.9021
35–450.9310.8971
15–25–350.94280.8969
15–25–450.94870.9034
15–35–450.95080.9034
25–35–450.89920.8992
Table 7. Accuracy comparison of synchronous and asynchronous sampling strategy.
Table 7. Accuracy comparison of synchronous and asynchronous sampling strategy.
CategoriesAccuracy by Synchronous SamplingAccuracy by Asynchronous Sampling
Bare Land90.65%92.43%
Building93.97%93.41%
Dock90.63%92.53%
Road86.4%88.92%
Shadow85.3%89.67%
Ship94.93%94.56%
Tree92.59%93.5%
Vehicle79.66%85.79%
Water93.37%95.62%
Overall90.95%91.53%

Share and Cite

MDPI and ACS Style

Zhou, K.; Ming, D.; Lv, X.; Fang, J.; Wang, M. CNN-Based Land Cover Classification Combining Stratified Segmentation and Fusion of Point Cloud and Very High-Spatial Resolution Remote Sensing Image Data. Remote Sens. 2019, 11, 2065. https://doi.org/10.3390/rs11172065

AMA Style

Zhou K, Ming D, Lv X, Fang J, Wang M. CNN-Based Land Cover Classification Combining Stratified Segmentation and Fusion of Point Cloud and Very High-Spatial Resolution Remote Sensing Image Data. Remote Sensing. 2019; 11(17):2065. https://doi.org/10.3390/rs11172065

Chicago/Turabian Style

Zhou, Keqi, Dongping Ming, Xianwei Lv, Ju Fang, and Min Wang. 2019. "CNN-Based Land Cover Classification Combining Stratified Segmentation and Fusion of Point Cloud and Very High-Spatial Resolution Remote Sensing Image Data" Remote Sensing 11, no. 17: 2065. https://doi.org/10.3390/rs11172065

APA Style

Zhou, K., Ming, D., Lv, X., Fang, J., & Wang, M. (2019). CNN-Based Land Cover Classification Combining Stratified Segmentation and Fusion of Point Cloud and Very High-Spatial Resolution Remote Sensing Image Data. Remote Sensing, 11(17), 2065. https://doi.org/10.3390/rs11172065

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop