Next Article in Journal
OSO-YOLOv5: Automatic Extraction Method of Store Signboards in Street View Images Based on Multi-Dimensional Analysis
Previous Article in Journal
Web Mapping and Real–Virtual Itineraries to Promote Feasible Archaeological and Environmental Tourism in Versilia (Italy)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Pattern Recognition and Segmentation of Administrative Boundaries Using a One-Dimensional Convolutional Neural Network and Grid Shape Context Descriptor

1
School of Resource and Environmental Sciences, Wuhan University, 129 Luoyu Road, Wuhan 430079, China
2
College of Surveying and Geo-Informatics, Tongji University, 1239 Siping Road, Shanghai 200092, China
*
Author to whom correspondence should be addressed.
ISPRS Int. J. Geo-Inf. 2022, 11(9), 461; https://doi.org/10.3390/ijgi11090461
Submission received: 16 June 2022 / Revised: 8 August 2022 / Accepted: 25 August 2022 / Published: 28 August 2022

Abstract

:
Recognizing morphological patterns in lines and segmenting them into homogeneous segments is critical for line generalization and other applications. Due to the excessive dependence on handcrafted features in existing methods and their insufficient consideration of contextual information, we propose a novel pattern recognition and segmentation method for lines, based on deep learning and shape context descriptors. In this method, a line is divided into a series of consecutive linear units of equal length, termed lixels. A grid shape context descriptor (GSCD) was designed to extract the contextual features for each lixel. A one-dimensional convolutional neural network (1D-U-Net) was constructed to classify the pattern type of each lixel, and adjacent lixels with the same pattern types were fused to obtain segmentation results. The proposed method was applied to administrative boundaries, which were segmented into components with three different patterns. The experiments showed that the lixel classification accuracy of the 1D-U-Net reached 90.42%. The consistency ratio was 92.41%, when compared with the manual segmentation results, which was higher than either of the two existing machine learning-based segmentation methods.

1. Introduction

In map space, various geographical entities are modeled as points, lines, polygons, or other element types. Lines are one of the most abundant elements, representing rivers, roads, administrative boundaries, and coastlines. Due to the heterogeneity inherent in geographical environments, lines typically differ in terms of their geometric properties and spatial structures. In particular, an individual line may have different shape pattern characteristics, which reveal the spatial distribution characteristics of the associated geographical entity. During the processing of line generalization, which simplifies the shape of a line by removing unwanted small details, it is important to recognize the morphological patterns of the lines and segment them into homogenous segments [1,2,3,4,5,6,7]. A simple example is shown in Figure 1, where a line consisting of two parts with different pattern characteristics has been simplified using two popular algorithms (i.e., the bend-based algorithm [8] and the orthogonality-preserving simplification algorithm [9]). The bend-based algorithm was appropriate for simplifying segment I with a complex hierarchy of bends, however, it generated an unsatisfying result for segment II with orthogonal characters. In contrast, the orthogonality-preserving algorithm performed well to maintain the orthogonality of segment II; however, it distorted the main bend structures of segment I. This example indicates that each line generalization algorithm has its own strengths and weaknesses for differing shape characteristics [10,11,12].
To overcome this drawback, one promising solution is to develop hybrid generalization methods which allow each part of the line to be processed using the appropriate algorithm. Specifically, for each input line, the shape characteristics of the line are first recognized and then divided into geometrically homogeneous segments. Next, the appropriate algorithm and parameter settings can be applied to each segment according to its pattern type. Since a large number of line generalization algorithms have been developed, methods for line pattern recognition and segmentation have become a focus of current research. This necessity has been highlighted by researchers during the generalization of various map features, including building outlines [11,13,14], administrative boundaries [12], road lines [15,16,17], coastlines [18], and land-use boundaries [5].
In the past few decades, some studies have been carried out devoted to line segmentation. Broadly, these methods can be classified into two categories: critical-point- and shape-analysis-based methods. Critical points, such as inflection points [19], are usually detected as break points, thus yielding segmentation results. Additionally, several studies implicitly achieved segmentation by applying point-compression models in which the retained points were used as the critical points [3,20]. Shape-analysis-based methods identify homogeneous segments with geometric measurements. For example, in the approach proposed by Plazanet et al. [4], an input line was recursively divided until all parts became geometrically homogeneous. The pattern type of each segment was then classified based on the shape measurements of the bends. Samsonov and Yakimova [12] proposed several procedures, including filtering, segmentation, and squaring, to decompose a line into homogeneous segments. They identified orthogonal and non-schematic segments based on analysis of the angle and distance measurements.
Intelligent methods have been explored for line segmentation based on shape analysis and geometric measures. For instance, Balboa and López [16,17] attempted to emulate expert human segmentation for road lines via a backpropagation artificial neural network (BANN). A set of geometric features was used to describe the shape characteristics of each road segment. A BANN classification model was then designed to establish mapping between the descriptive features and the pattern types. Finally, the line segmentation results were obtained by applying the classifier to predict the class of each component derived from a window moving along the lines. Liu and Yang [18] developed a similar method for coastline segmentation that processes the geometrical features of a segment using principal component analysis (PCA). They adopted the Bayesian model to construct a classifier to predict the pattern types of segments extracted by a moving window, and output the segmentation results by merging the segments belonging to identical classes. Although many methods have been developed for the purpose, the effective analysis and segmentation of lines remains challenging. First, it is difficult to comprehensively and objectively reveal the shape characteristics of a line, resulting in a lack of information for distinguishing segments with different shape characteristics. Second, existing methods lack an effective mechanism to consider local contextual information, which limits the performance of pattern recognition and segmentation.
From a visual cognition perspective, the line segmentation problem is similar to the image segmentation problem (i.e., combining units with similar characteristics to form local continuous homogeneous structures). In the image segmentation field, convolutional neural networks (CNNs) have excellent properties, including local perception and multiscale feature characterization, rendering them the most advanced technology for solving this problem [21,22]. Recently, CNNs and their variants have been successfully applied to map data processing, including geographical pattern recognition and classification [23,24,25], shape representation and classification [26,27], and cartographical generalization [7,28,29]. In this study, we employed CNN technology to construct a method for pattern recognition and segmentation of lines. However, unlike image data with grid-topology structures, each line consisted of a sequence of points with uneven distributions; the distance between two adjacent points was not identical. Thus, the processing units were not stationary. Therefore, we divided the lines into a series of consecutive linear units with equal lengths, also known as lixels [30], which serve as the basic units that compose a line. Based on this, each line was organized into a list of lixels, sequentially ordered and evenly distributed. In this manner, the line segmentation problem was transformed into a lixel classification problem, which could be analyzed and processed using advanced image segmentation techniques, to solve the shortcomings inherent in existing methods that include excessive dependence on handcrafted features and insufficient consideration of contextual information.
We propose a grid shape context descriptor (GSCD) to describe the contextual characteristics of each lixel and its neighbors. This descriptor provides a standardized computation for the contextual features with respect to each lixel, thereby reducing the subjective impact of manually defined features encountered in previous methods. Subsequently, with reference to the classic architecture [31], a novel one-dimensional convolutional neural network (1D-U-Net) that combines operations such as one-dimensional convolution, pooling, and transposed convolution, was constructed to analyze the extracted contextual features of the lixels and assign a pattern type to each. Finally, adjacent lixels with the same pattern types were fused to output the final segmentation results. To verify the proposed method, administrative boundaries were chosen as the experimental data. This is because human and natural factors both influence the geometrical morphologies of administrative boundaries, thus presenting a variety of pattern types conducive to effectively testing the proposed method.
The remainder of this paper is organized as follows. Section 2 introduces the experimental datasets and pattern types for the administrative boundaries. Section 3 details the proposed segmentation method using GSCD and 1D-U-Net. Section 4 presents the experimental design and results, as well as a detailed analysis, comparison, and discussion. Finally, Section 5 concludes this study.

2. Experimental Datasets and Shape Patterns

2.1. Experimental Datasets

In this study, two sets of administrative boundary data for southern China extracted from a 1:50,000 land-use database were used as the training and test datasets, respectively (Figure 2). The training dataset contained 102 administrative boundaries, with a total length of 927.781 km and an average length of 9.096 km. The test dataset contained 52 administrative boundaries, with a total length of 783.857 km and an average length of 15.074 km. The boundaries of the two datasets had complex and diversified morphological characteristics, suitable for confirming the effectiveness of the proposed method.

2.2. Administrative Boundary Shape Pattern Types

Numerous variants and combinations of shape patterns for lines were included in this study. For administrative boundary data, components may derive from natural objects, such as rivers and coastlines, as well as anthropogenic objects, such as roads and built-up areas. Natural objects are usually characterized by a hierarchical bend structure, whereas anthropogenic objects tend to have artificially sharp or right-angled features. Samsonov and Yakimova [12] demonstrated that these differences in line shape patterns can be conceptualized in three aspects: smoothness, schematism, and regularity. The degree of smoothness indicates whether the shape change along a line is smooth (i.e., whether it has a gradually changing tangent direction or contains a large deviation angle). Schematism indicates whether the shape composition of a line is simple or if it contains complex hierarchical curved structures of various sizes. Regularity refers to whether a line shape has repetitive characteristics.
Based on the shape pattern space constructed from these three dimensions, as well as the characteristics of the experimental data, we classified the shape patterns of the boundaries into three pattern types: smooth irregular schematic (SIS), sharp regular schematic (SRS), and sharp irregular non-schematic (SIN), as listed in Table 1. SIS segments were characterized as smooth, with no repetitive characteristics and uncomplicated curved structures, and were derived from natural and artificial environments. SRS segments were characterized by sharp orthogonal regularities similar to building outlines, mainly derived from artificial environments. SIN segments were characterized as sharp, with no repetitive characteristics and complex hierarchical bends of various sizes, mainly derived from the natural environment.

3. Methodology

Figure 3 shows the overall framework of the proposed method, including four main steps: lixel generation and labeling, lixel feature extraction, classification, and segmentation.
  • Lixel generation and labeling: Each administrative boundary was converted into a series of lixels via equidistant subdivision; the pattern type of each lixel was labeled;
  • Feature extraction for lixels: Automatic extraction of the contextual features for each lixel, using GSCD;
  • Lixel classification using 1D-U-Net: Construction of a 1D-U-Net to classify the pattern type of each lixel based on the extracted features;
  • Segmentation: Obtaining the segmentation results by fusing adjacent lixels with the same pattern type.

3.1. Lixel Generation and Labeling

Figure 4 illustrates the lixel generation and labeling processes. First, each administrative boundary was manually segmented according to its morphological characteristics. The pattern type of each segment was classified according to the pattern category described in Section 2.2, and the visual classification was performed at a fixed scale to eliminate the influence of the data display scale on pattern discrimination. The boundary was then divided into a series of lixels with a fixed length. Note that the last divided part of the boundary that did not satisfy the predefined length was considered a lixel.
The pattern type of each lixel was determined based on the pattern type of the segment to which it belonged. If a lixel was associated with multiple segments, it was labeled with the pattern type of the segment where it had the longest overlap length. Let E n × 1 = e 1 , e 2 , , e n ( n > 1 ) denote the sequence of lixels of an administrative boundary. The label information was organized as sequence data: T n × 1 = < t 1 , t 2 , , t n > , where t i { [ 1 , 0 , 0 ] , [ 0 , 1 , 0 ] , [ 0 , 0 , 1 ] }   ( i = 1 , 2 , , n ) is a three-dimensional one-hot vector that represents the pattern type of lixel e i .

3.2. Extracting Lixel Features Using GSCD

The morphological characteristics of a lixel were determined by its neighbors within a certain range (i.e., the contextual information). In this study, we employed a GSCD commonly used for shape analysis and pattern recognition [32] to extract the contextual features of the lixels. For each lixel on an administrative boundary, the GSCD was computed as follows:
  • A regular grid centered on the midpoint of the lixel was created. The grid contained p × p cells and the cell edges were always horizontal and vertical. The length of the cell edges was set to the fixed length of the lixels.
  • The length of the boundary located in each cell was counted and normalized by dividing the total length of the boundary within all cells.
  • The normalized values of all cells were arranged from left to right and from bottom to top into a feature vector that was used to describe the contextual features of the lixel.
Figure 5 shows the GSCDs with 5 × 5 cells for three different lixels. The grayscale value of each grid cell represents the normalized feature value. The GSCDs of the three lixels in different contexts also significantly varied, indicating that the GSCD method had good feature characterization ability.
After computing the GSCDs of all lixels along an administrative boundary, we obtained sequence data, F n × ( p × p ) , with n lixels and ( p × p ) -dimensional features, which served as input for the 1D-U-Net for subsequent pattern classification and segmentation.

3.3. Classifying Lixels Using a 1D-U-Net

This study referred to the classic U-Net [21] to construct the 1D-U-Net for classifying lixel patterns. The basic idea was to map the input features, F n × ( p × p ) , into labels, T n × 1 , through multiple 1-D convolution, pooling, and upsampling operations. Figure 6 shows the detailed architecture of the 1D-U-Net, which consisted of four downsampling blocks and four upsampling blocks. Each downsampling block contained two 1-D convolutions and one max pooling; each upsampling block contained one skip connection, one 1-D transposed convolution, and two 1-D convolutions. The last layer consisted of one 1-D convolution and one softmax activation function that mapped each feature vector into a label vector.
The following sections described the operations performed on the sequence data for the lixels used in this model.

3.3.1. One-Dimensional Convolution and Pooling Operations

One-dimensional convolution and pooling operations were used to extract high-level multi-scale features for lixel pattern classification from the shallow GSCD features of the lixels. The convolution operation processes local data in a window through a sliding kernel and generates new features using a nonlinear activation function. A sequence, X = { X 1 , X 2 , , X n } , was convolved using a kernel, k = { w 1 , w 2 , , w l 1 } , with a window size of l 1 to generate a new sequence, C . Each feature, c i , in C was computed as follows:
c i = f ( k · X i ( l 1 1 ) 2 : i + ( l 1 1 ) 2 + b )
where f ( · ) denotes the nonlinear Rectified Linear Unit (ReLU) activation function and b denotes the bias. The generated sequence is denoted as C = { c 1 , c 2 , , c f l o o r [ n + 2 d 1 l 1 s 1 ] + 1 } , where f l o o r [ · ] is the largest integer function, d 1 is the width of the padding, and s 1 is the stride length of the window.
The pooling operation obtained coarser-grained features by representing multiple features in a local window as one feature, which was conducive to reducing parameters, accelerating computations, and preventing overfitting, thereby improving the characterization and generalization ability of the model. Max pooling was used in this study. For multiple features, C ( i 1 ) s 2 + 1 : ( i 1 ) s 2 + l 2 , with a window size of l 2 , the maximum value was applied as the output feature, expressed as follows:
p i = m a x j = 0 l 2 1 ( C ( i 1 ) s 2 + 1 + j )
where s 2 is the stride length of the window. The generated sequence can be expressed as P = { p 1 , p 2 , , p f l o o r [ n l 2 s 2 ] + 1 } .
Figure 7 provides an example of the convolution and pooling operations, where the window size l 1 of the kernel was three and the window size l 2 of the pooling was two. A sequence C with a size of n and a sequence P with a size of n/2 were obtained after the convolution and max pooling operations, respectively.

3.3.2. One-Dimensional Upsampling Operation and Skip Connection

The pooling operation changed the size of the input features. The upsampling operation, implemented during the transposed convolution, was used to restore the feature size to the original input size.
The key step in the transposed convolution was to construct a transposed matrix, W , and multiply it with the feature vector, P . For the input features, P , with size n × 1 , the transposed matrix, W , was obtained by sliding a convolution kernel, k = { w 1 , w 2 , , w l 3 } , with a window size of l 3   n times, where the vertical stride length of the sliding was s 3 and the horizontal stride length s 4 was 1. The transposed matrix W was computed:
W i j = { w i s 3 ( j 1 ) , s 3 ( j 1 ) + 1 i s 3 ( j 1 ) + l 3     0 , otherwise
where i = { 1 , 2 , , s 3 ( n 1 ) + l 3 } and j = { 1 , 2 , , n } denotes the number of rows and columns, respectively. The transposed convolution operation can be expressed:
V ( s 3 ( n 1 ) + l 3 ) × 1 = W ( s 3 ( n 1 ) + l 3 ) × n × P n × 1
where W ( s 3 ( n 1 ) + l 3 ) × n denotes the transposed matrix and V ( s 3 ( n 1 ) + l 3 ) × 1 denotes the output features.
Figure 8 presents an example of a transposed convolution operation, where the vertical stride length s 3 was two and the convolution kernel width l 3 was two. The size of the transposed convolution matrix, W , is 2n × n, and thus a new feature vector, V , with a size of 2n × 1 is obtained.
We note that the pooling and upsampling operations may omit important spatial location information from the original input features. To alleviate this problem, a skip connection operation was put in place between the 1-D convolution and 1-D transposed convolution. As illustrated in Figure 6, the output feature vectors after the 1-D transposed convolution were connected to the feature vectors after the 1-D convolution with the same number of channels. Then, the values of each element in the feature vector were corrected using a convolution operation to restore the number of channels.

3.3.3. Definition of Loss Function

After convolution processing in the final layer, a sequence the same length as the input sequence was output; its number of channels was three, corresponding to the three pattern types. Finally, the softmax function was employed to activate the output features to obtain the predicted probabilities. For the i -th ( i = 1 , 2 , , n ) lixel of the output sequence, the probability ( a i ) j ( j = 1 , 2 , 3 ) that it belongs to the j -th pattern type was computed as follows:
( a i ) j = e z j k = 1 3 e z k
where z j denotes the feature of the j-th channel of the lixel. The output probability vector, a i = ( a i ) 1 , ( a i ) 2 , ( a i ) 3 , for the i -th lixel satisfied j = 1 3 ( a i ) j = 1 . The pattern type with the highest probability was considered the predicted pattern for this lixel.
The training process allowed the differences (i.e., the loss value, E ) to be minimized between the predicted probability vectors, a 1 , a 2 , , a n , and the labeling one-hot vectors, t 1 , t 2 , , t n , where n was the number of lixels. In this study, the cross-entropy function was applied to measure the difference, expressed as follows:
E = 1 n i = 1 n t i l o g ( a i )
The smaller the loss value, E , the closer the predictions were to the labels. Here, 1D-U-Net was trained using a backpropagation algorithm. During training, the predicted value for each mini-batch of data was obtained via forward propagation and the loss value was calculated. The trainable parameters were then updated gradually according to the learning rate and partial derivatives of each parameter relative to the loss value.

3.4. Obtaining Segmentation Results

The trained 1D-U-Net was used to predict the pattern types of all lixels along an administrative boundary. Homogeneous segments were then obtained by merging adjacent lixels with the same pattern types. However, there were some predictions for a few lixels that were incorrect, resulting in extremely short segments after the merging operation. As these short segments interfered with the segmentation results, post-processing was required to eliminate them. In this study, an iterative fusion method was used. As illustrated in Figure 9, post-processing was implemented as follows:
  • The segmentation results of the administrative boundary were traversed and the segment z i with the smallest length was identified;
  • If the length of z i was smaller than the predefined threshold S , z i was merged with its neighbor with a longer length;
  • Steps (1) and (2) were repeated until there were no segments smaller than S .

4. Experiments

The GSCD to automatically extract the features for lixels was implemented as an ArcMap add-in (Environmental Systems Research Institute, Redlands, CA, USA) with C#, and the constructed 1D-U-Net to classify lixels was implemented with Python in TensorFlow. This section presents the experimental design, lixel classification performance, line segmentation results and analysis, and a discussion on the parameter sensitivities.

4.1. Experimental Design

4.1.1. Sample Dataset Generation

To generate the samples, two volunteers with specialized knowledge independently segmented the administrative boundaries in the training and test datasets according to the criteria listed in Table 1. If the segmentation results from the two volunteers were different, a third volunteer with extensive cartographic experience rechecked and made the final decision. To ensure the consistency of cartographical details, the display scale when segmenting the boundaries was fixed at the data scale (i.e., 1:50,000). Next, each boundary was divided into a series of lixels, as described in Section 3.1. The setting of lixel size is critical for perceiving line patterns [33,34]. In this current study, by referring to the concept of the smallest visual object (SVO) discussed in the work of Li and Openshaw [33], and their comparative studies on the size settings of this parameter, the length of each lixel was set to 0.5 mm (map distance), corresponding to a ground distance of 25 m at a scale of 1:50,000. Consequently, there were 37,235 and 31,425 lixels for the training and test datasets, respectively.
To fully train the lixel classification model, two data augmentation methods were used to increase the sample size of the training dataset. As illustrated in Figure 10a, each administrative boundary was first rotated every 30°; the sample size was increased 11 times. Next, we implemented the sliding window method, sliding a fixed window with a size of 112 lixels along each boundary at steps of 30 lixels. As shown in Figure 10b, each window slide yielded a sample; thus, multiple samples with partial overlaps were obtained from one administrative boundary. Through data augmentation, 9432 training samples were obtained.
For the test dataset, data augmentation was not applied and the lixel sequence of each administrative boundary was divided into samples with a length of 112 lixels. If the last sample did not reach the required length, zeros were added. Finally, 308 test samples were obtained.

4.1.2. Parameter Settings

After preparing all the samples, the contextual features of each lixel were computed using the GSCD; a grid of 5 × 5 cells was used to construct the GSCD in the experiments. Therefore, each lixel was described using 25 dimensional features. The model was trained using the Adam optimizer for 50 epochs, with a learning rate of 0.0001.
Two existing segmentation methods based on machine learning were implemented for comparison; the backpropagation artificial neural network (BANN) [16,17] and naïve Bayesian (NB) [18] methods. Both methods used the sliding window method to generate samples from the administrative boundaries. The key step was to cut the administrative boundaries according to the fixed-size window which moved with an increment of a certain length. Referring to the experimental parameter settings discussed in the literature [17,18], the window size and length of the increment were set to 1500 and 150 m, respectively. Ten features were extracted to describe the morphological structures of the samples, including the segment length and baseline length ratio, mean of the bend lengths, median vertical distance from each point to the baseline, coefficient of variance of the bend baseline length, coefficient of variance of the bend length and baseline length ratio, median distance between two continuous points, coefficient of variance of the bend area and squared baseline length ratio, median ratio of the bend length to baseline length, fractal dimension, and mean of the turning angles. For more information on the definitions and computations of these features, please refer to the works of Ariza López and Balboa [16,17] and Liu et al. [18]. Based on principal component analysis (PCA), seven features with a sum of information greater than 90% were used as the inputs for the BANN and NB models. The number of neurons in the hidden layer of BANN was set to 15 and the ReLU activation function was used. A Gaussian model was used for the NB method.

4.2. Lixel Classification Performance Using 1D-U-Net

Figure 11 presents the accuracy and loss values of 1D-U-Net for lixel classification during the training phase. The classification accuracy and training loss changed rapidly during the first five rounds, gradually stabilized after ten epochs, and reached a peak after 50 epochs. After the training converged, the classification accuracy of the model on the training set reached 99.04%. Subsequently, the trained model was used to classify the lixels in the test samples, with a classification accuracy of 90.42%. This result indicates that 1D-U-Net can classify the pattern types of lixels with high accuracy.
The Precision, Recall, and F 1 -score were used to quantitatively evaluate the classification results. For each pattern type, the three metrics are defined as follows:
P r e c i s i o n = T P n u m T P n u m + F P n u m × 100%
R e c a l l = T P n u m T P n u m + F N n u m × 100%
F 1 - s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
where T P n u m denotes the number of lixels that were both automatically predicted and manually labeled with the same pattern type; F P n u m is the number of lixels that were automatically predicted to be this pattern type but were manually identified as a different pattern type; and F N n u m denotes the number of lixels that were manually identified as this pattern type but were incorrectly predicted as a different pattern types.
Table 2 lists the confusion matrix of the lixel classification results for the test samples using the trained model, and the three metrics. The F 1 -score value for the SIN pattern reached 0.96 while the F 1 -score values for the SIS and SRS patterns were slightly lower, at 0.90 and 0.88, respectively. This is because misclassification occurred mainly between the SIS and SRS lixels. Overall, the proposed model could extract the shapes’ contextual features and achieve lixel classification for the test boundaries with relatively high accuracy.

4.3. Segmentation Result Evaluation

4.3.1. Qualitative Evaluation

Based on the classification results, homogeneous segments were obtained by merging adjacent lixels with the same pattern types. To make the comparison fairer, the iterative fusion method was also employed in the BANN and NB methods after classifying each increment. In all three methods, the length threshold S for the short segment needing to be fused was set to 900 m (i.e., 36 lixels in the proposed method and six increments in the BANN and NB methods). Figure 12 shows the segmentation results for the test boundaries using different methods.
All three methods performed well in the segmentation of the test boundaries. By using the iterative fusion method, many short segments were removed and the segmentation results became more coherent. Careful comparison revealed that the proposed method had fewer segmentation errors than the BANN and NB methods. For example, the segment marked by the green ellipse in Figure 12a was manually labeled as an SRS pattern, but both the BANN and NB methods incorrectly predicted it as SIS or SIN; only the results of the proposed method were consistent with the manual labeling.
Figure 13 shows the differences between the manual and predicted results based on different methods, where predictions that coincided with manual results are rendered in gray while colored results indicate inconsistent results. There were significantly fewer inconsistent segments in the results of the proposed method than in the other two methods. The length of the inconsistent segments in the results of the proposed method was 59.5 km while the BANN and NB methods produced totals of 167.64 and 184.18 km of inconsistent segments, respectively.
Inconsistent segmentation can be divided into two classes. The first class is characterized by inconsistent segments that are short. This type of inconsistency was produced by deviations in the segment points; deviations within a certain range had a negligible impact on subsequent analyses. Instances in the second class included long inconsistent segments, produced by the incorrect classification of pattern types. As shown in Figure 13, both classes of inconsistencies in the results obtained by the BANN and NB methods were significantly greater than those of the proposed method.
To further investigate the two classes of inconsistencies that occurred in the segmentation results using different methods, we analyzed the inconsistencies in four typical boundaries, as shown in Figure 14. The deviations in the segmentation points occurred mainly at the junctions of the SIS and SRS segments. In terms of incorrect classification, the identification of SRS segments generated the worst performances, particularly in the results of the BANN and NB methods. Lines with SRS patterns may have long straight segments on both sides or may exhibit consecutive bends. In the former case, the local characteristics are similar to those of SIS lines in gently sloping areas, whereas in the latter case, SRS and SIN lines exhibit overall similarity. Overall, accurately identifying the SRS segments requires understanding the local and global characteristics of the lines.

4.3.2. Quantitative Evaluation

A consistency ratio (CR) metric was defined to quantitatively analyze the segmentation performance of the different methods. For the i-th pattern type, the C R i metric was computed as follows:
C R i = L c i L t i × 100%
where L c i and L t i denote the total length of the segments correctly identified as the i-th pattern type and the total length of the segments labeled as the i-th pattern type, respectively. The overall CR (OCR) was calculated as follows:
O C R = i L c i L t × 100%
where L t denotes the total length of all test boundaries. Table 3 lists the CR and OCR metrics for the segmentation results obtained using different methods. The OCR of the 1D-U-Net method reached 92.41%, which was higher than for either of the two existing methods. Moreover, the identification of SIN segments had the highest CR, reaching 97% using all three methods. In contrast, the three methods each had the lowest CR for SRS segments, with CRs for the two existing methods that were less than 60%; the CR of the proposed method approached 90%. Overall, the proposed method demonstrated the best segmentation performance in terms of CR and OCR metrics; it was advantageous in the identification of SRS segments, compared with the traditional BANN and NB methods.

4.4. Discussion

As previously mentioned, the GSCD can capture the local contextual features of each lixel and alleviate the subjective influence of artificially defined features. However, the grid size (i.e., the number of cells) had a significant impact on the lixel feature extraction and may have affected the segmentation performance. For descriptors with different grid sizes, the extracted features differed (Figure 15).
To discuss the sensitivity of the proposed method to the size of the GSCD, an experiment was conducted by varying the grid sizes from 3 × 3 to 7 × 7. Table 4 lists the lixel classification accuracies. With an increase in grid size, the lixel classification accuracy of the model initially increased and then decreased. The model performed best when the grid cell size was 5 × 5. A possible reason for this result is that as the number of cells increased from 3 × 3, the receptive field gradually increased; therefore, the contextual features of the lixels were enriched, thus improving the classification performance. However, when the grid size was greater than 5 × 5, many features were zero (Figure 15c) and the contextual information may have been disturbed by long-distance boundaries, thus leading to the degradation of classification performance.

5. Conclusions and Outlook

This study proposed a novel deep-learning approach for pattern recognition and segmentation of administrative boundaries, based on 1D-U-Net. In the model, a lixel was used as the basic processing unit and a GSCD was employed to extract the descriptive features of each lixel. Subsequently, a 1D-U-Net architecture was constructed to predict the lixels’ pattern types. Finally, the predicted results were iteratively fused to obtain the final segmentation results for the administrative boundaries. The experimental results showed that the lixel classification accuracy of 1D-U-Net reached 90.42% for the test administrative boundaries; the OCR of the segmentation results for the test samples was 92.41%, which was higher than that of the BANN- and NB-based segmentation methods.
Unlike existing methods, the proposed method is derived from image segmentation. It transforms the line segmentation problem into lixel classification by representing unstructured vector-based data as regular lixel-based sequence data. Additionally, the GSCD provides a promising means for characterizing local contextual information in lixels. These two advantages allow the proposed method to effectively identify the shape characteristics of a line, which significantly improves segmentation performance. Follow-up studies should focus on certain aspects. The method should be applied to other geographical lines, such as coastlines, rivers, and roads. Model optimization and parameter settings for the method, such as lixel size, should be further investigated. In addition, high-quality sample libraries should be constructed to further improve the model.

Author Contributions

Conceptualization, Min Yang and Xiongfeng Yan; methodology, Min Yang and Xiongfeng Yan; software, Haoran Huang and Yiqi Zhang; validation, Min Yang and Xiongfeng Yan; formal analysis, Min Yang, Haoran Huang, Yiqi Zhang and Xiongfeng Yan; investigation, Min Yang, Haoran Huang, Yiqi Zhang and Xiongfeng Yan; writing—original draft preparation, Min Yang, Haoran Huang, Yiqi Zhang and Xiongfeng Yan; writing—review and editing, Min Yang and Xiongfeng Yan; supervision, Min Yang and Xiongfeng Yan; funding acquisition, Min Yang and Xiongfeng Yan. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant numbers 42071450, 42001415.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Buttenfield, B.P. A Rule for Describing Line Feature Geometry; Longman: London, UK; Wiley: New York, NY, USA, 1991; pp. 150–171. [Google Scholar]
  2. Dutton, G. Scale, sinuosity, and point selection in digital line generalization. Cartogr. Geogr. Inf. Sci. 1999, 26, 33–54. [Google Scholar] [CrossRef]
  3. Visvalingam, M.; Williamson, P.J. Simplification and generalization of large scale data for roads: A comparison of two filtering algorithms. Cartogr. Geogr. Inf. Syst. 1995, 22, 264–275. [Google Scholar] [CrossRef]
  4. Plazanet, C. Modelling geometry for linear feature generalization. In Geographic Information Research: Bridging the Atlantic; Craglia, M., Coucleis, H., Eds.; Taylor & Francis: London, UK, 1997; pp. 264–279. [Google Scholar]
  5. Ai, T.; Ke, S.; Yang, M.; Li, J. Envelope generation and simplification of polylines using Delaunay triangulation. Int. J. Geogr. Inf. Sci. 2017, 31, 297–319. [Google Scholar] [CrossRef]
  6. Samsonov, T.; Yakimova, O. Regression modeling of reduction in spatial accuracy and detail for multiple geometric line simplification procedures. Int. J. Cartogr. 2020, 6, 47–70. [Google Scholar] [CrossRef]
  7. Du, J.; Wu, F.; Xing, R.; Gong, X.; Yu, L. Segmentation and sampling method for complex polyline generalization based on a generative adversarial network. Geocarto Int. 2021, 37, 4158–4180. [Google Scholar] [CrossRef]
  8. Wang, Z.; Müller, J.C. Line generalization based on analysis of shape characteristics. Cartogr. Geogr. Inf. Sci. 1998, 25, 3–15. [Google Scholar] [CrossRef]
  9. Wang, Z.; Lee, D. Building simplification based on pattern recognition and shape analysis. In Proceedings of the 9th International Symposium on Spatial Data Handling, Beijing, China, 10–12 August 2000; pp. 58–72. [Google Scholar]
  10. García Balboa, J.L.; Ariza López, F.J. Sinuosity pattern recognition of road features for segmentation purposes in cartographic generalization. Pattern Recognit. 2009, 42, 2150–2159. [Google Scholar] [CrossRef]
  11. Park, W.; Yu, K. Hybrid line simplification for cartographic generalization. Pattern Recognit. Lett. 2011, 32, 1267–1273. [Google Scholar] [CrossRef]
  12. Samsonov, T.E.; Yakimova, O.P. Shape-adaptive geometric simplification of heterogeneous line datasets. Int. J. Geogr. Inf. Sci. 2017, 31, 1485–1520. [Google Scholar] [CrossRef]
  13. Wei, Z.; Liu, Y.; Cheng, L.; Ding, S. A progressive and combined building simplification approach with local structure classification and backtracking strategy. ISPRS Int. J. Geo-Inf. 2021, 10, 302. [Google Scholar] [CrossRef]
  14. Yang, M.; Yuan, T.; Yan, X.; Ai, T.; Jiang, C. A hybrid approach to building simplification with an evaluator from a backpropagation neural network. Int. J. Geogr. Inf. Sci. 2022, 36, 280–309. [Google Scholar] [CrossRef]
  15. Mustiere, S. Cartographic generalization of roads in a local and adaptive approach: A knowledge acquistion problem. Int. J. Geogr. Inf. Sci. 2005, 19, 937–955. [Google Scholar] [CrossRef]
  16. García Balboa, J.L.; Ariza López, F.J. Generalization-oriented road line classification by means of an artificial neural network. GeoInformatica 2008, 12, 289–312. [Google Scholar] [CrossRef]
  17. Ariza López, F.J.; García Balboa, J.L. Generalization-oriented road line segmentation by means of an artificial neural network applied over a moving window. Pattern Recognit. 2008, 41, 1593–1609. [Google Scholar] [CrossRef]
  18. Liu, P.; Yang, Q. Coastline segment model research for map generalization based on Bayesian method. Comput. Eng. Appl. 2016, 52, 174–179. [Google Scholar] [CrossRef]
  19. Plazanet, C.; Affholder, J.G.; Fritsch, E. The importance of geometric modeling in linear feature generalization. Cartogr. Geogr. Inf. Syst. 1995, 22, 291–305. [Google Scholar] [CrossRef]
  20. Douglas, D.H.; Peucker, T.K. Algorithms for the reduction of the number of points required to represent a digitized line or its caricature. Cartogr. Int. J. Geogr. Inf. Geovis. 1973, 10, 112–122. [Google Scholar] [CrossRef]
  21. Minaee, S.; Boykov, Y.Y.; Porikli, F.; Plaza, A.J.; Kehtarnavaz, N.; Terzopoulos, D. Image segmentation using deep learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 3523–3542. [Google Scholar] [CrossRef] [PubMed]
  22. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention (MICCAI), Cham, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  23. Yan, X.; Ai, T.; Yang, M.; Yin, H. A graph convolutional neural network for classification of building patterns using spatial vector data. ISPRS J. Photogramm. Remote Sens. 2019, 150, 259–273. [Google Scholar] [CrossRef]
  24. Yang, M.; Jiang, C.; Yan, X.; Ai, T.; Cao, M.; Chen, W. Detecting interchanges in road networks using a graph convolutional network approach. Int. J. Geogr. Inf. Sci. 2022, 36, 1119–1139. [Google Scholar] [CrossRef]
  25. Yu, H.; Ai, T.; Yang, M.; Huang, L.; Yuan, J. A recognition method for drainage patterns using a graph convolutional network. Int. J. Appl. Earth Obs. Geoinf. 2022, 107, 102696. [Google Scholar] [CrossRef]
  26. Yan, X.; Ai, T.; Yang, M.; Tong, X. Graph convolutional autoencoder model for the shape coding and cognition of buildings in maps. Int. J. Geogr. Inf. Sci. 2021, 35, 490–512. [Google Scholar] [CrossRef]
  27. Liu, C.; Hu, Y.; Li, Z.; Xu, J.; Han, Z.; Guo, J. TriangleConv: A deep point convolutional network for recognizing building shapes in map space. ISPRS Int. J. Geo-Inf. 2021, 10, 687. [Google Scholar] [CrossRef]
  28. Feng, Y.; Thiemann, F.; Sester, M. Learning cartographic building generalization with deep convolutional neural networks. ISPRS Int. J. Geo-Inf. 2019, 8, 258. [Google Scholar] [CrossRef]
  29. Courtial, A.; El Ayedi, A.; Touya, G.; Zhang, X. Exploring the potential of deep learning segmentation for mountain roads generalisation. ISPRS Int. J. Geo-Inf. 2020, 9, 338. [Google Scholar] [CrossRef]
  30. He, Y.; Ai, T.; Yu, W.; Zhang, X. A linear tessellation model to identify spatial pattern in urban street networks. Int. J. Geogr. Inf. Sci. 2017, 31, 1541–1561. [Google Scholar] [CrossRef]
  31. Kim, Y. Convolutional neural networks for sentence classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1746–1751. [Google Scholar] [CrossRef]
  32. Fan, H.; Zhao, Z.; Li, W. Towards measuring shape similarity of polygons based on multiscale features and grid context descriptors. ISPRS Int. J. Geo-Inf. 2021, 10, 279. [Google Scholar] [CrossRef]
  33. Li, Z.; Openshaw, S. Algorithms for automated line generalization based on a natural principle of objective generalization. Int. J. Geogr. Inf. Sci. 1992, 6, 373–389. [Google Scholar] [CrossRef]
  34. Karsznia, I.; Gołębiowska, I.M.; Korycka-Skorupa, J.; Nowacki, T. Searching for an optimal hexagonal shaped enumeration unit size for effective spatial pattern recognition in choropleth maps. ISPRS Int. J. Geo-Inf. 2021, 10, 576. [Google Scholar] [CrossRef]
Figure 1. Simplified results (in red) of a line with different shape characteristics using two different algorithms: (a) the original line; (b) result using the bend-based simplification algorithm; (c) result using the orthogonality-preserving simplification algorithm.
Figure 1. Simplified results (in red) of a line with different shape characteristics using two different algorithms: (a) the original line; (b) result using the bend-based simplification algorithm; (c) result using the orthogonality-preserving simplification algorithm.
Ijgi 11 00461 g001
Figure 2. Experimental data: (a) administrative boundaries in the training dataset, and (b) administrative boundaries in the testing dataset.
Figure 2. Experimental data: (a) administrative boundaries in the training dataset, and (b) administrative boundaries in the testing dataset.
Ijgi 11 00461 g002
Figure 3. Framework of the segmentation method for administrative boundaries using the one-dimensional convolutional neural network (1D-U-Net).
Figure 3. Framework of the segmentation method for administrative boundaries using the one-dimensional convolutional neural network (1D-U-Net).
Ijgi 11 00461 g003
Figure 4. Lixel generation and labeling for each administrative boundary. Blue, red, and black colors of the divided segments indicate SIS, SRS, and SIN patterns, respectively.
Figure 4. Lixel generation and labeling for each administrative boundary. Blue, red, and black colors of the divided segments indicate SIS, SRS, and SIN patterns, respectively.
Ijgi 11 00461 g004
Figure 5. Grid shape context descriptors (GSCDs) with 5 × 5 cells for describing the contextual characteristics of three different lixels (in green) on an administrative boundary.
Figure 5. Grid shape context descriptors (GSCDs) with 5 × 5 cells for describing the contextual characteristics of three different lixels (in green) on an administrative boundary.
Ijgi 11 00461 g005
Figure 6. Architecture of the one-dimensional convolutional neural network (1D-U-Net) for classifying lixel pattern types.
Figure 6. Architecture of the one-dimensional convolutional neural network (1D-U-Net) for classifying lixel pattern types.
Ijgi 11 00461 g006
Figure 7. Illustration of the one-dimensional convolution and pooling operations.
Figure 7. Illustration of the one-dimensional convolution and pooling operations.
Ijgi 11 00461 g007
Figure 8. Illustration of the one-dimensional transposed convolution operation.
Figure 8. Illustration of the one-dimensional transposed convolution operation.
Ijgi 11 00461 g008
Figure 9. Post-processing for short segments to obtain fine segmentation results.
Figure 9. Post-processing for short segments to obtain fine segmentation results.
Ijgi 11 00461 g009
Figure 10. Two data augmentation methods for the training dataset: (a) rotation transformation method and (b) sliding window method.
Figure 10. Two data augmentation methods for the training dataset: (a) rotation transformation method and (b) sliding window method.
Ijgi 11 00461 g010
Figure 11. Training accuracy and loss in 1D-U-Net for lixel classification.
Figure 11. Training accuracy and loss in 1D-U-Net for lixel classification.
Ijgi 11 00461 g011
Figure 12. Segmentation results for the test administrative boundaries: (a) manual segmentation, (b) 1D-U-Net method, (c) BANN method, and (d) NB method.
Figure 12. Segmentation results for the test administrative boundaries: (a) manual segmentation, (b) 1D-U-Net method, (c) BANN method, and (d) NB method.
Ijgi 11 00461 g012
Figure 13. Differences in the manual segmentation results and predicted results of different methods for the test administrative boundaries: (a) 1D-U-Net, (b) BANN method, and (c) NB method. Gray indicates that the segmentation results are consistent; blue, red, and black denote the segments with inconsistent predictions as SIS, SRS, and SIN patterns, respectively.
Figure 13. Differences in the manual segmentation results and predicted results of different methods for the test administrative boundaries: (a) 1D-U-Net, (b) BANN method, and (c) NB method. Gray indicates that the segmentation results are consistent; blue, red, and black denote the segments with inconsistent predictions as SIS, SRS, and SIN patterns, respectively.
Ijgi 11 00461 g013
Figure 14. Segmentation results for four typical test boundaries (numbered 1–4) using different methods: (a) manual identification, (b) 1D-U-Net, (c) BANN method, and (d) NB method. Boxes highlight incorrect classifications and circles mark deviations in the segmentation points.
Figure 14. Segmentation results for four typical test boundaries (numbered 1–4) using different methods: (a) manual identification, (b) 1D-U-Net, (c) BANN method, and (d) NB method. Boxes highlight incorrect classifications and circles mark deviations in the segmentation points.
Ijgi 11 00461 g014
Figure 15. Grid shape context descriptors (GSCDs) with different sized grids for each lixel: (a) grid size of 3 × 3; (b) grid size of 5 × 5; (c) grid size of 7 × 7.
Figure 15. Grid shape context descriptors (GSCDs) with different sized grids for each lixel: (a) grid size of 3 × 3; (b) grid size of 5 × 5; (c) grid size of 7 × 7.
Ijgi 11 00461 g015
Table 1. Line characteristics of pattern types for administrative boundary segments.
Table 1. Line characteristics of pattern types for administrative boundary segments.
Pattern TypeExampleSmoothnessRegularitySchematism
Smooth irregular schematic
(SIS)
Ijgi 11 00461 i001Smooth
(gradually changing tangent directions)
Irregular
(no repetitive characteristics)
Schematic
(simple shape)
Sharp regular schematic
(SRS)
Ijgi 11 00461 i002Sharp
(angles with large deviations)
Regular
(repetitive right angles)
Schematic
(simple shape)
Sharp irregular non-schematic (SIN) Ijgi 11 00461 i003Sharp
(angles with large deviations)
Irregular
(no repetitive characteristics)
Non-schematic
(complex hierarchical bends with various sizes)
Table 2. Confusion matrix and three evaluation metrics for lixel classification of the test samples using 1D-U-Net.
Table 2. Confusion matrix and three evaluation metrics for lixel classification of the test samples using 1D-U-Net.
Automatically
Predicted
SIS PatternSRS PatternSIN PatternPrecision
(%)
Recall
(%)
F 1 - Score
Manually Labeled
SIS pattern11,642127718490.7688.850.90
SRS pattern1083985115786.9288.820.88
SIN pattern102206692395.3195.740.96
Table 3. Consistence ratios (CRs) and overall CRs (OCRs) of the segmentation results for the test boundaries using different methods.
Table 3. Consistence ratios (CRs) and overall CRs (OCRs) of the segmentation results for the test boundaries using different methods.
MethodCR (%)OCR (%)
SIS PatternSRS PatternSIN Pattern
BANN91.5950.2898.5178.61
NB83.5154.2697.8776.50
1D-U-Net91.2390.5497.4092.41
Table 4. Classification accuracies of the proposed model using GSCDs with different grid sizes.
Table 4. Classification accuracies of the proposed model using GSCDs with different grid sizes.
Length of Cell Edge
(m)
Grid Size
(Lixel × Lixel)
Classification Accuracy
(%)
253 × 389.17
4 × 489.58
5 × 590.42
6 × 689.68
7 × 788.53
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, M.; Huang, H.; Zhang, Y.; Yan, X. Pattern Recognition and Segmentation of Administrative Boundaries Using a One-Dimensional Convolutional Neural Network and Grid Shape Context Descriptor. ISPRS Int. J. Geo-Inf. 2022, 11, 461. https://doi.org/10.3390/ijgi11090461

AMA Style

Yang M, Huang H, Zhang Y, Yan X. Pattern Recognition and Segmentation of Administrative Boundaries Using a One-Dimensional Convolutional Neural Network and Grid Shape Context Descriptor. ISPRS International Journal of Geo-Information. 2022; 11(9):461. https://doi.org/10.3390/ijgi11090461

Chicago/Turabian Style

Yang, Min, Haoran Huang, Yiqi Zhang, and Xiongfeng Yan. 2022. "Pattern Recognition and Segmentation of Administrative Boundaries Using a One-Dimensional Convolutional Neural Network and Grid Shape Context Descriptor" ISPRS International Journal of Geo-Information 11, no. 9: 461. https://doi.org/10.3390/ijgi11090461

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop