Next Article in Journal
Rethinking Attention Mechanisms in Vision Transformers with Graph Structures
Next Article in Special Issue
Image Filtering to Improve Maize Tassel Detection Accuracy Using Machine Learning Algorithms
Previous Article in Journal
Mechanical Properties and Fatigue Life Analysis of Motion Cables in Sensors under Cyclic Loading
Previous Article in Special Issue
A Ship Detection Model Based on Dynamic Convolution and an Adaptive Fusion Network for Complex Maritime Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Double-Branch Multi-Scale Contextual Network: A Model for Multi-Scale Street Tree Segmentation in High-Resolution Remote Sensing Images

1
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
2
University of Chinese Academy of Sciences, Beijing 101408, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(4), 1110; https://doi.org/10.3390/s24041110
Submission received: 29 December 2023 / Revised: 25 January 2024 / Accepted: 5 February 2024 / Published: 8 February 2024
(This article belongs to the Special Issue AI-Driven Sensing for Image Processing and Recognition)

Abstract

:
Street trees are of great importance to urban green spaces. Quick and accurate segmentation of street trees from high-resolution remote sensing images is of great significance in urban green space management. However, traditional segmentation methods can easily miss some targets because of the different sizes of street trees. To solve this problem, we propose the Double-Branch Multi-Scale Contextual Network (DB-MSC Net), which has two branches and a Multi-Scale Contextual (MSC) block in the encoder. The MSC block combines parallel dilated convolutional layers and transformer blocks to enhance the network’s multi-scale feature extraction ability. A channel attention mechanism (CAM) is added to the decoder to assign weights to features from RGB images and the normalized difference vegetation index (NDVI). We proposed a benchmark dataset to test the improvement of our network. Experimental research showed that the DB-MSC Net demonstrated good performance compared with typical methods like Unet, HRnet, SETR and recent methods. The overall accuracy (OA) was improved by at least 0.16% and the mean intersection over union was improved by at least 1.13%. The model’s segmentation accuracy meets the requirements of urban green space management.

1. Introduction

Street trees refer to the trees planted on both sides of a road for shade and noise reduction [1]. They are an important part of urban green spaces and urban ecosystems. During the process of street tree management, it is necessary to quickly obtain the information regarding their distribution. There are two traditional ways to obtain this information: one is manual measurement, while the other is using vehicle-mounted LiDAR systems to scan all the street trees along a road [2]. These two methods are inefficient and face difficulty in meeting the requirements of information acquisition speed.
Satellite imaging technology has the advantage of allowing the rapid observation of a large area [3], and satellite remote sensing images are suitable for performing quick segmentation of street trees. However, due to their different species, seasons and growing years [4], street trees show different distribution patterns in satellite images, mainly consisting of concentrated distribution and decentralized distribution [5]. As a result, the sizes of the targets to be segmented are different, so it is necessary to consider both large and small targets in the images used. The methods that researchers have used in their studies in this field are generally based on band calculations, fractional operators, traditional machine learning, etc. [6]. However, these methods cannot demonstrate good performance on both large and small targets. Hong et al. presented two methods of a hierarchical classification technique to distinguish street trees from the neighboring grasslands and roads based on Quickbird images [7]. However, this method easily confused street trees and other trees and missed some small targets, so ensuring its segmentation accuracy proved difficult. Zhao et al. proposed a method for detecting the edges of street trees based on Fourier spectral features [8]. The road center line is pre-extracted using a Gabor filter in frequency domain. Then, the edges of the street trees are extracted according to the buffer of the road center line. The notable problem with this method is that the buffer of the road needs to be set manually. If the buffer is incorrect, the edge of some large targets will be incomplete. Therefore, it is necessary to build a method for the quick segmentation of street trees that considers both large and small targets.
Deep learning has been widely used in many fields, including the segmentation of remote sensing images, because of its excellent performance [9,10]. Since fully convolutional networks (FCNs) first achieved the aim of image segmentation [11], this field has developed rapidly. Many high-performance networks have been proposed, such as Unet [12], the Deeplab series [13], HRnet [14], etc., which have shown good results on different datasets. However, the networks should be improved to consider both small and large targets and adapt to the segmentation of the street trees; their ability to extract multi-scale context features needs to be enhanced.
In deep neural networks, feature information is extracted by the encoder. The encoder in traditional FCN models mainly extracts the local features of the image [15]. However, due to the limited receptive field of convolutional layers, traditional models are ineffective in extracting larger-scale and global features. Some scholars have proposed the use of pluggable modules in the encoder to improve the feature extraction ability and keep more features from loss [16]. This helps the model retain more features, but it shows little improvement in larger-scale applications.
The self-attention-mechanism-based Transformer structure has been applied to image segmentation in recent years [17]. The Transformer structure converts images into one-dimensional sequences for input, paying attention to the global features of images at all stages, but not to the local features of images. Therefore, an encoder combined with both an FCN and Transformer can make up for the shortcomings of both and improve the accuracy of street tree segmentation.
In addition, the brightness of the street trees in the analyzed images differs due to the acquisition time, illumination and other factors, which may reduce the accuracy of multi-scale street tree segmentation. Therefore, it is necessary to introduce other data sources into the segmentation task; some studies have already been carried out in this area [18]. Considering that vegetation index information is easy to obtain from urban green space management, the normalized vegetation index (NDVI) can be input into the network to reduce that impact. The range of the NDVI is [−1, 1], which is different from the RGB bands, so these two parts of information should be input and encoded separately.
To conquer this problem, in this paper, we proposed a Double-Branch Multi-Scale Contextual Network (DB-MSC Net) for the multi-scale segmentation of street trees from high-resolution remote sensing images.
The contributions of this paper are as follows:
  • The DB-MSC Net is proposed to enhance the ability to segment street trees. The overall accuracy is improved by at least 0.16% and the mIoU is improved by at least 1.13% compared to typical networks.
  • We designed a double-branch structure in the network to adapt to the input of both RGB images and the NDVI.
  • The MSC block is proposed to improve the ability to extract multi-scale features. It uses a CNN–Transformer hybrid structure to extract both local and global features.
The rest of this paper is organized as follows. Section 2 summarizes the typical segmentation methods based on deep neural networks. In Section 3, the structure of the DB-MSC Net and MSC block is described in detail. Relevant experiments are presented in Section 4. Discussions on our study are outlined in Section 5. Conclusions are provided in Section 6.

2. Related Work

In this section, we review several existing, representative methods and improvements for multi-scale segmentation in high-resolution remote sensing images based on deep neural networks.
There are two main ways to improve multi-scale segmentation ability. One is improving the CNN-based networks, and the other is using Transformer-based networks.
FCNs were the first networks to realize image segmentation based on deep neural networks. Subsequently, since the precision of multi-scale segmentation is not ideal, researchers began to explore methods of enhancing the ability to extract multi-scale features in neural networks. Chen et al. [19] proposed fully connected CRFs in Deeplab V1 and overcame the localization characteristics of deep neural networks. They proposed Deeplab V2 in 2017 [13], using atrous convolution and a spatial pyramid pooling module to expand the receptive field of the network. In Deeplab V3 and V3+ [20,21], they extended the ASPP module with different dilation rates of atrous convolution, which improved its multi-scale feature extraction ability. In addition, Zhao et al. [22] proposed a pyramid pooling module (PPM), which contains information between different scales and different subregions. Qin et al. [23] proposed an autofocus convolutional layer with parallel convolutional layers possessing different dilation rates and an attention module to learn the weight of different branches, which can adaptively change the size of receptive fields to extract multi-scale features. Gu et al. [24] used two parallel encoders to extract information from different scales and a single decoder to concatenate the information. Tokunaga et al. [25] used three parallel CNNs and weighted concatenation to extract multi-scale information.
The Transformer structure mainly consists of Multi-Head Attention and Feed-Forward Networks [26], showing good performance when extracting global features. Inspired by this advantage, researchers began to improve the Transformer structure for the field of image segmentation. Segmentation Transformer (SETR) [27] first replaced convolutional-layer-based encoders with a pure transformer. The self-attention model is used in global feature learning. Liu et al. [28] proposed the Swin Transformer, which is based on shifted windows to improve computational efficiency and uses hierarchical feature maps to obtain multi-scale feature maps.
In recent years, researchers have conducted many studies on tree extraction using remote sensing images, which can accurately represent the segmentation of street trees. Ye et al. [29] proposed a method of automatically extracting olive crowns, combining RGB images captured by unmanned aerial vehicles (UAVs) and a U2-Net neural network, which achieved high accuracy when extracting tree crown numbers in four typical subareas. Zhang et al. [30] proposed a new method for individual tree segmentation and identification based on the improved Mask R-CNN network. Their results showed that this method has more advantages in broadleaf canopy segmentation and number detection. Schürholz et al. [31] applied convolutional neural networks for instance segmentation to accurately delineate individual tree canopies for certain species and identify the area coverage of some mangrove tree species as well as the surrounding land-cover classes. Lv et al. [32] proposed a novel Mask-CSP-attention-coupled network (MCAN) to enhance detail information detection and improve individual tree detection accuracy. The results showed that this network can perform high-precision segmentation in many contexts. Zheng et al. [33] improved the structure of a High-Resolution Network (HR-Net) algorithm to make it more suitable for forest extraction in remote sensing images. However, the networks mentioned above did not focus on the ability to perform multi-scale feature extraction. Liu et al. [34] proposed a Multi-Scale Channel Spatial Attention Network (MSCSA-Net) and improved the overall accuracy of tree segmentation, but this network only used convolutional layers to build the channel and spatial attention module. It did not apply a Transformer structure and its ability to extract global features can be improved. So, it is important to build a CNN–Transformer hybrid network to enhance the ability to conduct global feature extraction.
In this article, we explore the method of achieving multi-scale street tree segmentation using high-resolution remote sensing images. Our method is a combination of CNN and Transformer which considers both multi-scale local features and global features. The following section provides the architectural details of our proposed model.

3. Materials and Methods

3.1. Dataset and Data Preprocessing

We proposed a benchmark dataset to test the improvement of our network. The experimental data comprised RGB images and NDVI products from the Beijing-3 International Cooperative (BJ-3N) Satellite. This satellite has an orbital altitude of 620 km and a revisit period of 1 day. The resolution of its panchromatic images is 0.3 m, and the multispectral resolution is 1.2 m. The wavelengths of the spectrum are 450–520 nanometers in blue, 530–590 nanometers in green and 620–690 nanometers in red.
Our study used 2 sets of remote sensing products. Each set contained one RGB image without a vegetation index and one NDVI. These two sets of products were both taken over Lhasa, Tibet Autonomous Region, China. One was acquired on 4 November 2021, and the other was acquired on 29 January 2022. We selected 12 street blocks in these two images and marked all the street trees as the label. All the images, NDVI and labels were clipped to fit the size of 256 × 256. We used random shifts, random rotations and horizontal and vertical flips for data augmentation. For RGB images, we used a random brightness jitter of up to ±10%. After that, the dataset was divided into the training set, validation set, and test set in the ratio of 8:1:1. Each set contained pictures with and without street trees. The number of each kind of image in the datasets is shown in Table 1. Figure 1 uses three typical scenarios to show the details of this dataset. The dataset is publicly available on https://pan.baidu.com/s/1xUp1Davs2-58i_oJxGUh7w?pwd=ngg3 (accessed on 25 December 2023).

3.2. The Overall Architecture of DB-MSC Net

The overall architecture of DB-MSC Net is shown in Figure 2. The numbers under each block represent the size or channels of the feature. As shown in the Figure, the DB-MSC Net model has five stages. For the first stage, we designed two input layers to encode image features without vegetation A 1 and vegetation features B 1 separately because of the different ranges of RGB images and the NDVI. A 1 was input into the MSC block at this stage and the output was F A 1 . Similarly, the output of B 1 was F B 1 . The process can be expressed by the following formula.
F A 1 = m s c ( A 1 , A 1 )
F B 1 = m s c ( B 1 , B 1 )
Starting from the second stage, the input RGB image and NDVI were 2 × 2 average-pooled from the previous stage. The features output from the previous stage, F A i 1 and F B i 1 , were also 2 × 2 max-pooled and then input into the MSC block together with the image or NDVI. This step is formulated as follows.
F A i = m s c M a x p o o l F A i 1 , A i i = 2 , 3 , 4 , 5
F B i = m s c M a x p o o l F B i 1 , B i i = 2 , 3 , 4 , 5
After encoding, the multi-scale image features F A i and vegetation features F B i can be extracted.
In the decoding process, first, F A 5 and F B 5 were concatenated by channel to form the feature Q 5 .
Q 5 = [ F A 5 , F B 5 ]
After 2 × 2 bilinear interpolation upsampling, 3 × 3 convolution and ReLU activation function operations for Q 5 , the weight of each channel was calculated and added through the channel attention mechanism (CAM) module to form the feature P 4 . Then, F A 4 , P 4 and F B 4 were concatenated by channel. These operations were then repeated for each stage. This step is formulated as follows.
P i = C A M U p C o n v ( Q i + 1 ) i = 1 , 2 , 3 , 4
Q i = F A i , P i , F B i i = 1 , 2 , 3 , 4
The CAM function is defined in Section 3.4. At the end of the model, Q 1 passes through a 3 × 3 convolutional layer to adjust the number of channels to output the result.

3.3. The Structure of the MSC Block

In DB-MSC Net, we propose a new encoding module, an MSC block, which is important for extracting the multi-scale features of street trees from remote sensing images and vegetation information. This section introduces the detailed structure of this module.
Taking the MSC block in the middle stage of the network as an example, the input of this module consists of two parts. One is the average pooled RGB image A i or NDVI B i , and the other is the max pooled output feature from the MSC block in the previous stage, F A i or F B i . The MSC block extracts the global features from the input image or NDVI and the local feature from the output of the previous MSC block. Finally, the global feature and the local feature are combined to obtain the multi-scale feature as the output of the MSC block.
The input RGB image or NDVI is first cut into a 16 × 16 size and linear-projected to a 2D matrix with a width of 256. Then, a zero matrix of the same size is initialized and added to that matrix through positional encoding and the matrix x is obtained. x is then mapped to two different subspaces using the learnable matrices W Q , W K and W V to obtain the matrices query ( q 1 , q 2 ), key ( k 1 , k 2 ) and value ( v 1 , v 2 ). The formula is as follows.
q i = W Q x i i = 1 , 2
k i = W K x i i = 1 , 2
v i = W V x i i = 1 , 2
Matrix q and the transposition of the matrix k are used for point multiplication. After being multiplied with d q , the square root of the dimensions of the matrix k and softmax function, the similarity probability α is obtained. α is multiplied with v to obtain the attention weight matrix y . Finally, after layer normalization, MLP and residual addition, the global feature matrix z is obtained. It is formulated as follows.
α 1 , i = s o f t m a x q 1 T · k i d q , k i = 1 , 2
α 2 , i = s o f t m a x q 2 T · k i d q , k i = 1 , 2
y 1 = i = 1 2 α 1 , i · v i
y 2 = i = 1 2 α 2 , i · v i
The Transformer module is repeated 6 times. After fully extracting global features, the feature is reshaped to the size same as the input image or NDVI and the matrix F 0 is obtained.
Matrix F A i 1 is input to the MSC block after 2 × 2 max pooling. The input matrix first goes through a 5-branch convolution module. The first four convolutional layers have a kernel size of 3 and stride of 1. One convolutional layer is not dilated, and the other 3 convolutional layers have dilation rates of 2, 3 and 5, respectively. The output matrices of each of these four branches, F 1 , F 2 , F 3 and F 4 , have the same length and width as matrix F A i 1 and the number of channels is half of F A i 1 . The last convolutional layer has a kernel size of 1 and stride of 1. The output matrix of this layer F 5 has the same length, same width and twice the channels as F A i 1 . Then, F 1 , F 2 , F 3 and F 4 are concatenated by channel and F 5 is added to obtain local features. Finally, the local and global features are concatenated by channel and the output multi-scale feature F A i is obtained. It is formulated as follows.
F A i = F 1 , F 2 , F 3 , F 4 + F 5 , F 0
The architecture of MSC block is shown in Figure 3.

3.4. Channel Attention Mechanism (CAM)

The channel attention mechanism (CAM) [35] is used to selectively emphasize informative features while suppressing less-useful features in RGB images and NDVI. Figure 4 shows the architecture of the CAM.
First, the input feature map is compressed in channel dimensions using max pooling and average pooling to obtain two one-dimensional vectors. Then, the multilayer perceptron (MLP) network is used, and vectors with weight value are generated. The two vectors are added and calculated using the sigmoid function σ . Finally, the vector is multiplied with the input feature and the weighted feature is output, which can be formulated as follows. F is the input feature and F is the output feature.
F = C A M F = σ ( M L P A v g p o o l F + M L P ( M a x p o o l ( F ) ) )

3.5. Experimental Design and Criteria

In our study, all experiments were implemented with Tensorflow 2.10.1 on an NVIDIA Geforce RTX 3080 Graphics Processing Unit (GPU) produced by ASUS in Shenzhen, China. CUDA 11.2 toolkit and cuDNN 8.1 were used to speed up the training process.
The DB-MSC model was compared with FCN-8s [11], Deeplab V3+ [21], PSPNet [22], UNet [12], HRNet [14], SETR [27] and MSCSA-Net [34]. We trained all networks using an Adam optimizer with a beta1 of 0.9, beta2 of 0.999 and adaptive learning rate. We initialized the learning rate at 0.001. The number of overall training iterations was 100 and each GPU batch size is 5. We used the CrossEntropy loss function to calculate the loss of these models. The loss is defined as follows. The experiment was repeated five times and averaged.
L = y t r u e log y p r e d ( 1 y t r u e ) l o g ( 1 y p r e d )
We conducted an ablation study to verify the effects of the improvements on the encoder. Three networks, Double-Branch UNet (DB-UNet), Double-Branch Channel Attention Mechanism UNet (DB-CAM-UNet) and DB-MSC Net, were trained to compare the improvements. The DB-UNet is a double-branch UNet-like network which allows for the separate input of the RGB image and NDVI. The DB-CAM-UNet adds CAM to the decoder based on DB-UNet.
The training hyperparameters of these three networks were the same as in the contrast experiment. The experiment was repeated five times and averaged.
The results of segmentation can be divided into 4 categories: True Positive (TP) refers to the pixels that are correctly segmented as street trees. False Positive (FP) refers to the pixels that are segmented as street trees but labeled as non-street trees. True Negative (TN) refers to the pixels that are correctly segmented as non-street trees. False Negative (FN) refers to the pixels that are segmented as non-street trees but labeled as street trees. To evaluate our method, four metrics, OA, F1, mIoU and Kappa, were used to evaluate the performance of each segmentation method. OA is the ratio of correctly segmented pixels to the total number of pixels. The F1-score (F1), mIoU and Kappa are defined as follows.
O A = T P + T N T P + T N + F P + F N
F 1 = T P 2 T P + F P + F N
m I o U = T P T P + F P + F N
p 0 = T P + T N
p e = T P + F P T P + F N + F P + T N F N + T N
K a p p a = p 0 p e 1 p e

4. Results

4.1. Network Training

The overall accuracy curves of the DB-MSC Net on the training set and verification set are shown in Figure 5. Loss curves of the DB-MSC Net in the training set and verification set are shown in Figure 6.
Figure 5 and Figure 6 show that the network converged after training on our dataset for 100 epochs.

4.2. Street Tree Segmentation

The quantitative results of different networks on our benchmark dataset are shown in Table 2.
The results show that compared to the networks using only convolutional layers like FCN-8s, PSPNet, Deeplab V3+, UNet, HRNet and MSCSA-Net, DB-MSC Net increased the mIoU by at least 1.13%, increased F1 by at least 1.42% and increased Kappa by at least 2.47%. Compared with networks using only Transformer in the encoder, such as SETR, the DB-MSC Net increased the mIoU by 5.11%, increased F1 by 4.57% and increased Kappa by 8.85%.
Four high-resolution remote sensing images with street trees were taken as segmentation examples. Original images, ground truth and segmentation results using FCN-8s, PSPNet, Deeplab V3+, UNet, HRNet, SETR, MSCSA-Net and our DB-MSC Net are shown in Figure 7.
Figure 7 shows that the networks using only convolutional layers under-segmented the large-size street trees. This is more obvious in the last image since the results of these networks are fragmented. The SETR network under-segmented the small-size street trees. In the first three images, it ignored some small street trees. However, the DB-MSC Net was not affected by the different sizes of street trees and showed a good performance. In all these images, the segmentation results of our method were closer to the ground truth.

4.3. Ablation Study

Table 3 shows the quantitative results of the ablation study, and the segmentation result examples are shown in Figure 8.
The quantitative results show that compared to the DB-UNet, the DB-CAM-UNet increased the mIoU by 0.19%, increased F1 by 0.92% and increased Kappa by 0.79%. The improvement was not obvious. However, our DB-MSC Net increased the mIoU by 2.33%, increased F1 by 4.53% and increased Kappa by 3.95%, which showed an obvious improvement.
Figure 8 shows that the DB-UNet under-segmented and over-segmented images. For example, some other trees were incorrectly identified as street trees in the first image. In the other three images, some street trees were not accurately segmented. The DB-CAM-UNet showed some improvements but still under-segmented images. In the second and the fourth image, some large-size street trees were not accurately segmented. By contrast, our DB-MSC Net showed obvious improvements in both large- and small-size street trees. The experiment shows that our improvements to the network were effective.

5. Discussion

There are significant challenges in the segmentation of street trees because of the different sizes of the targets. In this paper, the DB-MSC Net was proposed for the multi-scale segmentation of street trees to improve accuracy. The DB-MSC Net has two branches for the separate input of an RGB image and the NDVI. In the encoding path, MSC blocks were used to replace convolutional blocks. The MSC block combined dilated convolution and Transformer, which enabled the extraction of both local and global features. To further improve the performance of the network, the channel attention model (CAM) was embedded into the decoder.
Compared to the traditional networks, the overall accuracy of DB-MSC Net showed an improvement of at least 1.13% in mIoU. From the segmentation results, the DB-MSC Net proposed in this paper showed better performance than networks such as FCN-8s, PSPNet, Deeplab V3+, UNet, HRNet, MSCSA-Net and SETR. Street trees of different sizes were accurately segmented.
The DB-MSC Net accurately segmented multi-scale street trees in high-resolution remote sensing images. However, the method we proposed in this paper did not show a good effect for some street trees in shadow-covered areas. In areas with dense buildings, the segmentation effect was worse due to the shadows of the buildings. This method presents great challenges in segmenting such areas.
In future work, we will focus on the problem of street tree segmentation in areas covered by the shadows of buildings. We will explore appropriate methods to reduce the effects of shadows.

6. Conclusions

This paper focused on the problem of multi-scale segmentation of street trees in high-resolution remote sensing images. We proposed the DB-MSC Net method for performing multi-scale segmentation. The experimental results showed that the proposed method for segmenting street trees in high-resolution remote sensing images had improved accuracy.
The contrast experiment and ablation study verified the effectiveness of the proposed network by comparing it with two different improved network stages and six typical segmentation models using our BJ-3N satellite image dataset. The results met the accuracy requirements for urban green space management.

Author Contributions

Conceptualization, H.Z. and S.L.; methodology, H.Z.; software, H.Z.; validation, H.Z.; formal analysis, H.Z.; investigation, H.Z.; resources, S.L.; data curation, H.Z.; writing—original draft preparation, H.Z.; writing—review and editing, S.L.; visualization, H.Z.; supervision, S.L.; project administration, S.L.; funding acquisition, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Aerospace Information Research Institute (Y9B0640HM3).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, Y.; Wu, Y.; Sun, Q.; Hu, C.; Liu, H.; Chen, C.; Xiao, P. Tree failure assessment of london plane (Platanus × acerifolia (aiton) willd.) street trees in nanjing city. Forests 2023, 14, 1696. [Google Scholar] [CrossRef]
  2. Yadav, M.; Khan, P.; Singh, A.K.; Lohani, B. Generating GIS database of street trees using mobile lidar data. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2018, IV-5, 233–237. [Google Scholar] [CrossRef]
  3. Shahtahmassebi, A.R.; Li, C.; Fan, Y.; Wu, Y.; Lin, Y.; Gan, M.; Wang, K.; Malik, A.; Blackburn, G.A. Remote sensing of urban green spaces: A review. Urban For. Urban Green. 2021, 57, 126946. [Google Scholar] [CrossRef]
  4. Branson, S.; Wegner, J.D.; Hall, D.; Lang, N.; Schindler, K.; Perona, P. From Google Maps to a fine-grained catalog of street trees. ISPRS J. Photogramm. Remote Sens. 2018, 135, 13–30. [Google Scholar] [CrossRef]
  5. Breuste, J.H. Investigations of the urban street tree forest of Mendoza, Argentina. Urban Ecosyst. 2013, 16, 801–818. [Google Scholar] [CrossRef]
  6. Zhang, X.; Boutat, D.; Liu, D. Applications of fractional operator in image processing and stability of control systems. Fractal Fract. 2023, 7, 359. [Google Scholar] [CrossRef]
  7. Hong, Z.H.; Xu, S.; Wang, J.; Xiao, P.F. Extraction of urban street trees from high resolution remote sensing image. In Proceedings of the 2009 Joint Urban Remote Sensing Event, Shanghai, China, 20–22 May 2009; Volume 1–3, pp. 1510–1514. [Google Scholar]
  8. Zhao, H.H.; Xiao, P.F.; Feng, X.Z. Edge detection of street trees in high-resolution remote sensing images using spectrum features. In Proceedings of the MIPPR 2013: Automatic Target Recognition and Navigation, Wuhan, China, 26 October 2013. [Google Scholar]
  9. Ma, L.; Liu, Y.; Zhang, X.; Ye, Y.; Yin, G.; Johnson, B.A. Deep learning in remote sensing applications: A meta-analysis and review. ISPRS J. Photogramm. Remote Sens. 2019, 152, 166–177. [Google Scholar] [CrossRef]
  10. Zhang, J.X.; Yang, T.; Chai, T. Neural network control of underactuated surface vehicles with prescribed trajectory tracking performance. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–14. [Google Scholar] [CrossRef] [PubMed]
  11. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. arXiv 2017, arXiv:1411.4038. [Google Scholar]
  12. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin, Germany, 2015; pp. 234–241. [Google Scholar]
  13. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef]
  14. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 5686–5696. [Google Scholar]
  15. Kattenborn, T.; Leitloff, J.; Schiefer, F.; Hinz, S. Review on convolutional neural networks (CNN) in vegetation remote sensing. ISPRS J. Photogramm. Remote Sens. 2021, 173, 24–49. [Google Scholar] [CrossRef]
  16. Zhang, S.; Zhang, C. Modified U-Net for plant diseased leaf image segmentation. Comput. Electron. Agric. 2023, 204, 107511. [Google Scholar] [CrossRef]
  17. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  18. Yan, H.; Zhang, J.X.; Zhang, X. Injected infrared and visible image fusion via L_1 decomposition model and guided filtering. IEEE Trans. Comput. Imaging 2022, 8, 162–173. [Google Scholar] [CrossRef]
  19. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected CRFs. arXiv 2014, arXiv:1412.7062. [Google Scholar]
  20. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  21. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Computer Vision—ECCV 2018, Proceedings of the 15th European Conference, Munich, Germany, 8–14 September 2018; Springer: Cham, Switzerland, 2018; pp. 833–851. [Google Scholar]
  22. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6230–6239. [Google Scholar]
  23. Qin, Y.; Kamnitsas, K.; Ancha, S.; Nanavati, J.; Cottrell, G.; Criminisi, A.; Nori, A. Autofocus layer for semantic segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2018, Proceedings of the 21st International Conference, Granada, Spain, 16–20 September 2018; Springer: Cham, Switzerland, 2018; pp. 603–611. [Google Scholar]
  24. Gu, F.; Burlutskiy, N.; Andersson, M.; Wilén, L.K. Multi-resolution networks for semantic segmentation in whole slide images. In Computational Pathology and Ophthalmic Medical Image Analysis, Proceedings of the First International Workshop, COMPAY 2018, and 5th International Workshop, OMIA 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 16–20 September 2018; Springer: Cham, Switzerland, 2018; pp. 11–18. [Google Scholar]
  25. Tokunaga, H.; Teramoto, Y.; Yoshizawa, A.; Bise, R. Adaptive Weighting Multi-Field-Of-View CNN for Semantic Segmentation in Pathology. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 12589–12598. [Google Scholar]
  26. Xiao, H.; Li, L.; Liu, Q.; Zhu, X.; Zhang, Q. Transformers in medical image segmentation: A review. Biomed. Signal Process. Control 2023, 84, 104791. [Google Scholar] [CrossRef]
  27. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P.H.S.; et al. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 6877–6886. [Google Scholar]
  28. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar]
  29. Ye, Z.; Wei, J.; Lin, Y.; Guo, Q.; Zhang, J.; Zhang, H.; Deng, H.; Yang, K. Extraction of olive crown based on UAV visible images and the U2-Net deep learning model. Remote Sens. 2022, 14, 1523. [Google Scholar] [CrossRef]
  30. Zhang, C.; Zhou, J.; Wang, H.; Tan, T.; Cui, M.; Huang, Z.; Wang, P.; Zhang, L. Multi-species individual tree segmentation and identification based on improved mask R-CNN and UAV imagery in mixed forests. Remote Sens. 2022, 14, 874. [Google Scholar] [CrossRef]
  31. Schürholz, D.; Castellanos-Galindo, G.A.; Casella, E.; Mejía-Rentería, J.C.; Chennu, A. Seeing the forest for the trees: Mapping cover and counting trees from aerial images of a mangrove forest using artificial intelligence. Remote Sens. 2023, 15, 3334. [Google Scholar] [CrossRef]
  32. Lv, L.; Li, X.; Mao, F.; Zhou, L.; Xuan, J.; Zhao, Y.; Yu, J.; Song, M.; Huang, L.; Du, H. A deep learning network for individual tree segmentation in UAV images with a coupled CSPNet and attention mechanism. Remote Sens. 2023, 15, 4420. [Google Scholar] [CrossRef]
  33. Zheng, Z.; Yu, S.; Jiang, S. A domain adaptation method for land use classification based on improved HR-Net. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4400911. [Google Scholar] [CrossRef]
  34. Liu, K.-H.; Lin, B.-Y. MSCSA-Net: Multi-scale channel spatial attention network for semantic segmentation of remote sensing images. Appl. Sci. 2023, 13, 9491. [Google Scholar] [CrossRef]
  35. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Computer Vision—ECCV 2018, Proceedings of the 15th European Conference, Munich, Germany, 8–14 September 2018; Springer: Cham, Switzerland; pp. 3–19.
Figure 1. Typical scenarios in the dataset.
Figure 1. Typical scenarios in the dataset.
Sensors 24 01110 g001
Figure 2. The overall architecture of DB-MSC Net.
Figure 2. The overall architecture of DB-MSC Net.
Sensors 24 01110 g002
Figure 3. The architecture of the MSC block.
Figure 3. The architecture of the MSC block.
Sensors 24 01110 g003
Figure 4. The architecture of the CAM.
Figure 4. The architecture of the CAM.
Sensors 24 01110 g004
Figure 5. Overall accuracy curves of the DB-MSC Net.
Figure 5. Overall accuracy curves of the DB-MSC Net.
Sensors 24 01110 g005
Figure 6. Loss curves of the DB-MSC Net.
Figure 6. Loss curves of the DB-MSC Net.
Sensors 24 01110 g006
Figure 7. Segmentation results for high-resolution remote sensing images with street trees using different methods.
Figure 7. Segmentation results for high-resolution remote sensing images with street trees using different methods.
Sensors 24 01110 g007
Figure 8. Segmentation results of ablation study.
Figure 8. Segmentation results of ablation study.
Sensors 24 01110 g008
Table 1. Number of images in the dataset.
Table 1. Number of images in the dataset.
CategoryTraining SetValidation SetTest SetTotal
Contain street trees71108858908885
No street trees8901151101115
Total80001000100010,000
Table 2. Performance of different networks.
Table 2. Performance of different networks.
ModelOA (%)mIoU (%)F1 (%)Kappa (%)
FCN-8s [11]95.0563.8171.1342.69
PSPNet [22]95.0764.6272.0544.48
Deeplab V3+ [21]95.8169.2175.9852.28
UNet [12]95.9869.7876.0552.43
HRNet [14]95.4466.6773.5147.50
SETR [27]94.1965.8473.0946.78
MSCSA-Net [34]95.8769.8276.2453.16
DB-MSC Net (Ours)96.1470.9577.6655.63
Table 3. Quantitative results of ablation study.
Table 3. Quantitative results of ablation study.
ModelOA (%)mIoU (%)F1 (%)Kappa (%)
DB-UNet95.6568.6273.1351.68
DB-CAM-UNet95.8768.8174.0552.47
DB-MSC Net (Ours)96.1470.9577.6655.63
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, H.; Liu, S. Double-Branch Multi-Scale Contextual Network: A Model for Multi-Scale Street Tree Segmentation in High-Resolution Remote Sensing Images. Sensors 2024, 24, 1110. https://doi.org/10.3390/s24041110

AMA Style

Zhang H, Liu S. Double-Branch Multi-Scale Contextual Network: A Model for Multi-Scale Street Tree Segmentation in High-Resolution Remote Sensing Images. Sensors. 2024; 24(4):1110. https://doi.org/10.3390/s24041110

Chicago/Turabian Style

Zhang, Hongyang, and Shuo Liu. 2024. "Double-Branch Multi-Scale Contextual Network: A Model for Multi-Scale Street Tree Segmentation in High-Resolution Remote Sensing Images" Sensors 24, no. 4: 1110. https://doi.org/10.3390/s24041110

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop