Next Article in Journal
Automatic Bluefin Tuna Sizing with a Combined Acoustic and Optical Sensor
Previous Article in Journal
Ultrasonic Guided Wave Testing on Cross-Ply Composite Laminate: An Empirical Study
Previous Article in Special Issue
Computer Vision System for Welding Inspection of Liquefied Petroleum Gas Pressure Vessels Based on Combined Digital Image Processing and Deep Learning Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

NAS-HRIS: Automatic Design and Architecture Search of Neural Network for Semantic Segmentation in Remote Sensing Images

1
College of Information and Computer Engineering, Northeast Forestry University, Harbin 150040, China
2
College of Computer Science and Engineering, Xi’an University of Technology, Xi’an 710048, China
3
Faculty of Applied Mathematics, Silesian University of Technology, 44-100 Gliwice, Poland
4
Department of Applied Informatics, Vytautas Magnus University, 44404 Kaunas, Lithuania
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(18), 5292; https://doi.org/10.3390/s20185292
Submission received: 20 July 2020 / Revised: 5 September 2020 / Accepted: 14 September 2020 / Published: 16 September 2020
(This article belongs to the Special Issue Computer Vision for Remote Sensing and Infrastructure Inspection)

Abstract

:
The segmentation of high-resolution (HR) remote sensing images is very important in modern society, especially in the fields of industry, agriculture and urban modelling. Through the neural network, the machine can effectively and accurately extract the surface feature information. However, using the traditional deep learning methods requires plentiful efforts in order to find a robust architecture. In this paper, we introduce a neural network architecture search (NAS) method, called NAS-HRIS, which can automatically search neural network architecture on the dataset. The proposed method embeds a directed acyclic graph (DAG) into the search space and designs the differentiable searching process, which enables it to learn an end-to-end searching rule by using gradient descent optimization. It uses the Gumbel-Max trick to provide an efficient way when drawing samples from a non-continuous probability distribution, and it improves the efficiency of searching and reduces the memory consumption. Compared with other NAS, NAS-HRIS consumes less GPU memory without reducing the accuracy, which corresponds to a large amount of HR remote sensing imagery data. We have carried out experiments on the WHUBuilding dataset and achieved 90.44% MIoU. In order to fully demonstrate the feasibility of the method, we made a new urban Beijing Building dataset, and conducted experiments on satellite images and non-single source images, achieving better results than SegNet, U-Net and Deeplab v3+ models, while the computational complexity of our network architecture is much smaller.

1. Introduction

In recent years, with the progress and popularization of remote sensing technology, satellite imaging and aerial photography are becoming more and more advanced [1]. We can get images which contain large amounts of information. These images have been applied in many fields, like agriculture [2], forestry, geology, military, environmental protectio n [3], urban planning [4], etc. High-resolution (HR) remote sensing images include high spatial, temporal and spectral resolution. The HR remote sensing image in this paper mainly refers to the high spatial resolution (2 m resolution and better) remote sensing image. The high spatial resolution remote sensing images capture the surface of the earth in great detail. With the increasing spatial resolution of remote sensing images, there is a need to improve and innovate the method of analyzing remote sensing images.
Image segmentation of remote sensing images can be used for land area estimation, fire monitoring, urban planning, crop detection and yield modelling and many other applications [5,6]. Moreover, it is essential for observing the growth and evolution of complex urban systems, including slum detection, suburban growth, change in temperature in urban heat island, identifying disaster-damaged urban infrastructures, etc. [7,8,9,10].
Image segmentation aims to partition an image into homogenous regions such that no union of two adjacent regions is homogenous [11]. Remote sensing image segmentation has always been an important part of the remote sensing preprocessing process; how to improve the segmentation accuracy is always a difficult point [11]. Traditional HR remote sensing image segmentation is classified into four categories according to the principle of segmentation: the first is pixel-based algorithms, including the simplest segmentation algorithm—thresholding algorithm and cluster algorithm [12]. The second is boundary-based algorithms; the boundary-based algorithm usually looks for the sharp transition of gray value in the image to determine the boundary of the object region. The third is region-based algorithms, which are mainly divided into local regions based on the similarity between adjacent pixels to achieve segmentation. The fourth is physical-model-based algorithms, and the physical model is obtained from the imaging procedure, which describes the relationship between images and factual detail of the Earth’s surface [13].
With the popularization of remote sensing image segmentation and the development of artificial intelligence, the data-driven methods are getting more attention. Remote sensing image segmentation has attracted more attention from the computer vision and machine learning community. Convolutional Neural Networks (CNNs) have achieved state-of-the-art results in many computer vision tasks, which bring semantic segmentation into a new era [14,15,16]. As an improved architecture of CNNs, Fully Convolutional Network (FCN) demonstrated the state-of-the-art results for semantic image segmentation. FCN adopts deconvolution filter to conduct up-sampling on the feature map of the ultimate convolutional Layer. Compared to CNNs, FCN can recognize images at the pixel level and ensure robustness and accuracy simultaneously [17,18].
U-Net innovatively adopts the encoder–decoder architecture for semantic segmentation, i.e., the first half is divided into feature extraction and the second half is divided into upper sampling. U-net employs a totally different feature fusion method where features are spliced together in channel dimension to form a thicker feature [19,20]. Furthermore, in addition to the encoder–decoder structure, the fully connected Conditional Random Field (CRF), Atrous Convolution, Atrous Spatial Pyramid Pooling (ASPP), depth-separated convolution, and Xception technique are applied to the models in Deeplab family. This effectively improves the accuracy of boundary segmentation and the speed of training [21,22,23].
Although the above methods based on deep learning greatly enhance the accuracy and efficiency of remote sensing image segmentation [24,25,26], a robust model usually requires relevant experts to spend a lot of time and energy to complete it. Feature extraction and fusion are key for robust and effective image processing in remote sensing [27]. Especially due to the diversity of sources for remote sensing images, and the fact that the image features obtained by different methods are quite different [28,29], a method is required that can automatically search the optimal architecture for different data. The emergence of Neural Architecture Search (NAS) solves this pain point.
As an important derivative of automatic machine learning (Auto-ML), it replaces the manual process of architecture design as the machine’s automatic search for the neural architecture. MIT [30] and Google [31] proposed using reinforcement learning in 2016 to let computers automatically search for neural network architectures. The model obtained from the NAS achieves good accuracy in the image classification task. However, the initial neural architecture search required a large number of computing resources. For example, Google conducted an architecture search on the CIFAR-10 dataset, used 800 Graphical Processing Units (GPUs) and trained for 28 days. Such high computational costs make the work of ordinary researchers unrealistic. Therefore, how to reduce the cost of the search has become a problem that the NAS has had to face since its birth. Researchers have done a lot of work in recent years to get rid of high memory consumption [32,33].
Before doing an architectural search, we need to define the search space. The common search space is chained and it is formed by stacking lays with operators. Many deep neural networks have many similar parts, which are gradually abstracted into a cell, so the search space is greatly simplified. A cell is usually designed as a directed acyclic graph (DAG) [32,34,35,36].
There are three main types of search strategy. The first one is based on reinforcement learning. The generation of the architecture is regarded as an agent choosing the action, and the reward is obtained through the effect prediction function on a test set [30,31]. The second type of strategy is based on Genetic Algorithm (GA), a derivative-free optimization algorithm that may yield a global optimal solution, but is less efficient relatively [36,37]. The gradient-based method makes discrete search space continuous, and the objective function becomes a differentiable function, making it possible to use a gradient-based optimization method to find the optimal structure. The cell-based search space was applied into our works, and we use the gradient descent search strategy to search the space [32].
Here, we propose an improved HR remote sensing image segmentation method based on a neural architecture search, named NAS-HRIS. We applied NAS-HRIS to three different types of HR remote sensing dataset, to efficiently search out suitable architectures themselves.
Summarizing, our contributions are listed as follows:
  • The NAS of the HR remote sensing image segmentation is explored for the first time;
  • Our work embeds DAG into the search space and designs the differentiable searching process, which enables learning an end-to-end searching rule by using gradient descent optimisation [38]. We use the Gumbel-Max trick to provide an efficient way to draw samples from a non-continuous probability distribution, and it improves the efficiency of searching and reduces the memory consumption;
  • We provide a new HR remote sensing image segmentation dataset: the Beijing building datasets (BBD) that can be useful for image segmentation applications such as building segmentation for urban planning; (Figure 1)
  • Conducted search on a variety of remote sensing images, and training was conducted in aerial images, satellite images and Google earth image, obtaining and we got 98.52% pix accuracy, and 90.44% Mean Intersection over Union (MIoU) by using NAS-HRIS on the WHU dataset.
Other parts of this paper are structured as follows. In Section 2, we provide our proposed methodologies in detail. The datasets, experimental settings and comparison results are presented in Section 3. At last, we discuss our work and put forward prospects for the future work in Section 4. We have released our code at https://github.com/zhangmingwei98/NAS-HRIS.

2. Methodology

In this article, we used NAS to construct the architecture of the encoder for the segmentation model Figure 2. The neural architecture search consists of three parts: search space design, search strategy formulation and evaluation method. We defined a search space composed of several cells, and we used the search strategy of gradient descent to select the weights of each edge of the directed acyclic graph, and so used the Gumbel-max trick to do continuous relaxation.

2.1. Architecture Search Space

2.1.1. Cell Level

In NAS-HRIS, all cells are represented as a DAG (see Figure 3): the node of the graph stands for the input image or feature map, and the edge of the graph represents the operation, such as convolution and pooling. Each DAG consists of seven nodes: two are input nodes, four are intermediate nodes, and one is the output node. The output node is defined as the concatenation of four intermediate nodes as in Figure 4 and Figure 5. Our cell is designed according to [32,34,35,36]. The preorder node n i becomes the subsequent node n j after the calculation of operation p as follows
N ( j ) = i < j p i , j ( N ( i ) ) s . t . p i , j P i , j
In NAS-HRIS, the candidate operations set P has nine operations: (1) identity, (2) 3 × 3 avg pooling, (3) 3 × 3 max pooling, (4) 3 × 3 separate conv, (5) 5 × 5 separate conv, (6) 7 × 7 separate conv, (7) 3 × 3 dilated spearate conv, (8) 5 × 5 dilated spearate conv, (9) none.

2.1.2. Network Level

We look for two different cells, i.e., a normal one and a reduction one. They are similar in structure, and their feature maps are padded. However, there is a difference between normal and reduction cells. The stride of all operations is set to 1 for the normal cell, whereas the stride is set to 2 for all operations at the reduction cell. The purpose of reduction cell is to reduce the feature map resolution.
In NAS-HRIS, a cell is treated as the basic block and stacked by certain rules to form neural network. We also apply DAG to structure the network topology. The two input nodes of cell C e l l k are the output nodes of the preorder C e l l k 1 and C e l l k 2 , respectively. Convolutions of 1 × 1 are filled in where necessary. In the network, reduction cells were set in the location of 1/3 and 2/3 of the total network depth. We define architecture variable as α and the weight of architecture as ω . α can be composed of α n o r m a l and α r e d u c t i o n , α n o r m a l and α r e d u c t i o n are shared by all the normal and reduction cells, respectively. In our work, we search for α n o r m a l and α r e d u c t i o n values. NAS-HRIS selects the optimal operations from candidate operations according to the weight value in the search procedure. In the training procedure, we update the value of the selected operation by gradient descent.

2.2. Continuous Relaxation and Search Strategy

As we can see the search space in Figure 3, before the NAS-HRIS search architecture, the operation of each edge in DAG is unknown (a). We set up a certain number of candidate operations on each edge to continuous relaxation of the search space (b). Each edge of the finalized by applying the reparameterization trick to sampling (c,d).
Our goal is to gain the optimal architecture α * and its weight ω * within all operations. We introduced the loss function L to achieve our goal. L t r a i n and L v a l i d are train loss and valid loss, respectively. We regard this problem as a bi-level optimization problem. We find α * that minimizes L v a l i d ( α * , ω * ) in the case of obtaining the optimal weight ω α * , as we can see in (2) and (3).
min α L v a l i d ( w α * , α )
s . t . ω α * = arg min ω L t r a i n ( ω , α )
An architecture α consists of many repeating cells: λ i , j p is the p-th element of a P -demensional learnable α i , j . We adopted the softmax function to get normalized probability f i , j p for sampled operation p between N i and N j . The process of selection a operation was relaxed, as can be seen in (4).
f i , j p = exp ( λ i , j p ) p P exp ( λ i , j p )
In order to back-propagate gradient though λ i , j , we propose using the Gumbel-Max trick [39,40] to re-formulate Equation (1), which makes it possible to sample from a discrete probability distribution in an efficient way, as can see in (5) and (6). This method is proposed to perform NAS for the first time in GDAS [41]. DARTS needs to keep all intermediate results in memory, but the Gumbel-Max trick selects only one operation at a time. Therefore, if there are P candidate operations, the computing resource consumption is about 1 / P . Because the search efficiency of DARTS is mainly limited by memory resources, NAS-HRIS has a faster search speed in an environment with the same memory
N j = i = 1 j p P φ i , j p p ( N i ; ω i , j p )
s . t . φ i , j = 1 , ( i , j ) = arg max ( λ i , j p + ς p ) 0 , o t h e r w i s e
where ς p are Gumbel-distributed noise which are identically distributed and independently drawn samples from G umbel ( 0 , 1 ) 1 in (7). The φ i , j vector we obtained is a o n e _ h o t vector, and we multiply this vector by the range vector of x, and we end up with the x that we’re sampling. ω i , j p is the weight of operation p P between N i and N j .
ς p = log ( log ( u ) ) w i t h u U n i f o r m [ 0 , 1 ]
We apply s o f t m a x to relax a r g m a x in Equation (6), hence Equation (5) is differentiable. We replace φ i , j p with approximately φ ˜ i , j p . This makes Equation (5) differentiable in back-propagation
φ ˜ i , j p = exp ( ( λ i , j p + ς p ) / τ ) p P exp ( ( λ i , j p + ς p ) / τ )
where τ is the softmax temperature.
NAS-HRIS use gradient descent to optimize L v a l i d , similar to using RL or evolutionary architecture search, where validation set performance is seen as reward or fitness. See Algorithm 1 for the detailed searching process, which uses the gradient descent method to fine-tune α and ω
ω = ω ξ ω L t r a i n ( ω * , α )
α = α ξ α L v a l ( ω ξ ω L t r a i n ( ω , α ) , α )
where ξ is learning rate.
Algorithm 1 NAS-HRIS Search Encoder for High-Resolution Remote Sensing Image Segmention
Require: D t r a i n : the training set; D v a l i d : the validation set; n: batch size; initialized operation set P;
Ensure:
  1: initialized the architecture variable α and the weights ω randomly, learning rate ξ , search epochs
  2: repeat
  3:    Sample batch of data D t from D t r a i n ;
  4:    compute L t r a i n ( ω , α ) D t ;
  5:    Updata ω by gradient descent:
             ω = ω ξ ω L t r a i n ( ω , α ) ;
  6:    Sample batch of data D v from D v a l i d ;
  7:    compute L v a l i d ( ω , α ) D v ;
  8:    Updata ω by gradient descent:
             α = α ξ α L v a l i d ( ω ξ ω L t r a i n ( ω , α ) , α ) ;
  9: until converge
Compared with DARTS [32], NAS-HRIS saves P times the GPU memory cost, making the implementation of NAS in large-scale datasets possible. This satisfies the large data characteristics of a high-resolution remote sensing image.

2.3. Evaluation criteria

There are many criteria to evaluate the segmentation effect, most of which are based on accuracy and IoU. And different criteria represent different evaluation meanings. We selected several representative indicators to represent the performance of the segmentation task. In order to easily represent these criteria, we set the number of positive samples correctly predicted as T P , the number of positive samples wrongly predicted as F P , the number of negative samples correctly predicted as T N , and the number of negative samples wrongly predicted as F N .

2.3.1. Pixel Accuracy (PA)

This is one of the simplest metrics, and it represents the percentage of pixels that are properly classified.
P A = T P + T N T P + T N + F P + F N

2.3.2. F 1 Score

F 1 Score is defined as the harmonic mean of the precision and recall.
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 · P r e c i s i o n · R e c a l l P r e c i s i o n + R e c a l l

2.3.3. Mean Intersection over Union (MIoU)

This is the standard metric for segmentation tasks. It represents the mean ratio of intersection to union of two sets.
M I o U = T P T P + T N + F P

3. Experiments and Results

We describe the implementation of NAS-HRIS on three different datasets in detail. All the experiments were done in a single Tesla V100 GPU which has 32G memory. Our experiments consist of three stages. First of all, we use NAS to search the optimal architecture on the specified dataset, according to Algorithm 1. After this step, we can get the certain normal cell and reduction cell. The second stage is to retrain the optimal architecture and obtain a better performance model. In the first two steps, the training set and validation set are used. At last, we use the testing set to assess the performance of the architecture we have searched. We define each cell as consisting of seven node and eight candidate operations, and the depth of the encoder is eight layers. The learning rate is 0.025.

3.1. Experiments on Aerial Dataset

We chose the WHUBuilding dataset [42] for aerial images. The dataset is composed of more than 22,000 independent buildings in Christchurch, New Zealand. These buildings are extracted from aerial images with a spatial resolution of 0.0075 m and a coverage area of 450 km 2 . Most of the images are down-sampled to 0.3 m spatial resolution and cropped into 8189 non-overlapping blocks to form the whole dataset. They are divided into three parts, 4736 images for training, 1036 images for validation, and 2416 images for testing.
The architecture search process was carried out on the WHUBuilding dataset for about 12 hours for 30 epochs, and the resulting normal cell is shown in Figure 4, and the reduction cell in Figure 5. We ran the NAS-HRIS three times and the deviations of the PA, F 1 , and MIoU were 0.12%, 0.38%, and 0.25%, respectively, indicating the MIoU being nearly invariant. We compared NAS-HRIS with SegNet, U-Net and Deeplab v3+. The comparison results are shown in Table 1 and Figure 6. As we can see, the MIoU was higher than 5.93% and the F 1 was higher, 4.81%, than SegNet. Due to the simple design of the search space, our model is very small, only 1/164 times the size of SegNet.
As can be seen in Figure 6, the ability of SegNet to divide independent buildings is strong, and there is little adhesion between buildings, but the integrity of building segmentation is not high. In the aerial HR remote sensing images, U-Net does not perform as well as in the field of medical images. Although the MIoU is higher than SegNet, the independence of segmentation is not strong, and it is difficult to distinguish the areas between buildings. In the three groups of control experiments, Deeplab v3+ is the most prominent; the edge of the building is clearly divided, but there will be regional misclassification in the middle part of the building. As can be seen from the third picture, the distinction between roads and houses is still a difficult point in building segmentation, especially in areas with similar features. Obviously, the best performance is NAS-HRIS, the edge is clear, and the building segmentation is complete.
We used search time and train time to measure our approach NAS-HRIS. Because Segnet, U-net and deeplab are fixed architecture, there is no search time, so we have listed the respective trian time in relevant experiments. It is worth mentioning that because the DARTS method consumes a lot of memory, especially in the case of high-resolution remote sensing images with such a large scale of data, experiments cannot run on 32G GPU, so we do not give the relevant data, which precisely reflects the significance of our method improvement.

3.2. Experiments on Satellite Dataset

Gaofen Image Dataset (GID) is a dataset for land cover classification. It contains 150 HR images captured from more than 60 cities in China [43]. Each original image is 7200 × 6800, and we cut them into 182 images, each with a size of 512 × 512. Due to some problems with image labels, we selected 10,000 images as our dataset. Among them, 6000 images are for training, 2000 images for validation, and 2000 images for testing. There are five classes of tag in GID, which are built-up, farmland, forest, meadow, and waters, as can be seen in Figure 7.
By analogy with WHUBuilding, we used the three architectures of SegNet, U-Net, Deeplab v3+ as a comparison. The MIoU of NAS-HRIS is 7.37 % to 8.84 % higher than the other three methods (see Table 2), which shows the superiority of the customized architecture obtained by architecture search in complex datasets. Because there are many unmarked parts in the source image, in order to show the effect, we deliberately selected four images and compared them in this experiment. As can be seen from Figure 7, in satellite images, the two methods are not satisfactory for the boundary control of segmentation. There are functional disorders in the classification of forest by Segnet and functional disorders in the classification of meadow by NAS-HRIS. Note that in the last image, there are some ships parked on the water; although it is not marked in detail in the label, both methods have reflected that.
Compared to the cells searched by NAS-HRIS on WHUBuilding_Dataset, the cells searched by NAS-HRIS on GID_Dataset in Figure 8 and Figure 9 have a large number of avg_pooling. The reason for our analysis is that GID_Dataset is a satellite image dataset, which has large area, many colors and complex features. Furthermore, avg_pooling retains more background information from a wide range of images.

3.3. Experiments on Non-single Source Dataset

In order to run NAS-HRIS in multiple environments, we have made a non-single source dataset, namely Beijing Building Dataset (BBD). It is worth mentioning that BBD not only meets the requirements of HR image segmentation labels, but also has the value of convenient application. BBD is an elevation satellite image dataset, which is integrated by satellite image and aerial photographs for building extraction and identification. It contains 2000 images from Google Earth History Map of five different areas in Beijing in November 24th, 2016, and all these images are 512 × 512 with a precision of 0.458 m. It covers more than 100 km 2 geographic areas of Beijing both in suburbs and urban areas. We split the dataset into three parts, 1200 images for training, 400 images for validation and 400 images for testing.
In this experiment, we used the architecture searched on the WHUBuilding datasets. On this basis, retrain was carried out. The results of NAS-HRIS compared with SegNet, U-Net and Deeplab v3+ are shown in Table 3 and Figure 10.

4. Discussions and Conclusions

We proposed an improved image segmentation algorithm for high-resolution (HR) remote sensing images based on a neural architecture search (NAS-HRIS). NAS-HRIS uses a gradient descent search strategy to search in a cell-based search space. Compared with traditional methods, NAS-HRIS realizes the automatic design of neural networks and reduces the memory resources used in the automatic search process. We created a new urban Beijing Building Dataset (BBD), which is an elevation satellite image dataset integrated by satellite image and aerial photograph for urban building extraction and identification. We applied NAS-HRIS to aerial images, satellite images, and non-single source images, and achieved 90.44% MIoU on the WHUBuilding dataset. Although NAS-HRIS performs well in the task of segmentation of the HR remote sensing datasets, it still needs to consume considerable computing resources in the process of searching the architecture. Therefore, in the following work, we will further optimize the search space and search strategy and get rid of the constraints of computing resources on the neural architecture search.

Author Contributions

Conceptualization, W.J. and W.W.; Formal analysis, M.W.; Investigation, J.L.; Methodology, N.F. and W.W.; Project administration, W.J.; Software, M.Z., W.J. and N.F.; Writing—review and editing, R.D. All authors have read and agreed to the published version of the manuscript.

Funding

The work described in this paper is supported by National Natural Science Foundation of China (31770768), Fundamental Research Funds for the Central Universities(2572017PZ04), Heilongjiang Province Applied Technology Research and Development Program Major Project(GA18B301,GA20A301) and China State Forestry Administration Forestry Industry Public Welfare Project (201504307).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Gorelick, N.; Hancher, M.; Dixon, M.; Ilyushchenko, S.; Thau, D.; Moore, R. Google Earth Engine: Planetary-scale geospatial analysis for everyone. Remote Sens. Environ. 2017, 202, 18–27. [Google Scholar] [CrossRef]
  2. Wei, W.; Polap, D.; Li, X.; Woźniak, M.; Liu, J. Study on Remote Sensing Image Vegetation Classification Method Based on Decision Tree Classifier. In Proceedings of the 2018 IEEE Symposium Series on Computational Intelligence (SSCI), Bangalore, India, 18–21 November 2018; pp. 2292–2297. [Google Scholar] [CrossRef]
  3. Capizzi, G.; Sciuto, G.; Woźniak, M.; Damaševičius, R. A clustering based system for automated oil spill detection by satellite remote sensing. Lect. Notes Comput. Sci. 2016, 9693, 613–623. [Google Scholar] [CrossRef]
  4. Sun, Z.; Lin, D.; Wei, W.; Wozniak, M.; Damasevicius, R. Road Detection Based on Shearlet for GF-3 Synthetic Aperture Radar Images. IEEE Access 2020, 8, 28133–28141. [Google Scholar] [CrossRef]
  5. Ma, L.; Li, M.; Ma, X.; Cheng, L.; Du, P.; Liu, Y. A review of supervised object-based land-cover image classification. ISPRS J. Photogramm. Remote Sens. 2017, 130, 277–293. [Google Scholar] [CrossRef]
  6. Chen, Y.; Ming, D.; Zhao, L.; Lv, B.; Zhou, K.; Qing, Y. Review on high spatial resolution remote sensing image segmentation evaluation. Photogramm Eng. Remote Sens. 2018, 84, 629–646. [Google Scholar] [CrossRef]
  7. Badmos, O.S.; Rienow, A.; Callo-Concha, D.; Greve, K.; Jürgens, C. Urban development in West Africa-monitoring and intensity analysis of slum growth in Lagos: Linking pattern and process. Remote Sens. 2018, 10, 1044. [Google Scholar] [CrossRef] [Green Version]
  8. Zhou, D.; Xiao, J.; Bonafoni, S.; Berger, C.; Deilami, K.; Zhou, Y.; Frolking, S.; Yao, R.; Qiao, Z.; Sobrino, J.A. Satellite remote sensing of surface urban heat islands: Progress, challenges, and perspectives. Remote Sens. 2019, 11, 48. [Google Scholar] [CrossRef] [Green Version]
  9. Li, W.; Liu, H.; Wang, Y.; Li, Z.; Jia, Y.; Gui, G. Deep Learning-Based Classification Methods for Remote Sensing Images in Urban Built-Up Areas. IEEE Access 2019, 7, 36274–36284. [Google Scholar] [CrossRef]
  10. Yi, Y.; Zhang, Z.; Zhang, W.; Zhang, C.; Li, W.; Zhao, T. Semantic segmentation of urban buildings from VHR remote sensing imagery using a deep convolutional neural network. Remote Sens. 2019, 11, 1774. [Google Scholar] [CrossRef] [Green Version]
  11. Dey, V.; Zhang, Y.; Zhong, M. A Review on Image Segmentation Techniques with Remote Sensing Perspective; ISPRS: Vienna, Austria, 2010; Volume 38, pp. 31–42. [Google Scholar]
  12. Kettig, R.L.; Landgrebe, D. Classification of multispectral image data by extraction and classification of homogeneous objects. IEEE Trans. Geosci. Remote Sens. 1976, 14, 19–26. [Google Scholar] [CrossRef] [Green Version]
  13. Carleer, A.; Debeir, O.; Wolff, E. Assessment of Very High Spatial Resolution Satellite Image Segmentations. Photogramm Eng. Remote Sens. 2010, 71, 1285–1294. [Google Scholar] [CrossRef] [Green Version]
  14. Guo, S.; Jin, Q.; Wang, H.; Wang, X.; Wang, Y.; Xiang, S. Learnable gated convolutional neural network for semantic segmentation in remote-sensing images. Remote Sens. 2019, 11, 1922. [Google Scholar] [CrossRef] [Green Version]
  15. Chen, G.; Li, C.; Wei, W.; Jing, W.; Woźniak, M.; Blažauskas, T.; Damaševičius, R. Fully convolutional neural network with augmented atrous spatial pyramid pool and fully connected fusion path for high resolution remote sensing image segmentation. Appl. Sci. 2019, 9, 1816. [Google Scholar] [CrossRef] [Green Version]
  16. Dong, L.; Xing, L.; Liu, T.; Du, H.; Mao, F.; Han, N.; Li, X.; Zhou, G.; Zhu, D.; Zheng, J.; et al. Very High Resolution Remote Sensing Imagery Classification Using a Fusion of Random Forest and Deep Learning Technique-Subtropical Area for Example. IEEE J. Sel. Top Appl. Earth Obs. Remote Sens. 2020, 13, 113–128. [Google Scholar] [CrossRef]
  17. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  18. Fu, G.; Liu, C.; Zhou, R.; Sun, T.; Zhang, Q. Classification for high resolution remote sensing imagery using a fully convolutional network. Remote Sens. 2017, 9, 498. [Google Scholar] [CrossRef] [Green Version]
  19. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical image computing and computer-assisted intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  20. Zhang, Z.; Liu, Q.; Wang, Y. Road extraction by deep residual u-net. IEEE Geosci. Remote. Sens. Lett. 2018, 15, 749–753. [Google Scholar] [CrossRef] [Green Version]
  21. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 40, 834–848. [Google Scholar] [CrossRef]
  22. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  23. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  24. Li, J.; Liang, B.; Wang, Y. A hybrid neural network for hyperspectral image classification. Remote Sens. Lett. 2020, 11, 96–105. [Google Scholar] [CrossRef]
  25. Liu, Q.; Basu, S.; Ganguly, S.; Mukhopadhyay, S.; DiBiano, R.; Karki, M.; Nemani, R. DeepSat V2: Feature augmented convolutional neural nets for satellite image classification. Remote Sens. Lett. 2020, 11, 156–165. [Google Scholar] [CrossRef] [Green Version]
  26. Pashaei, M.; Kamangir, H.; Starek, M.J.; Tissot, P. Review and evaluation of deep learning architectures for efficient land cover mapping with UAS hyper-spatial imagery: A case study over a wetland. Remote Sens. 2020, 12, 959. [Google Scholar] [CrossRef] [Green Version]
  27. Xiong, W.; Xiong, Z.; Cui, Y.; Lv, Y. Deep multi-feature fusion network for remote sensing images. Remote Sens. Lett. 2020, 11, 563–571. [Google Scholar] [CrossRef]
  28. Gabryel, M.; Damaševičius, R. The image classification with different types of image features. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2017; Volume 10245, pp. 497–506. [Google Scholar]
  29. Zhou, B.; Duan, X.; Ye, D.; Wei, W.; Woźniak, M.; Połap, D.; Damaševičius, R. Multi-level features extraction for discontinuous target tracking in remote sensing image monitoring. Sensors 2019, 19, 4855. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Baker, B.; Gupta, O.; Naik, N.; Raskar, R. Designing neural network architectures using reinforcement learning. arXiv 2016, arXiv:1611.02167. [Google Scholar]
  31. Zoph, B.; Le, Q.V. Neural Architecture Search with Reinforcement Learning. arXiv 2016, arXiv:1611.01578. [Google Scholar]
  32. Liu, H.; Simonyan, K.; Yang, Y. Darts: Differentiable architecture search. arXiv 2018, arXiv:1806.09055. [Google Scholar]
  33. Huang, S.Y.; Chu, W.T. PONAS: Progressive One-shot Neural Architecture Search for Very Efficient Deployment. arXiv 2020, arXiv:2003.05112. [Google Scholar]
  34. Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning transferable architectures for scalable image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8697–8710. [Google Scholar]
  35. Liu, C.; Zoph, B.; Neumann, M.; Shlens, J.; Hua, W.; Li, L.J.; Fei-Fei, L.; Yuille, A.; Huang, J.; Murphy, K. Progressive neural architecture search. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 19–34. [Google Scholar]
  36. Real, E.; Aggarwal, A.; Huang, Y.; Le, Q.V. Regularized evolution for image classifier architecture search. In Proceedings of the aaai conference on artificial intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 4780–4789. [Google Scholar]
  37. Real, E.; Moore, S.; Selle, A.; Saxena, S.; Suematsu, Y.L.; Tan, J.; Le, Q.V.; Kurakin, A. Large-scale evolution of image classifiers. In Proceedings of the 34th International Conference on Machine Learning—JMLR.org, Sydney, Australia, 6–11 August 2017; Volume 70, pp. 2902–2911. [Google Scholar]
  38. Wei, W.; Zhou, B.; Maskeliūnas, R.; Damaševičius, R.; Połap, D.; Woźniak, M. Iterative Design and Implementation of Rapid Gradient Descent Method. Lect. Notes Comput. Sci. 2019, 11508, 530–539. [Google Scholar] [CrossRef]
  39. Jang, E.; Gu, S.; Poole, B. Categorical reparameterization with gumbel-softmax. arXiv 2016, arXiv:1611.01144. [Google Scholar]
  40. Maddison, C.J.; Mnih, A.; Teh, Y.W. The concrete distribution: A continuous relaxation of discrete random variables. arXiv 2016, arXiv:1611.00712. [Google Scholar]
  41. Dong, X.; Yang, Y. Searching for a robust neural architecture in four gpu hours. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 1761–1770. [Google Scholar]
  42. Ji, S.; Wei, S.; Lu, M. Fully convolutional networks for multisource building extraction from an open aerial and satellite imagery data set. IEEE Trans. Geosci. Remote Sens. 2018, 57, 574–586. [Google Scholar] [CrossRef]
  43. Tong, X.Y.; Xia, G.S.; Lu, Q.; Shen, H.; Li, S.; You, S.; Zhang, L. Learning transferable deep models for land-use classification with high-resolution remote sensing images. arXiv 2018, arXiv:1807.05713. [Google Scholar]
Figure 1. Example image from the Beijing building dataset (BBD), (a) is original data, and (b) is the associated label.
Figure 1. Example image from the Beijing building dataset (BBD), (a) is original data, and (b) is the associated label.
Sensors 20 05292 g001
Figure 2. The model uses the encoder-decoder structure. And the encoder is composed of cells which are searched by the NAS.
Figure 2. The model uses the encoder-decoder structure. And the encoder is composed of cells which are searched by the NAS.
Sensors 20 05292 g002
Figure 3. NAS-HRIS used the gradient descent search strategy to search the architecture of encoder in the cell-based search space and optimized the parameters by continuously evaluating the performance of the architecture. (a) operation of each edge in DAG is unknown; (b) candidate operations on each edge to continuous relaxation of the search space are set; (c,d) each edge is finalized by applying the reparameterization trick to sampling.
Figure 3. NAS-HRIS used the gradient descent search strategy to search the architecture of encoder in the cell-based search space and optimized the parameters by continuously evaluating the performance of the architecture. (a) operation of each edge in DAG is unknown; (b) candidate operations on each edge to continuous relaxation of the search space are set; (c,d) each edge is finalized by applying the reparameterization trick to sampling.
Sensors 20 05292 g003
Figure 4. The normal cell searched by NAS-HRIS on WHUBuilding_Dataset.
Figure 4. The normal cell searched by NAS-HRIS on WHUBuilding_Dataset.
Sensors 20 05292 g004
Figure 5. The reduction cell searched by NAS-HRIS on WHUBuilding_Dataset.
Figure 5. The reduction cell searched by NAS-HRIS on WHUBuilding_Dataset.
Sensors 20 05292 g005
Figure 6. Examples of segmentation results with the SegNet, U-Net, Deeplab v3+ and NAS-HRIS, respectively, on the aerial dataset. (a) Image. (b) Label. (c) SegNet. (d) U-Net. (e) Deeplab v3+. (f) NAS-HRIS.
Figure 6. Examples of segmentation results with the SegNet, U-Net, Deeplab v3+ and NAS-HRIS, respectively, on the aerial dataset. (a) Image. (b) Label. (c) SegNet. (d) U-Net. (e) Deeplab v3+. (f) NAS-HRIS.
Sensors 20 05292 g006
Figure 7. Representative cases of image segmentation results with SegNet and NAS-HRIS, respectively. (a) Image. (b) Label. (c) SegNet. (d) NAS-HRIS.
Figure 7. Representative cases of image segmentation results with SegNet and NAS-HRIS, respectively. (a) Image. (b) Label. (c) SegNet. (d) NAS-HRIS.
Sensors 20 05292 g007
Figure 8. The normal cell searched by NAS-HRIS on GID_Dataset.
Figure 8. The normal cell searched by NAS-HRIS on GID_Dataset.
Sensors 20 05292 g008
Figure 9. The reduction cell searched by NAS-HRIS on GID_Dataset.
Figure 9. The reduction cell searched by NAS-HRIS on GID_Dataset.
Sensors 20 05292 g009
Figure 10. Examples of segmentation results with the SegNet and NAS-HRIS, respectively, on the non-single source dataset. (a) Image. (b) Label. (c) SegNet. (d) NAS-HRIS.
Figure 10. Examples of segmentation results with the SegNet and NAS-HRIS, respectively, on the non-single source dataset. (a) Image. (b) Label. (c) SegNet. (d) NAS-HRIS.
Sensors 20 05292 g010
Table 1. Testing result on WHUBuilding_Dataset
Table 1. Testing result on WHUBuilding_Dataset
ArchitecturesParameters (M)PA (%) F 1 (%)MIoU (%)Search Time (h)Train Time (h)
SegNet29.444197.7788.9684.51-7.4
U-Net23.356598.3093.5688.41-6.2
Deeplab v3+13.395398.0994.4790.20-4.0
NAS-HRIS0.186898.5293.7790.4412.116.4
Table 2. Testing result on GID_Dataset
Table 2. Testing result on GID_Dataset
ArchitecturesParameters (M)PA (%) F 1 (%)MIoU (%)Search Time (h)Train Time (h)
SegNet29.444179.9671.5063.19-18.3 h
U-Net23.356580.3773.7164.66-13.2 h
Deeplab v3+13.395382.4271.8363.82-14.9 h
NAS-HRIS0.123288.4878.3567.0310.619.5 h
Table 3. Testing result on BBD_Dataset
Table 3. Testing result on BBD_Dataset
ArchitecturesParameters (M)PA (%) F 1 (%)MIoU (%)Search Time (h)Train Time (h)
SegNet29.444195.4882.1174.12-5.4
U-Net23.356595.2183.5674.66-2.8
Deeplab v3+13.395394.4284.4375.19-3.3
NAS-HRIS0.204896.2885.3175.2112.15.8

Share and Cite

MDPI and ACS Style

Zhang, M.; Jing, W.; Lin, J.; Fang, N.; Wei, W.; Woźniak, M.; Damaševičius, R. NAS-HRIS: Automatic Design and Architecture Search of Neural Network for Semantic Segmentation in Remote Sensing Images. Sensors 2020, 20, 5292. https://doi.org/10.3390/s20185292

AMA Style

Zhang M, Jing W, Lin J, Fang N, Wei W, Woźniak M, Damaševičius R. NAS-HRIS: Automatic Design and Architecture Search of Neural Network for Semantic Segmentation in Remote Sensing Images. Sensors. 2020; 20(18):5292. https://doi.org/10.3390/s20185292

Chicago/Turabian Style

Zhang, Mingwei, Weipeng Jing, Jingbo Lin, Nengzhen Fang, Wei Wei, Marcin Woźniak, and Robertas Damaševičius. 2020. "NAS-HRIS: Automatic Design and Architecture Search of Neural Network for Semantic Segmentation in Remote Sensing Images" Sensors 20, no. 18: 5292. https://doi.org/10.3390/s20185292

APA Style

Zhang, M., Jing, W., Lin, J., Fang, N., Wei, W., Woźniak, M., & Damaševičius, R. (2020). NAS-HRIS: Automatic Design and Architecture Search of Neural Network for Semantic Segmentation in Remote Sensing Images. Sensors, 20(18), 5292. https://doi.org/10.3390/s20185292

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop