Next Article in Journal
Slow Breathing Exercise with Multimodal Virtual Reality: A Feasibility Study
Next Article in Special Issue
A Blanket Accommodative Sleep Posture Classification System Using an Infrared Depth Camera: A Deep Learning Approach with Synthetic Augmentation of Blanket Conditions
Previous Article in Journal
Time Switching Based Wireless Powered Relay Transmission with Uplink NOMA
Previous Article in Special Issue
GMSRI: A Texture-Based Martian Surface Rock Image Dataset
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fast and Accurate Object Detection in Remote Sensing Images Based on Lightweight Deep Neural Network

School of Computer and Information Technology, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(16), 5460; https://doi.org/10.3390/s21165460
Submission received: 11 June 2021 / Revised: 31 July 2021 / Accepted: 9 August 2021 / Published: 13 August 2021
(This article belongs to the Special Issue Deep Learning Image Recognition Systems)

Abstract

:
Deep learning-based object detection in remote sensing images is an important yet challenging task due to a series of difficulties, such as complex geometry scene, dense target quantity, and large variant in object distributions and scales. Moreover, algorithm designers also have to make a trade-off between model’s complexity and accuracy to meet the real-world deployment requirements. To deal with these challenges, we proposed a lightweight YOLO-like object detector with the ability to detect objects in remote sensing images with high speed and high accuracy. The detector is constructed with efficient channel attention layers to improve the channel information sensitivity. Differential evolution was also developed to automatically find the optimal anchor configurations to address issue of large variant in object scales. Comprehensive experiment results show that the proposed network outperforms state-of-the-art lightweight models by 5.13% and 3.58% in accuracy on the RSOD and DIOR dataset, respectively. The deployed model on an NVIDIA Jetson Xavier NX embedded board can achieve a detection speed of 58 FPS with less than 10W power consumption, which makes the proposed detector very suitable for low-cost low-power remote sensing application scenarios.

1. Introduction

With the rapid development of satellite and imaging technology, optical remote sensing images with high spatial resolution are obtained more conveniently than ever before [1]. Studies on analyzing and understanding remote sensing images have drawn wide attention. Image classification, segmentation, object detection, and tracking have become the hot topics in the field of remote sensing [2,3,4]. Among them, object detection has presented a broader application prospect in real-world applications and is sought after by researchers in recent years [5].
For object detection tasks, deep neural network-based schemes have shown superior performance over traditional approaches [6,7]. In general, these schemes can be divided into two major categories: (1) one-stage neural network which adopts a fully convolutional architecture that outputs a fixed number of predictions on the grid, such as SSD [8], YOLO [9], and M2Det [10], and (2) two-stage network that leverages a proposal network to find regions of interest that have a high probability to contain an object and a second network to get the classification score and spatial offsets, such as FPN [11] and Faster R-CNN [12]. These detectors have been successfully utilized in many applications, such as robotics, autonomous vehicles, and surveillance systems.
However, a direct utilization of generic detectors in remote sensing images usually does not deliver satisfactory results. The major reason is that there are many distinct features different from natural images in remote sensing images, such as very complex geometric background, dense object distributions, and variety of objects with large variant in shapes and scales [13,14]. To address these new design challenges faced in object detection in remote sensing images, many studies have been proposed in the literature [15,16,17,18,19,20]. For instance, Wang et al. [15] designed a new pyramid structure to optimize the Faster-RCNN detector by adding a feature-reflowing pathway from the lower level for each scale to enrich the feature expression. Huang et al. [16] proposed a cross-scale fusion module based on the M2Det detector to extract sufficient comprehensive semantic information from features for performing multi-scale fusion. Zhao et al. [17] improved the SSD detector by adding a channel attention module to strengthen the long-term semantic dependence between objects to improve the discriminative ability of the deep features.
In this paper, we focus on the issue of the large variant in object distributions and scales in remote sensing images. Traditional anchor-based detectors [21,22] match regions of possible objects by a set of pre-allocated anchors with pre-defined aspect ratios; therefore, the final accuracy of the trained neural network model highly relies on the anchor configuration. However, in remote sensing images, the scale and distribution of the target objects vary in a very wide range. For instance, Figure 1 compares the sizes of the four classes of objects in the RSOD [23] dataset. Due to the fact that remote sensing images are often acquired by sensors with the same hardware settings, and because the shooting angle is also fixed [24], the sizes of the target object in remote sensing images are directly related to the real-world object scale. As a result, there is a large gap in the size of the ground truth bounding boxes among all objects within one training dataset. In Figure 1, the aircraft and oiltanks have much smaller scales but larger quantities in the image when compared to the overpass and playground. The current anchor selection scheme, such as K-means clustering [25,26,27], tends to allocate more anchors for objects with larger quantities. This makes generic detector perform very badly on objects that does not have sufficient corresponding samples during anchor configuration. To demonstrate this phenomenon, we compare some representative detection results of using the YOLOv4-Tiny network on the RSOD dataset in Figure 1. For the object classes of overpass and playground, the detection results are very inaccurate.
Beside the above problems, this study also deals with the efficiency of the deep neural network which consists most parts of the detector. In real-world applications, remote sensing object detection tasks are commonly used in rescue, military, and other scenarios [28,29]. This requires the detectors to be as lightweight as possible and be able to efficiently deploy on low-cost low-power embedded devices. Regardless of the type of detection framework used, optimizations of the detailed algorithm, such as the backbone neural network, multi-scale feature fusion, and adaptive anchor setting, are also important in trading off between the detection accuracy and speed to meet the requirements of the target application.
The contributions of this paper are the following. (1) We propose a lightweight backbone deep neural network design, which can achieve the optimal balance between model size and detection accuracy for fast processing on low-cost low-power embedded hardware platforms. (2) We propose an automatic anchor configuration scheme based on differential evolution (DE), which can minimize the average distance between ground truth bounding boxes and selected anchors, and improve the accuracy of object matching. (3) Comprehensive experiments on multiple datasets are conducted, and the results show that the proposed lightweight detector outperforms state-of-the-art detection approaches by 5.13% and 3.58% on the RSOD and DIOR datasets, respectively. We have also deployed the proposed detector on a embedded hardware platform, i.e., an NVIDIA Jetson Xavier NX board [30], achieving a real-time detection speed of 58 FPS with less than 10 W power consumption.
The reminder of this paper is organized as follows. The related work on existing studies of deep neural network design for remote sensing object detection tasks is summarized in Section 2. The proposed neural network design and automatic anchor configuration scheme is introduced in Section 3. The experimental setup is shown in Section 4. Section 5 describes the experimental results in detail, whereas Section 6 shows the performance of the deployment on embedded platform. Finally, Section 7 concludes this paper.

2. Related Work

The identification of objects in remote sensing images is a subset of a wider research field in object detection. Taking into account the needs of neural network deployment, the existing studies can be categorized into three main groups: lightweight object detection network, visual attention mechanism, and optimized anchor configuration.

2.1. Lightweight Object Detection Network

State-of-the-art approaches can be divided into region-based and single-shot detectors. The former ones, such as R-CNN [31], Fast R-CNN [32], and Faster R-CNN, use a set of Regions-of-Interest (ROIs) to extract sub-promotions that may contain objects, then fine-grained detection and classification modules analyze each ROI. Although region-based schemes have high detection accuracy, the complex network structures often result in high computation workload and low processing speed. Single-shot detectors, such as SSD, YOLO, and FCOS [33], tackle the input image in a single pipeline. As a benefit of the end-to-end structure, these detectors can achieve real-time detection speed, which is very suitable as a lightweight object detection network structure. Designers can further balance the accuracy and model size according to the requirements of their own target application.
As the most typical end-to-end network structure, YOLO [9,27,34,35] is favored by many application scenarios [36,37,38]. YOLO-Tiny is the lightweight version of the YOLO detector. YOLO detector is generally composed of three basic parts: backbone, neck, and head [35]. The backbone neural network, as an important part of the network to extract deep image information, generally performs as the image classification network, such as Darknet53, ResNet [39], Vgg [40], and Mobilenet [41]. The key information of object is extracted in the neck, and used to predict the position and category in the head. Although YOLO-Tiny meets the actual deployment requirements, it could not meet the accuracy requirements on remote sensing images due to the following drawbacks:
  • The shallow backbone and prediction network structures are not sufficient to extract deep semantic information, which limits the performance of the network in complex scenarios, for instance, very small objects or complicated backgrounds [42]. In addition, the simple organization of prediction layers cannot effectively cover objects of various proportions, especially when remote sensing images with dense object distributions were considered.
  • The performance of the detector is especially sensitive to anchor configurations, which not only affects the speed of training, but also the robustness of the network. As been explained in previous section, the small amount of anchors with fixed scales in YOLO-Tiny will deliver poor detection results due to the large variation in object scales.

2.2. Visual Attention Mechanism

Aiming at highlighting the salient feature of images, the attention module has been widely used in various types of detectors. It has became the mainstream approach of improving network accuracy. For instance, Wang et al. [43] proposed the residual attention network, which improved the expressive ability of the network. A channel-level attention mechanism through feature recalibrating was proposed by SENet [44], which could improve neural network’s performance in classification, detection, and other tasks to a new extent. CBAM [45], which combined channel attention and spatial attention, was also proposed to further improve the model’s accuracy. The CA module [46] that integrated the channel and the spatial direction not only reduced the amount of parameters, but also had stronger capabilities. In general, a typical attention module can be divided into two parts: channel attention module and spatial attention module, which have their own advantages and disadvantages for different applications, and the specific effects were often evaluated based on experiments on the target dataset.
Although the attention module can deliver increased detection accuracy, the improvements often come from objects of large scales, which means that, due to the special calculation utilized, some unobvious features are ignored, and the attention model usually focuses on feature-rich large objects, while small objects [46] are often ignored. For remote sensing images with a large number of small objects, the effectiveness of attention modules still needs to be further investigated.

2.3. Optimized Anchors

Anchor is used by the YOLO framework in the detection pipeline, which is extremely important in remote sensing image-based object detection tasks. In previous studies, there are two commonly used methods to allocate the anchors for the detection network:
  • Manual configuration. The anchors selected by the manual correction method are more straightforward and robust. However, it requires the designer to have rich experience in the application field and perform comprehensive manual experiments before determining the best setting.
  • Automatic configuration based on optimizations. According to distribution of the data set, this type of scheme can automatically find the best anchor position, which greatly relieves the effort of searching for the optimal configurations and also delivers higher accuracy and faster training speed.
K-means clustering is widely used as the optimization scheme for automatic anchor selection in YOLO. Given a specific number of anchors, the best anchor positions are selected by clustering the bounding boxes in the datasets. The IOU score is further used as the relative distance for anchor configurations in YOLOv4 [35]. Junos et al. [47] used a similar scheme to optimize the anchor in YOLOv3 and applied the network to the crop harvesting system. Zlocha et al. [48] optimized the anchor configuration based on a differential evolution search algorithm in RetinaNet. However, in previous studies, all objects were unified into one big category, which was not reasonable for datasets with unevenly distributed object scales. As we have pointed out in the introduction section, traditional scheme tends to shift the anchor selection toward the category with the maximum number of ground truth samples, resulting in poor detection performance for other objects.

3. Methodology

In this paper, we aim to design a lightweight deep neural network to accurately detect objects of large variant in scale and quantity in remote sensing images. The proposed detector framework is illustrated in Figure 2. In the proposed detector, we have designed a YOLOv4-like backbone network followed by three prediction layers to capture and combine rich contextual features, while, at the same time, minimizing the network’s computational cost. Moreover, efficient channel attention is also developed to form a downsampling module, namely, the Cross Stage Partial connections with Attention (CSPA), to efficiently extract high-level feature information for object classification and localization. Finally, an automatic optimal anchor selection approach based on differential evolution (DE) is proposed to address the problem of biased anchor allocation due to large variant of object scale and quantity in remote sensing images.

3.1. Lightweight Neural Network

In this section, the details of the proposed neural network architecture are presented. Our network design emphasizes the lightweight feature and computational efficiency at the minimum cost of detection accuracy. The overall neural network architecture is illustrated in Figure 2.
The proposed network is organized in two major components: a downsampling backbone and a prediction pipeline. The input images are first resized to an appropriate size (for instance, 416 × 416 pixels), and then the downsampling backbone network, which has a YOLOv4-like architecture consisting of two convolutional layers and three CSPA modules, extract the salient features form the input image at different layers. To enhance the network’s ability of capturing important features within one channel and fusing information among different input channels, we propose to add an efficient channel attention layer after the third convolution layer in each of the CSPA module.
In the proposed network, a prediction pipeline consisting of feature pyramids and three-scale prediction layers is designed to transform features from multiple branches into prediction results. The feature pyramid can effectively fuse meaningful semantic information obtained from salient feature maps of low resolution and finer-grained information extracted from the earlier branches. The three-scale prediction networks are used to predict encoding parameters of bounding box and class predictions. The final inference results are obtained after filtering the prediction box via non-maximal suppression (NMS).
In the original YOLOv4-Tiny structure, only two layers of networks were utilized in the prediction part to alleviate the computational complexity of the excessively large network in YOLOv4. However, this shallow network structure is extremely unfriendly to remote sensing object detection task. For instance, given an input image of the resolution of 416 × 416 pixels, the dimensions of the output feature maps of the second prediction layer in the original YOLOv4-Tiny structure are 26 × 26. According to the principle of the YOLO algorithm, the input image is divided into many grid cells of equal dimension, each of which will detect objects that appear within it. In the case of using two prediction layers, the image will be divided into 26 × 26 grid cells, which means that each cell is of the size 16 × 16. Generally, when the scales of the target object are too small relative to the size of the grid cells, the object could not be precisely located by the YOLO algorithm. In remote sensing images, there exists a large amount of small objects which cannot be precisely located by grid cells larger than 16 × 16. Therefore, in this work, we propose to add one extra prediction layer with an output feature map dimensions of 56 × 56, which corresponds to grid cells of 8 × 8 dimension on the input image. From Figure 3, it can be seen that, in the proposed network, the corresponding grid cells mapped on the input images well fit the scales of the small objects. Moreover, based on quantitative analysis (please refer to the detailed result on RSOD in Section 5.1), we have found that the proposed three-layer prediction network was optimal and there was no need to introduce extra layers at the expense of larger model size for very little improvement in detection accuracy.

3.2. Efficient Channel Attention

Instead of utilizing both channel and spatial level attentions, we propose to use a single channel attention module to enhance the visual attention of the backbone network. It is based on two design considerations: First, channel attention modules normally have a relatively small amount of parameters compared to spatial ones, which can facilitate rapid training convergence and deliver a faster prediction speed during deployment. Second, we have conducted experiments showing that using channel attention modules was sufficient to improve the detection accuracy to the desired level for remote sensing images. In this work, an efficient channel attention module was designed and utilized in the backbone network by following the basic network structure presented in [49] as shown by Figure 4.
The detailed architecture of the Efficient Channel Attention (ECA) module is summarized as follows: Given the input feature map I, I R C × W × H . In the first step, an average pooling operation is designed to compress I along the spatial dimension W × H to obtain the channel descriptor ( C × 1 × 1 ). In the following layer, it is converted into another channel descriptor ( R 1 × C ) through transpose and squeeze operations. Then, the local cross-channel interaction is completed through fast 1 D convolution of the kernel size k s i z e , where k s i z e also represents the coverage ratio of interaction. In the following parts, transpose and squeeze operation are utilized again to convert channel descriptor back to the original channel dimensions ( R C × 1 × 1 ). Finally, the channel weight is generated through the Sigmoid function, and multiplied by I.
To compare the performance of the classic channel attention module (Squeeze-and-Excitation (SE) [44]) with the proposed ECA module in terms of feature extraction capability and operation efficiency, we have conducted experiment to quantitatively measure the increment in model size and improvement in detection accuracy for both schemes when adopted in the proposed backbone network. The results are reported in Table 1. It can be observed that the SE module outperforms the baseline model by 3.10%, but with an increment of 2688 parameters, while ECA module outperforms SE module by 0.22% but contributes to 74% less parameters. It can be concluded that the ECA module can capture sufficient cross-channel interaction in an efficient way to improve the detection performance with minimal cost in model size and computational complexity.

3.3. Optimal Anchor Configuration Based on Differential Evolution

The anchor configuration, a hyperparameter for the training of the network model, affects the performance of the model in a great degree. Optimal anchor configuration can improve the network accuracy without additional consumption. However, when facing the large variation of object scales in remote sensing images, the anchor configuration scheme used in YOLO, i.e., K-means clustering, will lead to biased anchor allocating setting. The improvement in accuracy tends to be concentrated in the categories with a larger number of objects. One major reason is that the key evaluation metric widely used by previous studies is the mAP score, which is embodied as an average AP score of all categories. Excessively increasing the AP score of a certain category can increase the overall accuracy of the model to a certain extent. However, further improvement is not possible due to the limited detection ability for the other categories with very low AP scores, which means that the trained model is biased. In general, a more reasonable approach is to consider both the distribution of the quantity and size of the objects for all categories, and develop a scheme that can improve the overall performance of the neural network and balance the accuracy of all target categories.
To better capture the relationship between object scale and quantity, we propose an improved anchor configuration scheme based on differential evolution (DE). This method takes the height and width of the anchors as variables, and the sum of the nearest distances from the ground truth bounding boxes to the anchors as a fitness function. In addition, a weight value is also added to the distance calculation to avoid possible biased training of the neural network. Finally, the minimum value of the fitness function is solved by using DE [50], reaching the goal of minimizing the distance.
More specifically, for a given dataset, the distance from one ground truth bounding box to one anchor can be formulated as follows:
d i s ( t r u t h , a n c h o r ) = 1 I O U ( t r u t h , a n c h o r )
where I O U ( t r u t h , a n c h o r ) represents the intersection over union of the ground truth and the anchor box, centered at the origin. The corresponding calculation formula is
I O U = S o v e r l a p S u n i o n
where S o v e r l a p refers to the overlap area between the ground truth and the anchor box. S u n i o n refers to the union area between them.
By denoting x i j as the j-th ground truth of the i-th category, and θ k as k-th anchor box, the distance between x i j and the anchor boxes can be expressed in the following form:
g ( x i j , θ ) = m i n ( d i s ( x i j , θ 1 ) , d i s ( x i j , θ 2 ) , , d i s ( x i j , θ k ) )
which means that we choose the anchor with the smallest distance as its best match. Therefore, the distance between all samples (X) and anchors ( θ ) can be calculated by
G = 1 n × i = 1 n [ 1 m i × j = 1 m i g ( x i j , θ ) ]
where m 1 , m 2 , …, m n represent the number of ground truth samples in the n-th category.
In Equation (4), the values of the weights for different categories in the fitness function are set to be inversely proportional to the number of objects, which helps to eliminate the attraction effect of large number of objects for more anchors.
Here, we use single-objective DE [51] to solve this optimization problem of Equation (4). In Equation (4), the real values in x i j and θ k are all scaled to the range of [0,1]. To prevent conflicting with the aforementioned variables, the decision variables in DE are replaced by P i , i.e., P i = θ ( k , 2 ) used as the fitness function of DE. Then, DE searches for the best anchors of the smallest function G through constant iterations. The pseudocode of the proposed optimal anchor selection scheme is described by Algorithm 1, where C r denotes crossover rate, F s represents the scaling factor, N p is the population size, and t is the iteration number.
Algorithm 1 Anchor configurations algorithm based on DE.
Input: input parameters C r , F s , N p
Output: output a r g m i n P i t G ( P i t ) and P i t
1:
Initialize population P = ( P 1 t , P 2 t , , P N p t )
2:
Counter t 0
3:
while stop condition not met do
4:
    for  i ( 1 , 2 , , N p )  do
5:
         ν i differential mutation ( F s ; i , P )
6:
         μ i crossover ( C r ; P i t , ν i )
7:
        if  G ( μ i ) G ( P i t )  then
8:
            P i t + 1 μ i
9:
        else
10:
            P i t + 1 P i t
11:
        end if
12:
    end for t t + 1
13:
end while
14:
return a r g m i n P i t G ( P i t ) and P i t
In the first step, the proposed algorithm performs an initialization operation on the population P consisting of decision variables P i t , where P i t represents the i-th individual in the t-th iteration. The initialized population is distributed in a certain defined area according to the population size N p , and each individual represents a candidate solution. Generally, the initial population should cover the whole search space. As the population size N p increases, the probability of obtaining the global optimal solution also increases. In this work, each individual is defined as a specific anchor configuration.
In the following iteration, the procedures of differential mutation, crossover, and selection are repeated performed. The goal of differential mutation is to create groups of new individuals which have a certain level of probability of being the optimal solution. The difference introduced by mutant between the parent and children is quantitatively controlled by a scaling factor F s . Later, elements θ j , ( j = 1 , 2 , , k ) in individuals are randomly swapped by the crossover operation for the current iteration and its differential mutated group. This procedure promotes population diversity, and the crossover probability is controlled by C r . Then, newly generated and contemporary individuals are compared and better individuals are selected for transmission to the next iteration. The iteration continues until the best individual and objective function value are found. Through the above algorithm, the minimum value of fitness function can be obtained, which corresponds to the best anchor value.

4. Experimental Settings

4.1. Hardware Platforms

The proposed lightweight model was expected to run on low-power embedded devices. However, to demonstrate the advantage of the proposed optimization schemes and the performance of the proposed detector more comprehensively, experiments were conducted on two different hardware platforms, including an NVIDIA GeForce RTX2080Ti desktop GPU and an NVIDIA Jeteson Xavier embedded board. The hardware specifications of the two platforms are summarized in Table 2. To deploy on the embedded board, all tested neural network model were quantized into FP16 data format to relieve the pressure on external memory bandwidth.

4.2. Datasets and Training Parameters

In this work, experiments on two public remote sensing datasets were conducted to verify the effectiveness of the proposed detector and further evaluate its accuracy and speed. The RSOD remote sensing dataset was selected to measure the key performance metrics of the detector. RSOD includes 4993 aircraft in 446 images, 1586 oiltanks in 165 images, 191 playgrounds in 189 images, and 180 overpasses in 176 images. The dataset was randomly divided into the training and test set according to a 7.5:2.5 ratio.
Considering the relatively limited number of samples in the RSOD, in order to further verify the performance of our method, we have also chosen DIOR to evaluate and test the performance of the proposed model. DIOR [52] contains a wider range of 20 object categories, a total number of 23,463 images, in which 192,472 examples were labeled, and 11,725 images were used for training and 11,738 images were used for testing.
The training parameters of the final network model was setting as follows: The initial weight was pre-trained on the COCO dataset; Adam was utilized as the optimizer, while the initial learning rate was 1 × 10 4 and the maximum training epoch was set to 100.

4.3. Evaluation Metrics

We used a total number of four metrics to evaluate the performance of the proposed method: (1) mAP; (2) FLOPs (floating-point operations); (3) number of parameters; (4) FPS (frames per second). The mean average precision (mAP), which is widely used by previous studies, is still used as the key accuracy score in this paper to compare with the state-of-the-art. FLOPs and parameters are used to evaluate the computational complexity and memory footprint of the neural network model, respectively. FPS is used to evaluate the processing speed of the model on target hardware devices.

5. Results

5.1. Improvements by Network Structure

In this paper, we have proposed two optimizations on the lightweight neural network: multi-scale prediction layers and attention modules. The performance gains of these two schemes were evaluated independently and the results are shown in the following sections.
As discussed in Section 3.1, a deeper prediction network corresponds to a finer grid on the input images, which can locate small objects more accurately. However, detection speed will often be sacrificed due to increased model size and computational complexity. To balance between processing speed and accuracy, we have conducted a series of experiments to explore the optimal structure of the prediction network. The obtained results are compared in Table 3. Compared to only using two prediction layers, adopting a three layer prediction network can improve the detection accuracy in terms of mAP score by 2%, while the increment in FLOPs is ~48%. In particular, the detection accuracy of overpass and aircraft, which often have smaller scales in the images, are improved by 2.27% and 7.74%, respectively. Although adding a fourth layer to the prediction network can further raise the average accuracy by around 1% (the improvement is mainly on the small object of aircraft), the computational complexity doubled compared to using three layers and is 3 × than that of using only two layers. In addition, the extra prediction layer also creates a large amount of unnecessary bounding boxes, which will also reduce the executing speed of the post-processing procedures, such as NMS. Therefore, it is concluded, based on the comparative experiment results, that the three layer prediction network architecture was the most cost-effective design.
For visual attention enhancement, the proposed ECA design was compared with several state-of-the-art attention modules, including SE, CBAM, and CA, as listed in Table 4. The parameter count only includes the parameters introduced by the attention module, while the other metrics correspond to the whole neural network. The proposed ECA module delivers a 3.32% boost in mAP with almost no loss in processing speed. Although the CBAM and CA modules have both channel and spatial attention, they failed to improve the detection results. The main reason behind this was that these two attention modules that were added at a position close to the input have introduced a larger amount of parameters (4 × t h a n t h a t o f t h e p r o p o s e d s c h e m e ) to the original model, causing the training process very sensitive to the initialization state of the backbone model and the original loss function failed to generate sufficient backpropagation information to update the new parameters. The SE module has a very close inference performance with the proposed ECA module, but the processing speed is 2% slower on desktop GPU. This cost in speed will be further enlarged when deployed on embedded platforms.
The proposed ECA module also has a hyperparameter k s i z e , i.e., the filter size of the 1D convolution. Table 5 shows the experimental results with different values of k s i z e , including setting fixed value and adaptive ones in all convolution layers. The ECA module achieves the best performance when k s i z e = 3 and k s i z e = 7 . Furthermore, note that the adaptive approach does not outperform fixed ones. We conjecture the main reason is that one layer of convolution with small filter size is sufficient to capture enough spatial feature information within one channel, and larger perception filter and more layers are redundant [49].

5.2. Improvements by Anchor Configuration

Population size ( N p ) and maximum iteration number are two key initialization parameters for the DE solver utilized in the proposed anchor configuration algorithm. To determine the most reasonable initialization parameters, the following experiments were carried out.
Firstly, we set P i = θ ( 6 , 2 ) for the DE optimizer, in which only six anchors were configured. Therefore, other parameters were set as follows: C r = 0.7 , F s = 0.5 , maximum iteration = 500. The measured average value of the fitness function under different population sizes are listed in Figure 5 and the detailed convergence time are reported in Table 6.
When N p = 300, the objective function reaches the lowest value among all different configurations, so N p = 300 was selected as the optimal parameter setting. In addition, it was also found that 500 iterations did not make the DE solver fully converge, so the maximum iteration number was increased to 1000. The final performance of DE is shown in Figure 6. As the iteration number increases, the minimum and average values of the fitness function gradually coincide. The best anchor settings obtained for the RSOD dataset are summarized in Table 7. The obtained anchors, except for the fourth one, tends to have a larger scale in the height dimension. This phenomenon reveals that our scheme has optimized the anchor settings to best match the objects at varies scales. To more clearly observe the advantage of our proposed algorithm, we have also visualized the distribution of the ground truth bounding boxes and the obtained anchor boxes by using a scatter plot illustrated in Figure 7, in which the results obtained by using K-means clustering are compared. The figures show that the anchors obtained by the proposed algorithm are more evenly distributed among the entire data set, i.e., anchor allocation fully takes into account the distribution of samples in each category. For instance, in the dimension scale of 0.2 to 0.5, the K-means clustering scheme only allocated a single anchor to capture all the objects with large variant in scales, which will inevitably cause degradation in detection accuracy. In contrast, the proposed scheme has allocated four anchors in this range, each of which also corresponds to the clustering centers of a specific category of object. Therefore, it can be concluded that the proposed algorithm can locate the target objects more accurately than the original YOLOv4-Tiny network as the example shows in Figure 8. Note that, in this experiment, both networks have the same number of prediction layers.
Besides the accuracy improvements, the proposed anchor configuration scheme also delivers a faster training speed over traditional approaches. The loss functions obtained by adopting the aforementioned two anchor selection algorithms in training of the neural network models are compared in Figure 9. From the curves, we can see that the training process which adopted the proposed anchor selection scheme converges more quickly, and the final loss drops by about 50% relative to the K-means clustering scheme in the case of using six anchors. Because anchors selected by the K-means clustering scheme are narrowed in a small region, it is difficult for the detector to capture information from those samples outside this region to achieve better matching results. The anchors obtained by the proposed algorithm distribute more evenly in the dataset, which greatly improves the overall learning efficiency of the neural network.
After obtained the optimal anchor setting, the detection accuracy achieved by adding three prediction layers are shown in Table 8. Compared with K-means clustering, the proposed algorithm can improve the detection accuracy by 1.13 % in terms of mAP when using 9 anchors.
Finally, to verify the generality of the proposed scheme, we have also tested the proposed detection framework on the DIOR dataset. As illustrated in Figure 10, the distribution of the obtained anchors by using the proposed scheme is also more evenly allocated than K-means clustering. More detailed results of the measured AP scores of different categories are compared in Table 9. It can be observed that the proposed scheme outperforms K-means clustering by 4.41 % in terms of mAP score.

5.3. Comparison with the State-of-the-Art

We first compare the performance of the proposed detector with three generic detectors, including SSD, YOLOv4, and the YOLOv4-Tiny lightweight network on both the RSOD and DIOR datasets, respectively. Table 10 shows the performance of different networks on RSOD. Detection speed was measured on the same desktop GPU. Compared to YOLOv4-Tiny, the proposed scheme can achieve a considerable 5.13% improvement in mAP, while the cost in speed is an ~50 FPS decline on desktop GPUs. However, the differences in speed will become negligible when deployed in embedded platforms. It can also be observed that the proposed network can even achieve a slightly higher accuracy over the SSD model, while the processing speed is 4 × faster.
In addition, the proposed model was also trained and evaluated on the DIOR dataset, and the experimental results were reported in Table 11 and Table 12. Compared with YOLOv4-Tiny, our method improved the detection accuracy by 3.58%.
Table 13 compares the proposed scheme with state-of-the-art detectors that have been optimized for remote sensing images. Among all the detectors, CSFF [53] and CF2PN [16] have the highest accuracy but the most complex network structure. For instance, CSFF adopted ResNet-101 as the backbone and an FPN as the prediction network, which greatly improved the accuracy of remote sensing object detection. However, the network model is 8× larger than that of the proposed scheme, resulting in a 15× slower processing speed on desktop GPUs. The extremely large neural network model and high computational workload have prohibited similar schemes like CSFF and CF2PN to be deployed on embedded hardware platforms. In contrast, Simple-CNN [54] and ASSD-lite [55] have used simpler backbone structures and more compact network design. These two methods can then achieve real-time processing speed (around 60 FPS) on desktop GPUs. However, the computational workload of these two detectors is still too large to meet the capacity of our target embedded device (i.e., to achieve a real-time speed of 60 FPS, the network model of the detector should have less than 8 GFLOPs workload). The only lightweight detector that can compete with the proposed scheme in terms of the number of parameters is LO-Det [56]. However, in LO-Det, the authors have designed a very complex FPN network, in which channel shuffle and split operations were repeatedly used in each layer. There operations can improve the network’s accuracy but are very unfriendly to parallel processing on GPUs. Therefore, LO-Det only achieved a 4× slower processing speed than our scheme. The proposed detector achieves a considerably higher processing speed of 227.9 FPS than all the reference schemes on the desktop GPU, which reveals that our scheme is not only lightweight in model structure but also very efficient to be executed on GPU devices for parallel processing.

6. Deployment on Embedded Platform

We have deployed the proposed object detection framework on the NVIDIA Jetson Xavier NX board installed on an UAV machine. The proposed lightweight neural network model was quantized into 16-bit floating-point numbers (FP16) by using the TensorRT toolkit. The final prediction accuracy and speed results are reported in Table 14, in which FP32 refers to the standard 32-bit floating-point precision. By quantizing the network model to a reduced precision, the Jetson NX platform can deliver twice the computational capacity than using the standard FP32 data format (i.e., 500 GFLOPs vs. 250 GFLOPs peak performance). In addition, the detection accuracy of the quantized model was well preserved. Thanks to the lightweight, yet efficient network structure proposed, the system can perform high accuracy real-time object detection tasks at the speed of 58.17 FPS on captured remote sensing images with the power consumption of 8.5 W. Therefore, the achieved computational performance is 294 GFLOPs. When compared to the original YOLOv4-Tiny model, our scheme has a significant advantage of 5.1% improvement in detection accuracy, while the sacrifice in speed is not noticeable for practical usage. Moreover, the proposed detector also presented a 15.6% higher computational efficiency over YOLOv4-Tiny. The main reason is that parallelism of the convolutional units is optimized in the device, especially for 3 × 3 conv. Large amount of computation in our proposed is stemming from the increase in 3 × 3 conv, which improves the utilization of device resources.

7. Conclusions

In this paper, we have proposed an efficient lightweight object detector for remote sensing images based on deep convolutional neural networks. To achieve the best balance between detection speed and accuracy, we first designed an improved YOLOv4-like backbone network with three prediction layers to alleviate the problem of multi-scale object detection. Further combined with efficient channel attention to obtain important features, the detector can detect small objects with improved accuracy and no significant overhead in computational workload. Then, an optimal anchor configuration scheme was proposed to solve the problem of obtaining biased anchors due to the large variation in object scales in remote sensing images. Finally, evaluation was conducted on both the RSOD and DIOR datasets, respectively, and comparisons with state-of-the-arts show that the proposed lightweight detector has a significant advantage in processing speed while the detection accuracy is maintained at a close level. Furthermore, real-world deployment on the NVIDIA Jetson Xavier NX verified that our scheme was very suitable for low-cost low-power real-time remote sensing object detection tasks.

Author Contributions

Conceptualization, L.L. and D.W.; methodology, L.L. and D.W.; software, L.L. and Q.Z.; validation, L.L. and Q.Z.; formal analysis, L.L., D.W. and K.X.; writing—original draft, L.L. and D.W.; writing—review and editing, D.W. and K.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially funded in part by National Key Research and Development Program of China under Grant 2019YFB2204200, Beijing Natural Science Foundation under Grant 4202063, Fundamental Research Funds for the Central Universities (Grant 2020JBM020), and BJTU-Kwai Industry-University-Research Cooperation Project.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets presented in this study are available through: https://onedrive.hyper.ai/home/RSOD-Dataset (accessed on 10 April 2021), http://www.escience.cn/people/gongcheng/DIOR.html (accessed on 10 April 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Khan, M.J.; Khan, H.S.; Yousaf, A.; Khurshid, K.; Abbas, A. Super-Resolution Modern Trends in Hyperspectral Image Analysis: A Review. IEEE Access 2018, 6, 14118–14129. [Google Scholar] [CrossRef]
  2. Hong, D.F.; Gao, L.R.; Yokoya, N.; Yao, J.; Chanussot, J.; Du, Q.; Zhang, B. More Diverse Means Better: Multimodal Deep Learning Meets Remote-Sensing Imagery Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4340–4354. [Google Scholar] [CrossRef]
  3. Chen, J.; Wang, G.B.; Luo, L.B.; Gong, W.P.; Cheng, Z. Building Area Estimation in Drone Aerial Images Based on Mask R-CNN. IEEE Geosci. Remote Sens. Lett. 2021, 18, 891–894. [Google Scholar] [CrossRef]
  4. Afaq, Y.; Manocha, A. Analysis on change detection techniques for remote sensing applications: A review. Ecol. Inform. 2021, 63, 101310. [Google Scholar] [CrossRef]
  5. Hoeser, T.; Bachofer, F.; Kuenzer, C. Object Detection and Image Segmentation with Deep Learning on Earth Observation Data: A Review-Part II: Applications. Remote Sens. 2020, 12, 3053. [Google Scholar] [CrossRef]
  6. Wang, P.; Wang, L.G.; Leung, H.; Zhang, G. Super-Resolution Mapping Based on SpatialSpectral Correlation for Spectral Imagery. IEEE Trans. Geosci. Remote Sens. 2021, 59, 2256–2268. [Google Scholar] [CrossRef]
  7. Zhao, Z.Q.; Zheng, P.; Xu, S.T.; Wu, X. Object Detection with Deep Learning: A Review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 142–149. [Google Scholar]
  9. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  10. Zhao, Q.; Sheng, T.; Wang, Y.; Tang, Z.; Ling, H. M2det: A single-shot object detector based on multi-level feature pyramid network. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 9259–9266. [Google Scholar]
  11. Lin, T.Y.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  12. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the Advances in Neural Information Processing Systems 28 (NIPS 2015), Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  13. Du, P.J.; Xia, J.S.; Zhang, W.; Tan, K.; Liu, Y.; Liu, S.C. Multiple Classifier System for Remote Sensing Image Classification: A Review. Sensors 2012, 19, 4764–4792. [Google Scholar] [CrossRef]
  14. Tsagkatakis, G.; Aidini, A.; Fotiadou, K.; Giannopoulos, M.; Pentari, A.; Tsakalides, P. Survey of Deep-Learning Approaches for Remote Sensing Observation Enhancement. Sensors 2019, 19, 3929. [Google Scholar] [CrossRef] [Green Version]
  15. Wang, J.Y.; Wang, Y.Z.; Wu, Y.L.; Zhang, K.; Wang, Q. FRPNet: A Feature-Reflowing Pyramid Network for Object Detection of Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2020, 1–5. [Google Scholar] [CrossRef]
  16. Huang, W.; Li, G.; Chen, Q.; Ju, M.; Qu, J. CF2PN: A Cross-Scale Feature Fusion Pyramid Network Based Remote Sensing Target Detection. Remote Sens. 2021, 13, 847. [Google Scholar] [CrossRef]
  17. Chen, L.C.; Liu, C.S.; Chang, F.L.; Li, S.; Nie, Z.Y. Multiscale object detection in high-resolution remote sensing images via rotation invariant deep features driven by channel attention. Int. J. Remote Sens. 2021, 42, 5754–5773. [Google Scholar]
  18. Qing, Y.H.; Liu, W.Y.; Feng, L.Y.; Gao, W.J. Improved YOLO Network for Free-Angle Remote Sensing Target Detection. Remote Sens. 2021, 13, 2171. [Google Scholar] [CrossRef]
  19. Li, X.G.; Li, Z.X.; Lv, S.S.; Cao, J.; Pan, M.; Ma, Q.; Yu, H.B. Ship detection of optical remote sensing image in multiple scenes. Int. J. Remote Sens. 2021. [Google Scholar] [CrossRef]
  20. Chen, L.C.; Liu, C.S.; Chang, F.L.; Li, S.; Nie, Z.Y. Adaptive multi-level feature fusion and attention-based network for arbitrary-oriented object detection in remote sensing imagery. Neurocomputing 2021, 415, 67–80. [Google Scholar] [CrossRef]
  21. Tian, Z.; Zhan, R.; Hu, J.; Wang, W.; He, Z.; Zhuang, Z. Generating Anchor Boxes Based on Attention Mechanism for Object Detection in Remote Sensing Images. Remote Sens. 2020, 12, 2416. [Google Scholar] [CrossRef]
  22. Mo, N.; Yan, L.; Zhu, R.; Xie, H. Class-Specific Anchor Based and Context-Guided Multi-Class Object Detection in High Resolution Remote Sensing Imagery with a Convolutional Neural Network. Remote Sens. 2019, 11, 272. [Google Scholar] [CrossRef] [Green Version]
  23. Long, Y.; Gong, Y.; Xiao, Z.; Liu, Q. Accurate object localization in remote sensing images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2486–2498. [Google Scholar] [CrossRef]
  24. Mhangara, P.; Mapurisa, W. Multi-Mission Earth Observation Data Processing System. Sensors 2019, 19, 3831. [Google Scholar] [CrossRef] [Green Version]
  25. Han, W.Y.; Liu, X.H. Clustering Anchor for Faster R-CNN to Improve Detection Results. In Proceedings of the IEEE International Conference on Artificial Intelligence and Computer Applications (ICAICA), Dalian, China, 27–29 June 2020; pp. 749–752. [Google Scholar]
  26. Chen, L.; Zhou, L.; Liu, J. Aircraft Recognition from Remote Sensing Images Based on Machine Vision. J. Inf. Process. Syst. 2020, 16, 795–808. [Google Scholar]
  27. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  28. Al-Naji, A.; Perera, A.; Mohammed, S.L.; Chahl, J. Life Signs Detector Using a Drone in Disaster Zones. Remote Sens. 2019, 11, 2441. [Google Scholar] [CrossRef] [Green Version]
  29. Nikulin, A.; de Smet, T.S.; Baur, J.; Frazer, W.D.; Abramowitz, J.C. Detection and Identification of Remnant PFM-1 ’Butterfly Mines’ with a UAV-Based Thermal-Imaging Protocol. Remote Sens. 2018, 10, 1672. [Google Scholar] [CrossRef] [Green Version]
  30. NVIDIA Developer. NVIDIA Embedded-Computing. Available online: https://developer.nvidia.com/embedded-computing (accessed on 20 May 2021).
  31. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014. [Google Scholar]
  32. Girshick, R. Fast r-cnn. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Santiago, Chile, 11–18 December 2015; pp. 1440–1448. [Google Scholar]
  33. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully Convolutional One-Stage Object Detection. arXiv 2019, arXiv:1904.01355. [Google Scholar]
  34. Joseph, R.; Ali, F. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  35. Bochkovskiy, A.; Wang, C.; Liao, H. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  36. Gao, J.F.; Chen, Y.; Wei, Y.M.; Li, J.N. Detection of Specific Building in Remote Sensing Images Using a Novel YOLO-S-CIOU Model. Case: Gas Station Identification. Sensors 2021, 21, 1375. [Google Scholar] [CrossRef]
  37. Hu, X.L.; Liu, Y.; Zhao, Z.X.; Liu, J.T.; Yang, X.T.; Sun, C.H.; Chen, S.H.; Li, B.; Zhou, C. Real-time detection of uneaten feed pellets in underwater images for aquaculture using an improved YOLO-V4 network. Comput. Electron. Agric. 2021, 185, 106135. [Google Scholar] [CrossRef]
  38. Singh, S.; Ahuja, U.; Kumar, M.; Kumar, K.; Sachdeva, M. Face mask detection using YOLOv3 and faster R-CNN models: COVID-19 environment. Multimed. Tools Appl. 2021, 80, 19753–19768. [Google Scholar] [CrossRef] [PubMed]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  40. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  41. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  42. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  43. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
  44. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  45. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  46. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. arXiv 2021, arXiv:2103.02907. [Google Scholar]
  47. Junos, M.H.; Khairuddin, A.; Thannirmalai, S.; Dahari, M. An optimized YOLO-based object detection model for crop harvesting system. IET Image Process. 2021. [Google Scholar] [CrossRef]
  48. Zlocha, M.; Dou, Q.; Glocker, B. Improving RetinaNet for CT Lesion Detection with Dense Masks from Weak RECIST Labels. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention (MICCAI), Shenzhen, China, 13–17 October 2019; pp. 402–410. [Google Scholar]
  49. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11531–11539. [Google Scholar]
  50. Price, K.V. Differential evolution: A fast and simple numerical optimizer. In Proceedings of the North American Fuzzy Information Processing, Berkeley, CA, USA, 19–22 June 1996; pp. 524–527. [Google Scholar]
  51. Opara, K.R.; Arabas, J. Differential Evolution: A survey of theoretical analyses. Swarm Evol. Comput. 2019, 44, 546–558. [Google Scholar] [CrossRef]
  52. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  53. Cheng, G.; Si, Y.J.; Hong, H.L.; Yao, X.W.; Guo, L. Cross-Scale Feature Fusion for Object Detection in Optical Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2021, 18, 431–435. [Google Scholar] [CrossRef]
  54. Li, L.; Cao, G.; Liu, J.; Tong, Y. Efficient Detection in Aerial Images for Resource-Limited Satellites. IEEE Geosci. Remote Sens. Lett. 2021, 1–5. [Google Scholar] [CrossRef]
  55. Xu, T.; Sun, X.; Diao, W.H.; Zhao, L.J.; Fu, K.; Wang, H.Q. ASSD: Feature Aligned Single-Shot Detection for Multiscale Objects in Aerial Imagery. IEEE Trans. Geosci. Remote Sens. 2021, 1–17. [Google Scholar] [CrossRef]
  56. Huang, Z.; Li, W.; Xia, X.G.; Wang, H.; Tao, R. LO-Det: Lightweight Oriented Object Detection in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2021, 99, 1–15. [Google Scholar] [CrossRef]
Figure 1. Demonstration of complicated remote sensing scenes in RSOD dataset. The detected bounding box are draw by using YOLOv4-Tiny.
Figure 1. Demonstration of complicated remote sensing scenes in RSOD dataset. The detected bounding box are draw by using YOLOv4-Tiny.
Sensors 21 05460 g001
Figure 2. The proposed lightweight neural network for remote sensing object detection.
Figure 2. The proposed lightweight neural network for remote sensing object detection.
Sensors 21 05460 g002
Figure 3. Different grid cells on input.
Figure 3. Different grid cells on input.
Sensors 21 05460 g003
Figure 4. Structure of ECA module.
Figure 4. Structure of ECA module.
Sensors 21 05460 g004
Figure 5. Average value of the fitness function under different population sizes.
Figure 5. Average value of the fitness function under different population sizes.
Sensors 21 05460 g005
Figure 6. The relationship between average and minimum value of the fitness function.
Figure 6. The relationship between average and minimum value of the fitness function.
Sensors 21 05460 g006
Figure 7. Distributions of the ground truth bounding boxes and the anchors obtained by using K-means clustering and the proposed scheme on the RSOD dataset.
Figure 7. Distributions of the ground truth bounding boxes and the anchors obtained by using K-means clustering and the proposed scheme on the RSOD dataset.
Sensors 21 05460 g007
Figure 8. Comparison of detection results of YOLOv4-Tiny with the proposed detector. The ground truth is in blue box, and the predicted target is in green color. (a) Original YOLOv4-Tiny network detection results. (b) The proposed algorithm detection results.
Figure 8. Comparison of detection results of YOLOv4-Tiny with the proposed detector. The ground truth is in blue box, and the predicted target is in green color. (a) Original YOLOv4-Tiny network detection results. (b) The proposed algorithm detection results.
Sensors 21 05460 g008
Figure 9. The loss function curve using anchors obtained by different algorithms.
Figure 9. The loss function curve using anchors obtained by different algorithms.
Sensors 21 05460 g009
Figure 10. Comparison of the anchor configuration results in the DIOR dataset.
Figure 10. Comparison of the anchor configuration results in the DIOR dataset.
Sensors 21 05460 g010
Table 1. Comparison of the performances of the SE and ECA modules. Reduction ratios are set to 16 and 3 for SE and ECA, receptively. CSPAn represents the n-th CSPA block.
Table 1. Comparison of the performances of the SE and ECA modules. Reduction ratios are set to 16 and 3 for SE and ECA, receptively. CSPAn represents the n-th CSPA block.
MethodParams/BytesmAP/%
CSPA1CSPA2CSPA3Total
None000080.02
SE1285122048268883.12
ECA9619238467283.34
Table 2. Characteristics of two different hardware platforms.
Table 2. Characteristics of two different hardware platforms.
DeviceNVIDIA GeForce RTX2080Ti Desktop GPUNVIDIA Jeteson Xavier
GPU4352 NVIDIA CUDA cores384-core NVIDIA Volta GPU and 48 Tensorc cores
CPUIntel Core i9-7960x6-core NVIDIA Carmel ARM®v8.2 64-bit CPU
6 MB L2 + 4 MB L3
Memory11 GB 352-bit GDDR6 616 GB/s8 GB 128-bit LPDDR4 51.2 GB/s
Storage4T Hard Disk DrivemicroSD
Power285 W10 W (low-power mode)/15 W
Table 3. Comparison of model size and accuracy under different number of prediction layers (Num.)
Table 3. Comparison of model size and accuracy under different number of prediction layers (Num.)
Num.ParamsFLOPsmAP/%FPSAircraftOiltankOverpassPlayground
25.881 M3.42 G80.02285.372.4697.8551.5698.22
36.527 M5.06 G82.00230.474.7395.9959.3097.98
46.732 M9.83 G83.09182.578.8896.8459.6796.96
Table 4. Comparison of different attention methods.
Table 4. Comparison of different attention methods.
MethodsParamsFLOPsmAP/%FPS
None0 bytes0 k80.02285.3
CBAM2988 bytes422 k79.94268.9
CA5424 bytes351 k79.95273.5
SE2688 bytes2.94 k83.12277.1
ECA672 bytes0.89 k83.34280.8
Table 5. Results of adding the ECA module to the backbone with various numbers of k s i z e .
Table 5. Results of adding the ECA module to the backbone with various numbers of k s i z e .
k size 3579Adaptive
mAP/%83.3481.7082.9780.7581.56
Table 6. The average convergence time under different population sizes.
Table 6. The average convergence time under different population sizes.
Population SizeAverage Value of
the Fitness Function
Average of Convergence Time
(s/Iteration)
1000.27187.99
2000.271315.29
3000.269725.73
4000.271036.37
5000.270046.04
6000.272953.67
7000.271458.69
8000.272759.98
Table 7. Best anchors (two prediction-layer setting) obtained by the proposed scheme.
Table 7. Best anchors (two prediction-layer setting) obtained by the proposed scheme.
Anchor ( θ )Best Decision VariablesAnchor Settings
( θ 1 ( 1 ) , θ 1 ( 2 ) ) ( 0.0327 , 0.0370 ) ( 14 , 15 )
( θ 2 ( 1 ) , θ 2 ( 2 ) ) ( 0.0573 , 0.0667 ) ( 24 , 28 )
( θ 3 ( 1 ) , θ 3 ( 2 ) ) ( 0.1065 , 0.1235 ) ( 44 , 51 )
( θ 4 ( 1 ) , θ 4 ( 2 ) ) ( 0.2296 , 0.4331 ) ( 96 , 180 )
( θ 5 ( 1 ) , θ 5 ( 2 ) ) ( 0.3669 , 0.3501 ) ( 153 , 146 )
( θ 6 ( 1 ) , θ 6 ( 2 ) ) ( 0.4491 , 0.5125 ) ( 187 , 213 )
Table 8. The performance of different anchor settings.
Table 8. The performance of different anchor settings.
MethodsKmAP(%)Anchors
K-means681.55(9,10), (15,15), (22,23), (31,34), (47,53), (145,178)
982.00(8,9), (12,12), (16,15), (19,20), (24,25), (30,32), (40,44), (51,58), (145,178)
Proposed681.69(14,15), (24,27), (44,51), (96,180), (153,146), (187,213)
983.13(13,13), (19,19), (30,31), (41,50), (50,60), (134,171), (99,197), (157,113), (176,214)
Table 9. Mode detailed mAP (%) results of ablation experiments on the DIOR dataset.
Table 9. Mode detailed mAP (%) results of ablation experiments on the DIOR dataset.
MethodsC1C2C3C4C5C6C7C8C9C10C11C12C13C14C15C16C17C18C19C20mAP/%
K-means42.0654.4571.0668.6720.2271.7843.8354.4847.5058.1855.9047.8444.9038.5950.0833.8764.5234.9923.6748.7448.74
DE57.9857.8071.4374.9422.8372.4343.7356.7149.1459.3864.8051.7247.0642.6854.7038.0879.7237.5326.6953.6453.15
Table 10. Measured detection results on the RSOD dataset.
Table 10. Measured detection results on the RSOD dataset.
MethodsSizeParamsFLOPsmAP/%FPSAircraftOiltankOverpassPlayground
SSD300 [8]30024.15 M30.64 G84.7154.270.1290.3478.43100.00
YOLOv4 [35]41663.95 M29.89 G92.5044.896.1398.3875.7899.71
YOLOv4-Tiny [35]4165.881 M3.42 G80.02285.372.4697.8551.5698.22
Proposed4166.527 M5.06 G85.13227.987.1098.9756.5897.86
Table 11. The explanation of each category in DIOR.
Table 11. The explanation of each category in DIOR.
C1C2C3C4C5C6C7C8C9C10
AirplaneAirportBaseball
field
Basketball
court
BridgeChimneyDamExpressway
service area
Expressway
toll station
Golf
course
C11C12C13C14C15C16C17C18C19C20
Ground track
field
HarborOverpassShipStadiumStorage
tank
Tennis
court
Train
station
VehicleWind mill
Table 12. Measured detection results on the DIOR dataset.
Table 12. Measured detection results on the DIOR dataset.
MethodsC1C2C3C4C5C6C7C8C9C10C11C12C13C14C15C16C17C18C19C20mAP
YOLOv4-Tiny [35]58.6155.9971.5774.5222.1972.1147.2654.8348.5060.1164.4651.0946.9241.9355.4237.1879.7836.2726.4952.2352.87
Proposed58.1655.6272.3976.0125.8673.0343.3155.4351.3958.9466.0351.3048.6970.4151.8253.3482.4638.7832.6063.3356.45
Table 13. Comparison with state-of-the-art detectors for remote sensing images. The performance data were measured on the DIOR dataset.
Table 13. Comparison with state-of-the-art detectors for remote sensing images. The performance data were measured on the DIOR dataset.
ApproachCSFF [53]CF2PN [16]Simple-CNN [54]ASSD-Lite [55]LO-Det [56]Proposed
Year202120212021202120212021
BackboneResNet-101VGG16VGG16MobileNetv2MobileNetv217-layer-CNN
Parameters>46 M91.6 M23.53 M>24 M6.93 M6.5 M
FPS15.2119.7133564.52227.9
mAP6867.2566.563.358.7356.45
DeviceRTX3090RTX2080TiGT710GTX 1080TiRTX3090RTX2080Ti
Table 14. Measured accuracy, speed, and efficiency on the NVIDIA Jetson Xavier NX board.
Table 14. Measured accuracy, speed, and efficiency on the NVIDIA Jetson Xavier NX board.
MethodsmAP (FP32)mAP (FP16)FPS (FP16)Efficiency (100%)
YOLOv4-Tiny80.02%80.22%63.2843.2%
Proposed85.13%85.33%58.1758.8%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lang, L.; Xu, K.; Zhang, Q.; Wang, D. Fast and Accurate Object Detection in Remote Sensing Images Based on Lightweight Deep Neural Network. Sensors 2021, 21, 5460. https://doi.org/10.3390/s21165460

AMA Style

Lang L, Xu K, Zhang Q, Wang D. Fast and Accurate Object Detection in Remote Sensing Images Based on Lightweight Deep Neural Network. Sensors. 2021; 21(16):5460. https://doi.org/10.3390/s21165460

Chicago/Turabian Style

Lang, Lei, Ke Xu, Qian Zhang, and Dong Wang. 2021. "Fast and Accurate Object Detection in Remote Sensing Images Based on Lightweight Deep Neural Network" Sensors 21, no. 16: 5460. https://doi.org/10.3390/s21165460

APA Style

Lang, L., Xu, K., Zhang, Q., & Wang, D. (2021). Fast and Accurate Object Detection in Remote Sensing Images Based on Lightweight Deep Neural Network. Sensors, 21(16), 5460. https://doi.org/10.3390/s21165460

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop