Next Article in Journal
A Watermarking Optimization Method Based on Matrix Decomposition and DWT for Multi-Size Images
Previous Article in Journal
Research on the Life Prediction Method of Meters Based on a Nonlinear Wiener Process
Previous Article in Special Issue
Infrared and Visible Image Registration Based on Automatic Robust Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SV-FPN: Small Object Feature Enhancement and Variance-Guided RoI Fusion for Feature Pyramid Networks

1
School of Science, Beijing University of Civil Engineering and Architecture, Beijing 100044, China
2
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(13), 2028; https://doi.org/10.3390/electronics11132028
Submission received: 18 May 2022 / Revised: 19 June 2022 / Accepted: 23 June 2022 / Published: 28 June 2022
(This article belongs to the Special Issue Computer Vision Techniques: Theory and Applications)

Abstract

:
Small object detection is one of the research difficulties in object detection, and Feature Pyramid Networks (FPN) is a common feature extractor in deep learning; thus, improving the results of small object detection based on FPN is of great significance in this field. In this paper, SV-FPN is proposed for a small object detection task, which consists of Small Object Feature Enhancement (SOFE) and Variance-guided Region of Interest Fusion (VRoIF). When using FPN as a feature extractor, an SOFE module is designed to enhance the finer-resolution level feature maps from which the small object features are extracted. VRoIF takes the variance of RoI features as the data driver to learn the completeness of several RoI features from different feature layers, which avoids wasting information and introducing noise. Ablation experiments on three public datasets (KITTI, PASCAL VOC 07+12 and MS COCO 2017) demonstrate the effectiveness of SV-FPN, and the mean Average Precision (mAP) of SV-FPN in the three datasets achieves 41.5%, 53.9% and 38.3%, respectively.

1. Introduction

In recent years, with the advent of the era of big data, object detection based on deep learning has developed rapidly. Object detection is one of the basic tasks in computer vision, which is widely used in many areas, such as medical imaging, social security, smart homes, and transportation, etc. However, small object detection is a challenging task, and this is mainly because: (1) small objects occupy fewer pixels, which will result in indistinguishable features after passing through the convolutional neural networks (CNNs); (2) unbalanced distribution of small objects in the dataset will skew the network training process; (3) the setting of the network structure is not friendly to small objects. Therefore, many meaningful works are proposed.
Initially, some works were proposed for multi-scale object detection tasks. Lin et al. proposed Feature Pyramid Networks (FPN) for multi-scale detection [1]. Singh et al. optimized the candidate boxes on each layer of the image pyramid to improve the multi-scale detection accuracy [2]. Cao et al. used high-level features to enhance the semantic information of low-level features to improve multi-scale object detection [3]. Ren et al. proposed an anchor box, which could effectively detect objects of different sizes by designing different sized anchor boxes [4]. As technology develops, more and more works are focused on improving small object detection. Hu et al. proposed Context-Aware Region of Interest (RoI) Pooling to prevent small objects from being distorted after RoI Pooling [5]. Mate et al. proposed a small object data enhancement method by oversampling, copying and pasting small objects to expand the number and distribution of small objects in the training set [6]. Li et al. used a Perceptual Generative Adversarial Network (PGAN) to map small low-resolution objects into super-resolution large objects [7]. Chen et al. proposed a loss function feedback mechanism [8]. During the training process, the input is dynamically adjusted by the proportion of the small object in the loss function so that the network can optimize the object balance of each size. Bharat et al. proposed Scale Normalization for Image Pyramids (SNIP), which selectively back-propagates the gradients of object instances of different sizes as a function of the image scale [2]. Bharat et al. proposed SNIPER to save training overhead and speed up network convergence [9]. These works were carried out for small objects, which may sometimes affect medium and large objects.
Our contributions are as follows:
First, we propose SV-FPN to improve the result of small object detection, which would not affect the medium or large objects.
Second, we design an SOFE module between C 2 and P 2 in FPN to enhance small object features.
Third, we propose VRoIF, which dynamically fuses RoI features by the variance to enhance object features.
Fourth, some extensive empirical studies are performed to evaluate the effectiveness of the proposed method. The results show the effectiveness of SV-FPN for small object detection.
This paper is organized as follows: related concepts are introduced in Section 2. Section 3 explains the proposed works in detail. Subsequently, some experiments are implemented to prove the superior implementation performances of our works in Section 4. Finally, Section 5 concludes this work.

2. Related Works

2.1. Small Object Detection

To the best of our knowledge, small objects have smaller areas and fewer pixels than medium and large objects. There are two definitions of a small object; one is that the object area is less than 32 × 32 pixels [10], and the other is that the ratio of the object area to the image area is between 0.08% and 0.58% [11].
Object detection based on deep learning is usually performed by convolution-activation-pooling in the feature extraction stage, and pooling would down-sample the features, resulting in the loss of object features. Therefore, the feature information of small objects is less than medium and large objects, and less available information leads to poor detection results. Therefore, many works were proposed to improve small object detection from other aspects.

2.1.1. Data Augmentation

Data enhancement is widely used in object detection. CutOut [12], MixUp [13], CutMix [14], and Mosaic were often used to adjust the input of images to improve the robustness of the network [15]. Kisantal et al. proposed a small object data enhancement method to expand the number and distribution of small objects in the training set to enrich the diversity of the dataset, thereby improving the detection performance of small objects [6]. Yu et al. proposed a scale matching strategy, which cuts according to the size of the object to narrow the gap between objects of different sizes and avoids the situation where the information of small objects is easily lost in conventional scaling operations [16].

2.1.2. Multi-Scale Learning

In the shallow layer of CNNs, the receptive field is small, the semantic information is weak, and the context information is inferior, but the shallow layer has more spatial and detailed feature information. Liu et al. proposed SSD (Single Shot Multibox Detector), which used shallower features to detect smaller objects and deeper features to detect larger objects [17]. In order to save computing resources and obtain better feature fusion effects, Lin et al. proposed FPN (Feature Pyramid Network), which is the most popular multi-scale network at present [1]. The construction of FPN involves a bottom-up pathway, a top-down pathway, and lateral connections, and FPN achieves the purpose of feature enhancement by fusing the features of adjacent layers.

2.1.3. Contextual Learning

In real-world scenarios, there is usually a relationship between “object to scene” and “object to object”, and using this relationship will help improve the detection performance of small objects. Treating context utilization as an optimization problem, Barnea et al. discussed how much context or other types of additional information could improve detection scores and showed that simple co-occurrence relationships are the most effective context information [18]. Chen et al. proposed a hierarchical context embedding framework to enhance the feature representation of candidate regions by mining contextual cues [19]. Fu et al. modeled and inferred the inherent semantics and spatial layout relationship between objects and retained the spatial information as much as possible when extracting the semantic features of small objects [20]. Pato et al. proposed a rescoring algorithm based on contextual rescoring, which used RNNs (Recurrent Neural Networks) and self-attention mechanisms to transfer information between candidate regions and generate contextual representation, and used the obtained context information to perform a secondary evaluation of the detection results [21].

2.1.4. Generative Adversarial Learning

Generative Adversarial Learning aims to achieve the same detection performance as larger-sized objects by mapping the features of small low-resolution objects into features equivalent to those of high-resolution objects. Li et al. proposed Perceptual Generative Adversarial Network (PGAN) specifically for small object detection, which learns high-resolution feature representations of small objects by confronting the generator and discriminator with each other [7]. Bai et al. proposed a multi-task generative adversarial network (MTGAN) for small objects. In MTGAN, the generator is a super-resolution network that can upsample small blurred images into fine images and recover detailed information for more accurate detection [22]. Noh et al. proposed a new feature-level super-resolution method, which maintains the same receptive field between the generated high-resolution object features and the low-resolution features generated by the feature extractor through dilated convolution, avoiding the problem of generating false over-score features due to mismatched receptive fields [23].

2.1.5. Other Strategies

Although these methods mentioned above can effectively improve the performance of small object detection, the performance gains brought by these methods are often limited by the computational cost. There are many excellent works that have improved small object detection from other aspects. Wang et al. proposed a guided anchoring strategy based on semantic features, which improved the performance of small object detection by simultaneously predicting the possible location of the object center and the scale and aspect ratio of the object [24]. Sun et al. proposed a multi-receptive field and small-object-focused weakly supervised segmentation network by using multiple receptive field blocks to focus on the object and its adjacent background and set weights according to different spatial positions to achieve enhanced feature availability [25].

2.2. Feature Pyramid Networks

The design goal of the FPN is to use the pyramid feature generated by CNNs to construct a feature pyramid with multi-resolution and use the feature pyramid for multi-scale object detection. FPN is currently the most popular multi-scale network [1]. Figure 1 shows the structure of FPN, where { C 2 , C 3 , C 4 , C 5 } is a bottom-up path and { P 2 , P 3 , P 4 , P 5 } is a top-down path.
To improve detection performance, there are many meaningful works developed on FPN [1]. In order to facilitate the growth of information flow in the recommended region-based instance segmentation framework, Liu et al. proposed PANet [26]. By analyzing the defects of FPN, AugFPN proposed targeted improvement plans, which were Consistent Supervision, Residual Feature Augmentation, and Soft RoI Selection [27]. NAS-FPN proposed merging cells to re-integrate the features extracted by FPN for better detection accuracy [28]. Auto-FPN used the search network to automatically find the most suitable connection between different layers for better features [29]. A 2 -FPN used attention-guided feature aggregation to improve multi-scale feature learning [30].
Although there are many works based on FPN, they generally focus on the defects of the network and do not focus on improving small object detection based on FPN.

2.3. Preliminaries

In FPN, the image is down-sampled by CNNs to obtain C 5 , and the feature maps { P 4 , P 3 , P 2 } are generated by the up-sampling of P 5 . The bottom-up feature map has accurately localized information because they are subsampled fewer times, so the feature maps are generated by the up-sampling of P 5 and would fuse with the bottom-up pathway to enhance their position information. However, the dimension (the number of channel) of { C 2 , C 3 , C 4 , C 5 } is {256, 512, 1024, 2048}, and the dimension of { P 2 , P 3 , P 4 , P 5 } is {256, 256, 256, 256}; therefore, the 1 × 1 convolution is used to reduce the dimension of { C 2 , C 3 , C 4 , C 5 } before the fusion of the two pathways.
RoI feature extraction is a research hotspot. Anchor-Base object detection uses RoI Pooling or RoI Align to obtain RoI features of the final feature map. When using FPN as a feature extractor, there are four feature maps, so various RoI feature extraction strategies are proposed. RoI feature extraction is based on two research directions: one is area assign, which is allocated RoI to a certain layer of FPN to extract features according to the area of RoI; the other is feature fusion, which combines the RoI features from { P 2 , P 3 , P 4 , P 5 }, then fuses them in some way.
FPN proposed that different scales of RoIs would be assigned to different feature pyramid levels decided by Equation (1) [1]:
k = k 0 + log 2 ( w h / 224 )
Here, k is the assigned feature level in FPN, so the range of k is [2, 5], k 0 is the assigned result when the area of the RoI is 224 × 224, w and h represent the length and width of the RoI, 224 is the canonical input size of ImageNet. It can be seen that the small objects will acquire features from P 2 in FPN. However, 224 is a hard-coded, which cannot adapt to input image scale variation, so Lee et al. proposed using the relative area of RoI in the image to replace the absolute area in Equation (1), and a new RoI assignment function is defined as Equation (2) [31]:
k = k m a x log 2 ( A i n p u t / A R o I )
Here, k is the assigned feature level in FPN, k m a x represents the high level of FPN, A R o I represents the area of the object in the input image, and A i n p u t represents the area of the image. However, RoI features obtained from one certain feature level may not be the optimal method, which may ignore useful information from other feature levels in FPN.
Considering that the importance of features may not be strongly correlated to the pyramid levels they belong to, PANet proposed adaptive feature pooling, which pooling feature from { P 2 , P 3 , P 4 , P 5 } and fused them with maximum operation [26]. Because the maximum operation may ignore some useful information and the fully connected layer in adaptive feature pooling resulted in additional parameters, AugFPN proposed adaptive spatial fusion, which used the result of global average pooling as the channel initial attention weight, and then the learned weight was applied to the features correspondingly; finally, four features were added pixel-by-pixel [27]. PANet and AugFPN considered fusing the feature to make full use of features but did not consider whether the fusion method is suitable for this task. In addition, the excessive difference between the input and output size of the RoI Align will result in perturbed features.
Therefore, this paper proposes VRoIF. Firstly, the variance of RoI features is used to replace the max operation or average pooling of the previous method to obtain the initial weight of the features on { P 2 , P 3 , P 4 , P 5 }. At the same time, in order to ensure the integrity of the RoI feature, the result of the area assigned would be weighted and added to the weighted features.

3. Materials and Methods

3.1. The Overall Structure of SV-FPN

Figure 2 shows the structure of SV-FPN. Small Object Features Enhancement (SOFE) module and Variance-guided RoI Fusion (VRoIF) were conducted in FPN to improve the result of small objects. The SOFE module was added to lateral connections between C 2 and P 2 , which was designed to enhance P 2 , from which small object features are extracted. We propose VRoIF to allow each RoI to use variance as weight initialization to fuse its feature from different feature levels in FPN, and the result of area assignment is weight-added to ensure feature integrity.

3.2. Small Object Features Enhancement

Convolution extracts certain specified features of an image, such as textures, gradients, etc. There are many different-sizes of convolution kernels. A larger convolution kernel has a larger receptive field and can obtain more comprehensive features. The role of 1 × 1 convolution: (1) enable cross-channel interaction and information integration; (2) increase nonlinear features; (3) reduce model parameters and calculation. It is worth noting that stacking multiple small convolution kernels can achieve the receptive field of a large convolution kernel with more non-linearity and less parameter computation.
Small object features are extracted from P 2 in FPN, which is obtained by adding C 2 and P 3 . C 2 has accurate localization and simple semantic information, but the lack of deep convolution would bring noise to P 2 . Therefore, it is necessary to design a deep convolution block for C 2 to improve its abstract semantic information and reduce detailed information.
We propose a Small Object Feature Enhancement (SOFE) module, which enhances the small object features based on FPN, and the flowchart of SOFE is shown in Figure 3. As illustrated in Figure 4, a SOFE module consists of 9 CBR modules, which are divided into 3 branches. The CBR module consists of three parts, namely CONV, BN and ReLU, and CBRn represents the size of the convolution kernel as n × n. The convolution kernels in the second CBR module of each branch are designed to different sizes: the 1 × 1 convolution branch is used to obtain inter channel correlation information and retain small object features, the 3 × 3 convolution branch is used to extract features, and the 5 × 5 convolution branch is used to capture the context information around the object so that the context information is helpful for detection. And the pseudocode of SOFE is shown in Algorithm 1. It is worth noting that in order to reduce model complexity and training time, the convolution kernel in the first and third CBR modules of each branch is set to 1 × 1, and the 5 × 5 convolution kernel here is replaced by two 3 × 3 convolution kernels.
By adding the SOFE module to FPN, with deep convolution to extract abstract features and positional information, P 2 is much more friendly to small object detection.
Algorithm 1 Small Object Feature Enhancement.
Input: feature map C 2
Output: enhanced C 2
1:
Enhance C 2 :
2:
    F 1 C B R 1 ( C B R 1 ( C B R 1 ( C 2 ) ) )
3:
    F 2 C B R 1 ( C B R 3 ( C B R 1 ( C 2 ) ) )
4:
    F 3 C B R 1 ( C B R 5 ( C B R 1 ( C 2 ) ) )
5:
Feature fusion:
6:
    F F 1 + F 2 + F 3
7:
Convolution:
8:
    F C o n v ( F )
9:
Residual connection:
10:
   Enhanced C 2 F + C 2

3.3. Variance-Guided RoI Fusion

For the first RoI assignment method, the correspondence of RoIs and FPN levels depends on Equation (1). Table 1 presents the range of RoI area assigned to { P 2 , P 3 , P 4 , P 5 }, where A R o I represents the area of an object in an input image, and A R o I represents the area of an object in feature maps.
The output size of RoI Align is usually 7 × 7 or 14 × 14. It can be seen from the last column of Table 1 that the relative size of the RoI assigned to each FPN layer of { P 2 , P 3 , P 4 , P 5 } is close to the output size of RoI Align, which provides a complete feature.
For the other RoI assignment method, small objects fuse the features from surrounding objects in the low-resolution feature map, which have strong semantic features, and max pooling or average pooling as the weight initialization would lead to larger feature weights when the feature is fused. Large objects in the high- resolution feature map would be lost and result in incomplete features.
Parameter initialization has a great impact on the model. A good parameter initialization can speed up the convergence of the network and improve the generalization ability of the network. The result of max pooling or average pooling as initial parameters may not lead to a better result. Therefore, we want to find an observation to measure the completeness and richness of the feature and use the result of the observation as the initial parameters.
In statistical descriptions, variance is used to calculate the difference between each observation and the population mean. Abstract object features should have obvious changes to provide more information for detection tasks, so it is worth thinking about using variance as an observation to measure the completeness of the RoI feature.
In order to verify the relationship between the variance of object features and the detection results, we applied variance statistics to the object features of the network with different mean Average Precisions (mAP). Taking Faster R-CNN, Mask R-CNN and Cascade R-CNN as experimental networks and using ResNet50 and ResNet101 as backbone networks, the statistical results are shown in Table 2.
In Table 2, 1× is the result of 12 training epochs, and 2× is the result of 24 training epochs. The first column under ResNet50 and ResNet101 is the mAP, and the second column is the mean variance of the regression box features in the validation set of MS COCO 2017. It can be seen from Table 2 that in the same network structure, the longer the training, the higher the accuracy, and the larger the feature variance of the regression box. Therefore, the feature map with large variance is more friendly to object detection.
Based on the above analysis, variance can effectively describe the completeness and richness of the feature. Therefore, we propose a Variance-guided RoI Fusion (VRoIF) module, which is shown in Figure 5 and the flowchart of VRoIF is shown in Figure 6. VRoIF uses the variance of RoI features as a data-driver to obtain the weight of RoI features in { P 2 , P 3 , P 4 , P 5 }, and the weighted RoI features from { P 2 , P 3 , P 4 , P 5 } are fused to obtain F v a r . The calculation of variance in feature maps is shown in Equation (3), where x is the pixel sample value of an RoI feature, x ¯ is the average value of pixel sample in the RoI features, and n is the number of pixel samples.
S 2 = i = 1 n ( x x ¯ ) 2 n 1
Considering the size difference of the input and output of RoI Align, F v a r is easily disturbed by adjacent objects, so the result of area assignment, F a r e a , will be added to F v a r . For each RoI, firstly, VRoIF extracts RoI features from { P 2 , P 3 , P 4 , P 5 }, and the area assignment result F a r e a is obtained by Equation (1). Secondly, the variance of RoI features is calculated, which is used as the original data to drive the learning weights of the RoI features. Thirdly, the learned weights are multiplied with the corresponding RoI features, and then a convolution is conducted to fuse RoI features. Finally, F v a r and F a r e a are weight-added to obtain the RoI features F R o I by Equation (4):
F R o I = α F a r e a + ( 1 α ) F v a r
Here, α is a hyper-parameter in the range of (0, 1), and the detection result is optimal when it is 0.5. And the pseudocode of VRoIF is shown in Algorithm 2.
Algorithm 2 Variance-guided RoI Fusion.
Input: feature map C 2
Output: enhanced C 2
1:
k k 0 + log 2 ( w h / 224 )
2:
    F a r e a is RoI feature extracted from P k
3:
F concatenating RoI features extracted from { P 2 , P 3 , P 4 , P 5 }
4:
W e i g h t C o n v ( C o n v ( V a r ( F ) ) )
5:
F v a r C o n v ( W e i g h t · F )
6:
F R o I α F a r e a + ( 1 α ) F v a r

4. Experiments

4.1. Datasets and Evaluation Metrics

In the experiments, we used the public datasets: KITTI [32], PASCAL VOC 07+12 [33] and MS COCO 2017. KITTI consists of more than 7K images, of which 3712 are in the training set and 3769 are in the validation set, with a total of 7 categories. PASCAL VOC 07+12 includes around 16K images, of which 4952 images are used for validation, with a total of 20 categories. MS COCO 2017 contains 118K images for training, 5K images for validation, with a total of 80 categories.
In the KITTI dataset, we used the standard evaluation metric proposed in PASCAL Visual Objects Classes: mAP, and the IoU threshold is set to 0.5. In PASCAL VOC 07+12 and MS COCO 2017, we follow the standard evaluation metric proposed in MS COCO: mAP, AP 50 , AP 75 , AP S , AP M and AP L , mAP is over 10 IoU thresholds (IoU = 0.50, 0.55, …, 0.90, 0.95) to calculate, AP 50 and AP 75 represent the AP value when the IoU threshold is 0.50 and 0.75, AP S , AP M and AP L can clearly describe the detection performance with different object scales.

4.2. Implementations Details

All experiments were implemented based on MMDetection [34] and exploited with Ubuntu18.04, PyTorch v1.6.0, CUDA 10.1 and CUDNN 7.1.4. We trained models with 1 GPU (2 images per GPU) for 12 epochs. The selected baseline network is Faster R-CNN, and the initial learning rate was set as 0.0025 and decreased by a ratio of 0.1 after the 8th and 11th epoch, respectively. All other hyper-parameters in this paper remain the same as MMDetection.

4.3. Experiment Results and Analysis

4.3.1. Weight Fusion Operation Selection

Max pooling and average pooling are widely used in deep learning for fusion operation. PANet proposed adaptive feature pooling, which used max pooling as the RoI feature fusion method, and AugFPN proposed Soft RoI Selection, which used average pooling as the initial weight of RoI feature fusion. However, max pooling and average pooling are not suitable for all feature fusion situations. Therefore, in order to verify the superiority of variance-guided methods, we take max pooling, average pooling and variance as weight initialization to conduct experiments on the PASCAL VOC 07+12. The experimental results are shown in Table 3.
As can be seen from Table 3, except for AR L , the rest of the evaluation metrics for the model initialized with variance-guided fusion outperforms max pooling and average pooling. In small object detection, the performance of variance-guided fusion is more eye-catching: compared with average pooling, AP S is increased by 4.1%, and AR S is increased by 2.6%; compared with max pooling, AP S is increased from 24.7% to 25.0%, AR S is increased from 33.4% to 37.3%. AP is an indicator that is used to measure the regression quality of the prediction box, and AR is an indicator to measure the number of detected objects. Therefore, the experimental data indicate that when using the variance-guided option in RoI feature fusion, not only the regression quality of the predict box would be improved but also much smaller objects would be detected. It also shows that the weights based on variance-guided fusion are better than max pooling and average pooling.

4.3.2. Proportional Statistics

In order to analyze the proportion of RoI features of different sizes from different feature layers in FPN, we count the sources of F v a r . Figure 7 shows that RoIs with different sizes all exploit features from several different levels with variance-guided fusion. The blue polyline represents small objects that were assigned to P 2 originally in FPN, and nearly 48% of the features are from P 2 with variance-guided fusion, which explains the integrity of features is useful for detection. The green polyline and red polyline represent large objects that were assigned to P 4 and P 5 originally in FPN, and it can be clearly seen that their peaks are slightly higher than those of the other two polylines, which indicates that P 4 and P 5 have stronger semantic information. The peaks of the four polylines all appear in the result of Equation (1), which is consistent with our above analysis.

4.3.3. Hyper-Parameter

In order to explore the influence of α in Equation (4) on the detection accuracy, we set α to {0, 0.1, 0.2, …, 0.8, 0.9} to conduct experiments on the PASCAL VOC 07+12 to find the optimal solution. The results are shown in Table 4. When α is 0.5, mAP reaches the optimal value, and when α is 0.6, AP S reaches the optimal value. It is worth noting that the result of non-zero is better than zero, which verifies that the result of the area assignment is beneficial to detection.

4.3.4. Component Ablation Studies

In order to verify the effectiveness of SV-FPN, some ablation experiments are performed in this section. We first take Faster R-CNN + FPN as the baseline experiment because the two modules in SV-FPN are proposed based on FPN. Then, we replace the lateral connection between P 2 and C 2 with the SOFE module to enhance all of the small object features. At the same time, all of the medium objects and a few large objects feature are enhanced. After that, VRoIF is added to fuse RoI features from different FPN levels, which could capture richer RoI features for multi-scale objects. Finally, SV-FPN, which combines the SOFE module and VRoIF, is used as the backbone of the detection networks. Table 5, Table 6 and Table 7 present the ablation experiment results on the public datasets KITTI, PASCAL VOC 07+12 and MS COCO 2017, respectively.
Objects in the KITTI dataset are mostly medium and small objects. It can be seen from Table 5 that after adding the SOFE module, the AP of most categories is increased in the range of 0.3–7.4%, except for the van, and mAR is increased by 3.2%; after adding the VRoIF module, in addition to the slight decrease in the AP of the car, a few objects are the same as the baseline networks, the rest of the categories have been improved to varying degrees, the AP of trams has increased by 8.3% especially, and mAR has been increased from 56% to 58.7%. Last but not least, when adding two modules, that is, when using SV-FPN as the feature extractor, AP and AR were greatly improved—from 38.8% to 41.5%, and from 56% to 60.6%, respectively. The improvement of AP means that the predicted boxes are more accurate, and classification accuracy would be improved; the improvement of mAR means that more objects are detected. Therefore, SV-FPN can ensure the improvement in detection accuracy and reduction in the missed detection rate.
It can be seen from Table 6 that the SOFE module significantly improves the accuracy of small objects on PASCAL VOC 07+12. AP S is increased from 22.4% to 24.4%, and mAP is increased by 0.6%, which reveals that the SOFE module has a positive impact on small object detection. In addition, the VRoIF module improves AP S by 4.1%, AR S by 3.0%, and mAR by 0.7%, which verifies that the VRoIF module can result in significant performance improvements for small objects. When both the SOFE module and VRoIF module are added to the network, mAP reaches the optimal value of 53.9%, and mAR also reaches 63.8%, which is 1.7% higher than that of the baseline network. However, the small object detection result of SV-FPN is worse than that of adding only one module.
It can be seen from Table 7 that the performance of the SOFE module and VRoIF module on MS COCO 2017 is consistent with the performance on PASCAL VOC 07+12: the two modules significantly improved AP S , which has increased by 0.8% and 1.0%, respectively, and AR S improved too. The two modules also have a corresponding improvement in the detection ability of the network, increasing mAP to 38.3%, but AP S is also worse than adding only one module.
From the above experimental results, it can be seen that VRoIF has improved multi-scale object detection results on three public datasets, which proves that VRoIF is better than the baseline method and effectively improved the multi-scale object prediction accuracy. SV-FPN, which embeds the SOFE module and VRoIF into FPN, achieved the best detection accuracy in all ablation experiments on the three datasets: the baseline network mAP is improved by 2.7%, 1.6% and 0.8%, respectively, proving that SV-FPN not only improves the small objects but also improves the detection effect of the medium and the large objects to a certain extent.
In addition, Figure 8 shows the mAP trend of FPN and SV-FPN during the training process. It was found that the mAP of SV-FPN outperforms FPN at the beginning of training, and the leading advantage is obvious. It is only when the training is relatively stable that the mAP of the two modules is not very different. However, after the decay of the learning rate, the mAP of SV-FPN is still significantly better than that of FPN, and the leading trend gradually stabilizes during the network convergence.
The ablation experiment results prove that the improvement in the SOFE module is mainly attributed to the small object detection, while VRoIF is not limited to small objects, and the multi-scale object detection results also improved. However, when the two components are used simultaneously, SV-FPN achieves the best performance, of which AP S is worse than when only using the SOFE module or VRoIF. This is mainly because the position of the two proposed components is close in the detection network, which will cause feature disturbance for small object features.

4.3.5. Example Display

For a fair observation and to verify the improvement of SV-FPN in small object detection, we visualized the detection results on the validation sets of three public datasets: KITTI, PASCAL VOC 07+12 and MS COCO 2017, and selected some images to show the differences between the baseline and SV-FPN in different scenarios.
The visualization results of the three datasets are shown in Figure 9, Figure 10 and Figure 11. The first row is the visual results of the baseline, and the last row is the visual results of Faster R-CNN + SV-FPN. In addition, there is a comparison of the results between the first row and the last row.
As shown in Figure 9, in the traffic scene image, it can be clearly seen that compared with the baseline detection results, SV-FPN can accurately detect small objects that are difficult to identify not only in the distance but also close by. The SOFE module and VRoIF in this scene are beneficial to improve the detection effects of small objects.
As shown in Figure 10 and Figure 11, in the multi-scene images of PASCAL VOC 07+12 and MS COCO 2017, SV-FPN significantly improved the detection results of all kinds of objects, such as person, cars, etc. Compared with FPN, the visual effect of SV-FPN is better. Although the SOFE module and VRoIF improve small object detection performance from different angles, the visual results show that they could effectively detect small objects, which are ignored by the baseline. The above visualization results verify the effectiveness of SV-FPN.

5. Conclusions

In this paper, we propose SV-FPN to improve the detection accuracy of small objects. Firstly, we design the SOFE module to enhance small object features, which is placed in the lateral connections between P 2 and C 2 . After that, in order to enhance the RoI features of the small objects, VRoIF uses variance as the data driver to adaptively weight the RoI features from different FPN layers to obtain the weighted RoI feature map F v a r , and the result of the area allocation method F a r e a will be weight-added to ensure the integrity of the object features. Some experiments on KITTI, PASCAL VOC 07+12 and MS COCO 2017 show that SV-FPN improves the detection accuracy and visual results of small objects.

Author Contributions

Conceptualization, Q.Y. and C.Z.; methodology, Q.Y.; software, Q.Y.; validation, Q.Y., C.Z. and H.W.; formal analysis, Q.Y. and C.Z.; writing—original draft preparation, Q.Y., C.Z. and H.W.; writing—review and editing, C.Z., Q.H. and L.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (nos. 62072024, 41971396), R&D Program of Beijing Municipal Education Commission (KM202210016002), the Projects of Beijing Advanced Innovation Center for Future Urban Design (no. UDC2019033324, UDC2017033322), and the Fundamental Research Funds for Municipal Universities of Beijing University of Civil Engineering and Architecture (nos. X20084, ZF17061).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  2. Singh, B.; Davis, L.S. An analysis of scale invariance in object detection snip. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3578–3587. [Google Scholar]
  3. Cao, G.; Xie, X.; Yang, W.; Liao, Q.; Shi, G.; Wu, J. Feature-Fused SSD: Fast detection for small objects. In Proceedings of the Ninth International Conference on Graphic and Image Processing (ICGIP 2017), International Society for Optics and Photonics, Qingdao, China, 14–16 October 2018; Volume 10615, p. 106151E. [Google Scholar]
  4. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28. [Google Scholar] [CrossRef] [PubMed]
  5. Park, H.; Sjosund, L.; Yoo, Y.; Monet, N.; Bang, J.; Kwak, N. Sinet: Extreme lightweight portrait segmentation networks with spatial squeeze module and information blocking decoder. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 1–5 March 2020; pp. 2066–2074. [Google Scholar]
  6. Kisantal, M.; Wojna, Z.; Murawski, J.; Naruniec, J.; Cho, K. Augmentation for small object detection. arXiv 2019, arXiv:1902.07296. [Google Scholar]
  7. Li, J.; Liang, X.; Wei, Y.; Xu, T.; Feng, J.; Yan, S. Perceptual generative adversarial networks for small object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1222–1230. [Google Scholar]
  8. Chen, Y.; Zhang, P.; Li, Z.; Li, Y.; Zhang, X.; Qi, L.; Sun, J.; Jia, J. Dynamic Scale Training for Object Detection. arXiv 2020, arXiv:2004.12432. [Google Scholar]
  9. Singh, B.; Najibi, M.; Davis, L.S. Sniper: Efficient multi-scale training. Adv. Neural Inf. Process. Syst. 2018, 31. [Google Scholar] [CrossRef]
  10. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  11. Chen, C.; Liu, M.Y.; Tuzel, O.; Xiao, J. R-CNN for small object detection. In Proceedings of the Asian Conference on Computer Vision, Taipei, Taiwan, 20–24 November 2016; pp. 214–230. [Google Scholar]
  12. DeVries, T.; Taylor, G.W. Improved regularization of convolutional neural networks with cutout. arXiv 2017, arXiv:1708.04552. [Google Scholar]
  13. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  14. Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 6023–6032. [Google Scholar]
  15. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  16. Yu, X.; Gong, Y.; Jiang, N.; Ye, Q.; Han, Z. Scale match for tiny person detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 4–8 January 2020; pp. 1257–1265. [Google Scholar]
  17. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  18. Barnea, E.; Ben-Shahar, O. Exploring the bounds of the utility of context for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7412–7420. [Google Scholar]
  19. Chen, Z.M.; Jin, X.; Zhao, B.; Wei, X.S.; Guo, Y. Hierarchical context embedding for region-based object detection. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 633–648. [Google Scholar]
  20. Fu, K.; Li, J.; Ma, L.; Mu, K.; Tian, Y. Intrinsic relationship reasoning for small object detection. arXiv 2020, arXiv:2009.00833. [Google Scholar]
  21. Pato, L.V.; Negrinho, R.; Aguiar, P.M. Seeing without looking: Contextual rescoring of object detections for ap maximization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 14610–14618. [Google Scholar]
  22. Bai, Y.; Zhang, Y.; Ding, M.; Ghanem, B. Sod-mtgan: Small object detection via multi-task generative adversarial network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 206–221. [Google Scholar]
  23. Noh, J.; Bae, W.; Lee, W.; Seo, J.; Kim, G. Better to follow, follow to be better: Towards precise supervision of feature super-resolution for small object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 9725–9734. [Google Scholar]
  24. Wang, J.; Chen, K.; Yang, S.; Loy, C.C.; Lin, D. Region proposal by guided anchoring. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seoul, Korea, 27 October–2 November 2019; pp. 2965–2974. [Google Scholar]
  25. Sun, S.; Yin, Y.; Wang, X.; Xu, D.; Zhao, Y.; Shen, H. Multiple receptive fields and small-object-focusing weakly-supervised segmentation network for fast object detection. arXiv 2019, arXiv:1904.12619. [Google Scholar]
  26. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  27. Guo, C.; Fan, B.; Zhang, Q.; Xiang, S.; Pan, C. Augfpn: Improving multi-scale feature learning for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12595–12604. [Google Scholar]
  28. Ghiasi, G.; Lin, T.Y.; Le, Q.V. Nas-fpn: Learning scalable feature pyramid architecture for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seoul, Korea, 27 October–2 November 2019; pp. 7036–7045. [Google Scholar]
  29. Xu, H.; Yao, L.; Zhang, W.; Liang, X.; Li, Z. Auto-fpn: Automatic network architecture adaptation for object detection beyond classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 6649–6658. [Google Scholar]
  30. Hu, M.; Li, Y.; Fang, L.; Wang, S. A2-FPN: Attention aggregation based feature pyramid network for instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15343–15352. [Google Scholar]
  31. Lee, Y.; Park, J. Centermask: Real-time anchor-free instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13906–13915. [Google Scholar]
  32. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef] [Green Version]
  33. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar]
  34. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. MMDetection: Open mmlab detection toolbox and benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar]
Figure 1. Illustration of FPN, with predictions made independently at all levels. { C 2 , C 3 , C 4 , C 5 } is a bottom-up path and { P 2 , P 3 , P 4 , P 5 } is a top-down path, 1 × 1 convolution is a lateral connection.
Figure 1. Illustration of FPN, with predictions made independently at all levels. { C 2 , C 3 , C 4 , C 5 } is a bottom-up path and { P 2 , P 3 , P 4 , P 5 } is a top-down path, 1 × 1 convolution is a lateral connection.
Electronics 11 02028 g001
Figure 2. Illustration of SV-FPN. (a) Small object feature enhancement module. (b) Variance-guided RoI Fusion.
Figure 2. Illustration of SV-FPN. (a) Small object feature enhancement module. (b) Variance-guided RoI Fusion.
Electronics 11 02028 g002
Figure 3. The flowchart of SOFE.
Figure 3. The flowchart of SOFE.
Electronics 11 02028 g003
Figure 4. (a) The structure of the SOFE module. (b) The details of the eSOFE module.
Figure 4. (a) The structure of the SOFE module. (b) The details of the eSOFE module.
Electronics 11 02028 g004
Figure 5. Illustration of VRoIF. F a r e a is the result of area assignment. Var is the variance of each channel of concatenating RoI features. Weight represents the completeness of each channel of concatenating RoI features. F v a r is the weighted RoI features. F R o I is the result of VRoIF.
Figure 5. Illustration of VRoIF. F a r e a is the result of area assignment. Var is the variance of each channel of concatenating RoI features. Weight represents the completeness of each channel of concatenating RoI features. F v a r is the weighted RoI features. F R o I is the result of VRoIF.
Electronics 11 02028 g005
Figure 6. The flowchart of VRoIF.
Figure 6. The flowchart of VRoIF.
Electronics 11 02028 g006
Figure 7. Ratio of RoI features from different feature levels. Each line represents a set of proposals that should be assigned to the same feature level in FPN, i.e., proposals with similar scales. The horizontal axis denotes the ratio of features. It shows that RoIs with different sizes all exploit features from several different levels.
Figure 7. Ratio of RoI features from different feature levels. Each line represents a set of proposals that should be assigned to the same feature level in FPN, i.e., proposals with similar scales. The horizontal axis denotes the ratio of features. It shows that RoIs with different sizes all exploit features from several different levels.
Electronics 11 02028 g007
Figure 8. Accuracy analysis of the proposed SV-FPN and FPN on COCO. The blue polyline is SV-FPN, and the orange polyline is FPN.
Figure 8. Accuracy analysis of the proposed SV-FPN and FPN on COCO. The blue polyline is SV-FPN, and the orange polyline is FPN.
Electronics 11 02028 g008
Figure 9. Comparison of visual results between our model and FPN on the KITTI dataset.
Figure 9. Comparison of visual results between our model and FPN on the KITTI dataset.
Electronics 11 02028 g009
Figure 10. Comparison of visual results between our model and FPN on the PASCAL VOC 07+12 dataset.
Figure 10. Comparison of visual results between our model and FPN on the PASCAL VOC 07+12 dataset.
Electronics 11 02028 g010
Figure 11. Comparison of visual results between our model and FPN on the MS COCO 2017 dataset.
Figure 11. Comparison of visual results between our model and FPN on the MS COCO 2017 dataset.
Electronics 11 02028 g011
Table 1. The area range of RoI assignment. ‘Related to input image’ represents the RoI area in image, and ‘Related to FPN level feature maps’ represents the RoI area in feature maps.
Table 1. The area range of RoI assignment. ‘Related to input image’ represents the RoI area in image, and ‘Related to FPN level feature maps’ represents the RoI area in feature maps.
FPN LevelsRelated to Input ImageRelated to Feature Maps
P 5 A R o I 448 2 A R o I 14 2
P 4 224 2 A R o I < 448 2 14 2 A R o I < 28 2
P 3 112 2 A R o I < 224 2 14 2 A R o I < 28 2
P 2 A R o I < 112 2 A R o I < 28 2
Table 2. Statistical results of object variance. Explore the relationship between the variance of object features and mAP.
Table 2. Statistical results of object variance. Explore the relationship between the variance of object features and mAP.
NetworkFaster R-CNNMask R-CNNCascade R-CNN
BackboneResNet50ResNet101ResNet50ResNet101ResNet50ResNet101
0.3741.0160.3941.1090.3821.4440.4001.5710.4032.0790.4202.311
0.3841.7390.3981.8550.3922.6850.4082.9310.4102.8770.4253.039
Table 3. Effect of weight initialization of RoI feature fusion. Detection performance is relatively robust to weight initialization of RoI feature fusion.
Table 3. Effect of weight initialization of RoI feature fusion. Detection performance is relatively robust to weight initialization of RoI feature fusion.
mAPAP 50 AP 75 AP S AP M AP L mARAR S AR M AR L
Soft RoI Selection (average pooling)0.5260.8370.5740.2090.3890.5740.6210.3470.4900.667
Adaptive Feature Pooling (max pooling)0.5290.8360.5760.2470.3870.5780.6260.3340.4910.672
Variance-guided0.5320.8420.5850.2500.3940.5790.6260.3730.4970.670
Table 4. Effect of the value of α . The proportion of F a r e a has an impact on the detection.
Table 4. Effect of the value of α . The proportion of F a r e a has an impact on the detection.
α 00.10.20.30.40.50.60.70.80.9
mAP0.5230.5280.5260.5240.5300.5320.5300.5270.5270.526
AP 50 0.8350.8360.8410.8380.8400.8420.8430.8380.8410.838
AP 75 0.5740.5760.5680.5700.5800.5850.5770.5770.5760.575
AP S 0.2240.2610.2220.2490.2300.2500.2650.2600.2420.246
AP M 0.3870.3900.3950.3870.3860.3940.4010.3950.4000.393
AP L 0.5710.5760.5740.5700.5780.5790.5760.5740.5730.574
Table 5. Effect of each component. Results are reported on KITTI. FPN: Feature Pyramid Networks. SOFE: Small Object Feature Enhancement. VRoIF: Variance-guided RoI Fusion.
Table 5. Effect of each component. Results are reported on KITTI. FPN: Feature Pyramid Networks. SOFE: Small Object Feature Enhancement. VRoIF: Variance-guided RoI Fusion.
FPNSOFEVRoIFmAPCarVanTruckPedestrianCyclistTramMiscmAR
0.3880.8790.3100.1280.5970.5130.1940.9300.560
0.4090.8820.3070.2020.6040.5220.2400.1060.592
0.4120.8760.3110.2070.5970.5130.2770.1000.587
0.4150.8820.3100.2120.5990.5010.2840.1160.606
Table 6. Effect of each component. Results are reported on PASCAL VOC 07+12. FPN: Feature Pyramid Networks. SOFE: Small Object Feature Enhancement. VRoIF: Variance-guided RoI Fusion.
Table 6. Effect of each component. Results are reported on PASCAL VOC 07+12. FPN: Feature Pyramid Networks. SOFE: Small Object Feature Enhancement. VRoIF: Variance-guided RoI Fusion.
FPNSOFEVRoIFmAPAP 50 AP 75 AP S AP M AP L mARAR S AR M AR L
0.5230.8350.5740.2240.3870.5710.6210.3590.4900.667
0.5280.8380.5760.2440.3930.5760.6260.3410.5010.671
0.5300.8430.5770.2650.4010.5760.6260.3890.4980.670
0.5390.8410.5930.2370.3980.5870.6380.3970.5040.684
Table 7. Effect of each component. Results are reported on COCO 2017. FPN: Feature Pyramid Networks. SOFE: Small Object Feature Enhancement. VRoIF: Variance-guided RoI Fusion.
Table 7. Effect of each component. Results are reported on COCO 2017. FPN: Feature Pyramid Networks. SOFE: Small Object Feature Enhancement. VRoIF: Variance-guided RoI Fusion.
FPNSOFEVRoIFmAPAP 50 AP 75 AP S AP M AP L mARAR S AR M AR L
0.3750.5830.4060.2130.4100.4890.5140.3220.5520.652
0.3760.5840.4070.2210.4120.4850.5160.3230.5570.653
0.3770.5880.4050.2230.4120.4900.5180.3250.5600.658
0.3830.5870.4140.2170.4200.4950.5240.3290.5650.661
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, Q.; Zhang, C.; Wang, H.; He, Q.; Huo, L. SV-FPN: Small Object Feature Enhancement and Variance-Guided RoI Fusion for Feature Pyramid Networks. Electronics 2022, 11, 2028. https://doi.org/10.3390/electronics11132028

AMA Style

Yang Q, Zhang C, Wang H, He Q, Huo L. SV-FPN: Small Object Feature Enhancement and Variance-Guided RoI Fusion for Feature Pyramid Networks. Electronics. 2022; 11(13):2028. https://doi.org/10.3390/electronics11132028

Chicago/Turabian Style

Yang, Qianhui, Changlun Zhang, Hengyou Wang, Qiang He, and Lianzhi Huo. 2022. "SV-FPN: Small Object Feature Enhancement and Variance-Guided RoI Fusion for Feature Pyramid Networks" Electronics 11, no. 13: 2028. https://doi.org/10.3390/electronics11132028

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop