Next Article in Journal
Refined 3D Solar Temperature Field and Effect Simulation of Ultra-High Steel Bridge Pylon
Previous Article in Journal
Elevation Angle Characterization for LEO Satellites: First and Second Order Statistics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Edge-YOLO: Lightweight Infrared Object Detection Method Deployed on Edge Devices

School of Information Science and Engineering, East China University of Science and Technology, Shanghai 200237, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(7), 4402; https://doi.org/10.3390/app13074402
Submission received: 26 February 2023 / Revised: 20 March 2023 / Accepted: 28 March 2023 / Published: 30 March 2023
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Existing target detection algorithms for infrared road scenes are often computationally intensive and require large models, which makes them unsuitable for deployment on edge devices. In this paper, we propose a lightweight infrared target detection method, called Edge-YOLO, to address these challenges. Our approach replaces the backbone network of the YOLOv5m model with a lightweight ShuffleBlock and a strip depthwise convolutional attention module. We also applied CAU-Lite as the up-sampling operator and EX-IoU as the bounding box loss function. Our experiments demonstrate that, compared with YOLOv5m, Edge-YOLO is 70.3% less computationally intensive, 71.6% smaller in model size, and 44.4% faster in detection speed, while maintaining the same level of detection accuracy. As a result, our method is better suited for deployment on embedded platforms, making effective infrared target detection in real-world scenarios possible.

1. Introduction

Visible light images are commonly used in target detection due to their high resolution, definition, and detailed visual information that is easily interpreted by the human eye. However, these images are sensitive to external factors such as weather and lighting conditions, which can reduce image quality and negatively impact target detection accuracy. This is where infrared imaging technology plays a crucial role. By overcoming these limitations, infrared imaging allows for image acquisition under diverse lighting and weather conditions, including foggy days and nighttime scenarios. As a result, this technology has been widely adopted in various fields, such as autonomous driving, security monitoring, and remote sensing, and offers a broader range of use cases than visible images. Therefore, the importance of infrared imaging technology in enabling efficient and reliable target detection cannot be overstated.
Traditional infrared target detection techniques [1,2,3] are mainly model-based methods, such as template matching, threshold segmentation, and the Hausdorff metric. However, with the development of deep learning, target detection techniques based on convolutional neural networks have emerged in recent years. These methods are primarily divided into two-stage algorithms (e.g., Faster-RCNN [4]) and single-stage algorithms (e.g., SSD [5] and YOLO [6]). The single-stage algorithm is designed to achieve a balance between detection speed and accuracy, resulting in a significant improvement in detection speed while maintaining accuracy compared with the two-stage algorithm. Consequently, the YOLO series has become a widely used representative of the single-stage algorithm. Among the mainstream YOLO algorithms, YOLOv5 (version 6.2) has achieved significant improvement in both detection accuracy and speed compared with its predecessor by using Mosaic data enhancement, C3 modules, and an improved SPPF module. Although YOLOv5 performs well on visible images, it encounters several challenges when applied to infrared detection. The first major issue is that infrared images suffer from poor contrast, high noise, and blurred imaging, leading to the loss of crucial target features during deep convolutional network processing. Moreover, since infrared images lack color information, and the difference between target and background features is minimal, it becomes challenging for deep convolutional neural networks to distinguish useful information from irrelevant data, reducing detection accuracy. Another significant challenge is that embedded devices commonly used in autonomous driving and security monitoring fields have limited computing power, storage space, and power consumption, which makes deploying large target detection models such as YOLOv5 difficult. Additionally, these fields require real-time detection, and collecting data on edge devices and sending them to the server for detection and analysis can lead to network latency and communication congestion problems in widely distributed areas.
The challenges mentioned above necessitate lightweight infrared target detection on edge devices. In summary, this paper proposes a solution to these issues by introducing Edge-YOLO, an infrared target detection algorithm that utilizes lightweight networks and attention mechanisms specifically designed for edge devices. The primary enhancements of this algorithm can be summarized as follows:
(1)
The bounding box loss function was redesigned, and a loss function with a power hyperparameter α was made to accelerate the convergence of the loss function and solve the uncertainty of the aspect ratio in CIoU;
(2)
A lightweight content-aware up-sampling operator was adopted, which can obtain a larger perceptual field than the original nearest-neighbor up-sampling method, while only introducing a small number of parameters and computational cost;
(3)
The feature extraction network was reconstructed based on the improved ShuffleNetv2, which enhances the extraction ability of strip features in IR scenes and the perception ability of salient features in IR images by embedding a newly designed strip depthwise convolutional attention module in ShuffleBlock while significantly reducing the computational power of the network.
The remaining sections of this paper are organized as follows: Section 2 provides an overview of related works on target detection with neural networks, including YOLOv5, and other algorithms in infrared target detection. Section 3 provides a detailed introduction to the Edge-YOLO algorithm proposed in this paper. Section 4 presents the results of various experiments conducted to evaluate the performance of Edge-YOLO. Finally, Section 5 summarizes the main contributions of this paper.

2. Related Works

The YOLO family of algorithms, known for their efficiency and simplicity, was first introduced by Redmon et al. in 2015. In the years that followed, Redmon et al. released YOLOv2 and YOLOv3 algorithms, which further reduced network complexity and improved detection speed compared with two-stage algorithms [7]. After Redmon withdrew from the field of computer vision, Glenn Jocher released YOLOv5 in 2020, which has since been updated to version 6.2. YOLOv5 is composed of a backbone feature extraction module, a neck feature fusion module, and a head detection module, as illustrated in Figure 1. The algorithm incorporates five different scales of n, s, m, l, and x, with larger scales delivering higher detection accuracy but slower real-time performance. However, the network structure of the models of different scales remains consistent, differing only in the number of partial layers, and is represented uniformly as “×n” in the figure. YOLOv5 uses CSPDarknet53 as its backbone network, which includes the Cross Stage Partial (CSP) structure [8]. The CSP structure integrates gradient changes in the feature map, reducing the problem of repeating gradient information in the backbone network. Moreover, YOLOv5 utilizes a bottleneck structure with residual connections in the backbone network to prevent network degradation due to gradient disappearance, and a bottleneck structure without residual connections in the feature fusion layer to reduce computational effort. Additionally, Jocher employs a modified Spatial Pooling Pyramid Fast (SPPF) structure in place of the original SPP [9]. The modified SPPF achieves the same computational results as the original parallel MaxPool layers of three different sizes by serializing multiple MaxPool layers of the same size, significantly reducing computational time.
In addition to YOLOv5, various target detection methods have been proposed for infrared scenes by researchers. Li et al. [10] proposed the YOLO-FIRI model, an infrared image area-free target detector based on YOLOv5. They achieved good infrared target detection performance by improving the CSP structure and introducing multiple detection heads. Fan et al. [11] improved the feature extraction capability by using dense connection blocks based on YOLOv5, and improved the detection accuracy by adding a channel focus mechanism and modifying the loss function. Dai et al. [12] proposed TIRNet, which adopted VGG as the feature extractor and used a continuous information fusion strategy to obtain more accurate and smoother detection results. Li et al. [13] designed a dense nested interactive module to achieve progressive interaction among high-level and low-level features. You et al. [14] utilized multiscale mosaic data augmentation to enhance the diversity of objects and proposed a parameter-free attention mechanism to enhance features. Although these methods can be applied to IR target detection, they have some drawbacks. For instance, striped targets in IR road scenes require a more reasonable combination of striped convolution and traditional convolution to extract features. The bounding box loss function of the algorithm needs to be more accurately adapted to the boundary regression of targets in IR images, and the model needs to be more lightweight to be suitable for practical edge devices. Therefore, the method in this paper focuses on improving the above shortcomings.

3. Methods

The structure of our Edge-YOLO model is shown in Figure 2 below.
Firstly, the backbone of the model uses the improved ShuffleBlock to replace the C3 module in YOLOv5 to enhance the feature extraction capability for the characteristics of IR images of road scenes while reducing the complexity of the model.
Secondly, in the feature up-sampling structure, the original nearest-neighbor up-sampling operator is replaced by the improved CAU-Lite module.
Thirdly, although not shown in the figure, we utilize the recently proposed EX-IoU instead of CIoU as the bounding box loss function of our model. This new loss function provides better and more accurate convergence during training, thus leading to improved detection performance.

3.1. Improved Bounding Box Loss Function EX-IoU

The latest version of the YOLOv5 algorithm (version 6.2) uses the Complete-IoU (CIoU) as its bounding box loss function, as proposed by Zheng [15]. The CIoU integrates three aspects of the intersection–union ratio (IoU) between the predicted box and the ground truth box: the ratio of the distance between their centroids to the length of the diagonal of the minimum outer rectangle, and the similarity of their aspect ratios. The equation for the CIoU is as follows:
{ L C I o U = 1 I o U + ρ 2 ( b , b g t ) c 2 + α v α = v ( 1 I o U ) + v v = 4 π 2 ( arctan w g t h g t arctan w h ) 2
where α is the weight coefficient, ρ ( b , b g t ) is the distance between the center point of the predicted box and the ground truth box, c is the diagonal length of the minimum outer rectangle, v indicates the difference between the aspect ratio of the predicted box and the ground truth box, and v is 0 if they have the same aspect ratio.
The CIoU metric used in the current version of the YOLOv5 algorithm is designed to integrate three aspects of the intersection–union ratio IoU between the prediction box and the ground truth box. However, the aspect ratio used in CIoU is a relative value, which can introduce uncertainty during calculation and potentially hinder the optimization of the model. To address this issue, we propose the use of Efficient-IoU (EIoU) as the bounding box loss function, as proposed by Zhang et al. [16]. EIoU splits the aspect ratio based on CIoU and replaces the original aspect ratio difference between the predicted box and ground truth box with the ratio of the width difference between the predicted box and the ground truth box to the width of the minimum circumscribed rectangle, and the ratio of the height difference to the height of the minimum outer rectangle. This approach leads to a more accurate bounding box loss function and facilitates better model optimization. The formula is as follows:
L E I o U = 1 I o U + ρ 2 ( b , b g t ) c 2 + ρ 2 ( w , w g t ) C w 2 + ρ 2 ( h , h g t ) C h 2
where C w and C h denote the width and height of the minimum outer rectangle, respectively. The literature [17] proposes to use the hyperparameter α as a power on each term in the loss function IoU, with the following simplified formula.
L α I o U = 1 I o U α
The parameter α is crucial in emphasizing the importance of the loss and gradient of objects with high IoU, thereby enhancing the accuracy of the bounding box regression. To improve the bounding box loss function, this paper incorporates the power α into the EIoU equation, resulting in a new function known as EX-IoU. This function exponentially magnifies the importance of the IoU value, centroid distance, width difference, or height difference between any predicted box and the ground truth box, leading to an exponential reduction effect on losses and an improvement in the accuracy of the bounding box regression. The optimal value of α is determined through experiments discussed in Section 4.
L E X I o U = 1 I o U α + ρ 2 α ( b , b g t ) c 2 α + ρ 2 α ( w , w g t ) C w 2 α + ρ 2 α ( h , h g t ) C h 2 α

3.2. Content-Aware Lightweight Up-Sampling Operator CAU-Lite

Up-sampling is a crucial operation in the Feature Pyramid Network (FPN [18]), and the two commonly used methods for up-sampling are linear interpolation and deconvolution. Linear interpolation methods, such as nearest neighbor interpolation and bilinear interpolation, only take into account the sub-pixel neighborhood of the current pixel, which results in insufficient semantic information and limited receptive fields. On the other hand, the deconvolution method expands the dimensionality of the feature map through convolution, but it applies the same convolution kernel across the entire feature map, which makes it challenging to capture local changes and variations in the feature map. Furthermore, this method introduces a large number of parameters and computational overhead into the network. The CARAFE proposed by Wang et al. [19] compensates the shortcomings of the above two types of methods to some extent: CARAFE perceives and aggregates contextual information within a larger reception field, and instead of applying a fixed convolution kernel to all features, it dynamically generates adaptive up-sampling kernels, and then reorganizes the features based on the predicted up-sampling kernels. The up-sampling kernel prediction module of CARAFE changes the dimensionality of the input feature map by convolution layer to generate a feature map with the channel number of σ 2 k u p 2 , where σ indicates the up-sampling rate (generally 2) and k u p indicates the size of the up-sampling kernel (the value in this paper is 5), and then uses the pixel shuffle method to expand the channel dimension in the spatial dimension to obtain an up-sampling kernel map with the shape of σ H × σ W × k u p 2 , which contains σ H × σ W up-sampling kernels. Each position of the output feature map is then mapped back to the input feature map in the feature reassembly module. The region of k u p × k u p , the center of the map, is taken out and dotted with the predicted up-sampling kernels at that point to obtain the result. Each channel at the same spatial location on the feature map uses the same up-sampling kernel.
The analysis reveals that CARAFE uses a zero-padding strategy on the feature map edge positions in the feature reorganization stage, which leads to imperfect edge information of the up-sampled images and it is difficult to correctly upsample the target features at the edge of the feature map. Based on this, this paper proposes an improved Content-Aware Up-Sampling-Lite (CAU-Lite) method to replace the nearest neighbor up-sampling method in YOLOv5. Before finding the neighborhood of k u p × k u p , the nearest neighbor interpolation is used to upsample the input feature map, so that the spatial dimension of the feature map is the same as the spatial dimension of the up-sampled kernel map. Then, at each spatial position in the feature map, the element with the size 1 × 1 × k u p 2 × C is taken out and reshaped to the size of k u p × k u p × C . At the same time, the upper sampling core of 1 × 1 × k u p 2 is reshaped at the corresponding position in the upper sampling core map to the size of k u p × k u p × 1 . The product of each channel of the feature map with the upper sampling core is dotted to obtain a result with the size 1 × 1 × C , and the result obtained for all channels is the result of the corresponding position in the output feature map. The improved CAU-Lite structure and calculation process are shown in Figure 3.

3.3. YOLOv5 Network Model Improvement

The YOLOv5 algorithm is mainly designed for the visible light domain and is better suited for deployment on GPUs due to its high number of model parameters, large computational requirements, and large model size. However, the objective of this paper is to create lightweight target detection networks for edge-embedded devices, making it reasonable to employ a lightweight network structure instead of the heavy CSPDarknet backbone network of YOLOv5. One promising candidate for such a lightweight network model is ShuffleNetv2, proposed by Ma et al. of Megvii’s team [20], which achieves a good balance between model accuracy and running speed. By using lightweight structures such as grouped convolution and depthwise convolution, ShuffleNetv2 is optimized for computational complexity, storage access cost, and parallelism, resulting in a noticeable improvement in actual running speed. In light of this, we chose to use an improved version of ShuffleNetv2 as the backbone network structure of the Edge-YOLO algorithm. The following problems usually exist in IR road scenes:
(1)
Compared with visible light images, original infrared images lack detailed texture and feature complexity, making them less susceptible to network perception, especially in deeper and complex networks where features are lost and additional noise is introduced.
(2)
In road scenes, pedestrians and bicycles are typically present in slender strips rather than regular shapes. However, traditional convolution layers use regular convolution kernels (N × N), which may result in a loss of feature information due to their inability to adapt to target shape changes.
To enhance the extraction capability of strip-shaped features in IR scenes and the perception of salient features in IR images without increasing computational effort, this paper proposes embedding the strip depthwise convolutional attention module (SDCA) into ShuffleBlock. As shown in Figure 4, SDCA takes an input and generates a shortcut branch. The local information is aggregated by a 5 × 5 depthwise convolution, and the output is convolved by three branches: a pair of 1 × 7 and 7 × 1 depthwise strip convolutions, a pair of 1 × 11 and 11 × 1 depthwise strip convolutions, and a shortcut branch. These branches capture multi-scale contextual information and strip features, and their results are summed and passed through a 1 × 1 normal convolutional layer to model the relationship between different channels. The output of this layer is then used as the attention weights to weigh the input of SDCA by multiplying it with the generated shortcut branch.
The improved structure obtained by embedding the strip depthwise convolutional attention module into ShuffleBlock, as well as the addition of SENet [21] as the channel attention mechanism in the right branch, are shown in Figure 5. Typically, 1 × 1 convolutions are used before and after depthwise convolutions to fuse information between channels and to increase or decrease the number of channels. However, the original ShuffleBlock uses 1 × 1 convolution layers before and after the depthwise convolution layer in its right branch, which generates redundancy. To reduce the number of parameters and computational demand, this paper removes the 1 × 1 convolution layer after the depthwise convolution.

4. Experiments

4.1. Experimental Environment and Dataset

The experiments in this paper were conducted using an Intel Xeon Platinum 8255C CPU and an NVIDIA RTX 3090 GPU with CUDA version 11.7. To evaluate the detection performance of Edge-YOLO, we used the publicly available FLIR dataset, which is an infrared dataset released by FLIR in 2018. The dataset consists of more than 10,000 images classified into four categories: Person, Bicycle, Car, and Dog. However, since there are only a few Dog images in the dataset, this paper only evaluated the detection performance of Edge-YOLO for the remaining three categories.

4.2. Bounding Box Hyperparameter Study

In the improved EX-IoU bounding box loss function of this paper, there is a hyperparameter α that affects the model’s accuracy performance. To determine the optimal value of α for the Edge-YOLO algorithm, we conducted multiple training and testing experiments using different values of α. The accuracy results obtained are shown in Figure 6. From the results, it can be observed that the highest mAP value of 78.8% is achieved when the value of α is set to 3, while the mAP value of the model decreases to 76.9% when the value of α is set to 8. This indicates that the model’s detection accuracy improves by 2.47% when using the optimal value of α, and the model achieves its best performance in terms of detection performance. As a result, this paper selects 3 as the power of each term in EX-IoU to obtain the best accuracy performance.

4.3. Model Lightweighting Experiment

By replacing the backbone feature extraction network of YOLOv5 with the improved ShuffleBlock in this paper, i.e., Edge-YOLO shown in Figure 2, the overall number of parameters, computation, and model size of the algorithm model can be effectively reduced, and Table 1 below shows the comparison of each parameter after the model is lightened and improved.
The table above shows that by replacing the backbone network with the improved ShuffleBlock, Edge-YOLO reduces the number of network parameters by 72.2%, the amount of computation by 70.3%, and the model size by 71.6% compared with YOLOv5m. This demonstrates the significant lightweight effect of the proposed method, which helps to reduce the storage and computation resources required by the model and is more suitable for deployment on edge-embedded devices.

4.4. Ablation Experiments

In this part, the original ShuffleNetv2 is firstly used as the backbone network of YOLOv5m, based on which the ablation experiments of several improvement strategies proposed in this paper are conducted to better understand the effects of different improvement strategies on the detection effect in Edge-YOLO, and the results are shown in Table 2 below.
As can be seen from Table 2, compared with the first group of experiments using only the basic model, the second group of experiments with the addition of EX-IoU solves the problem of uncertainty in the aspect ratio of CIoU by improving the loss function of the bounding box and accelerating the convergence of the loss function, and the detection accuracy is improved by 1.2% from the results, while the remaining parameters remain unchanged. The third group of experiments replaces the original nearest neighbor interpolation up-sampling with the CAU-Lite up-sampling operator proposed in this paper, which senses and aggregates contextual information within a larger reception field, dynamically generates adaptive up-sampling kernels, and performs feature reorganization based on the generated up-sampling kernels. It can be seen that with CAU-Lite, the detection accuracy of the model is improved by 1.6%, but the FPS is also slightly reduced. The fourth group of experiments applies the strip depthwise convolutional attention module proposed in this paper, which replaces the original ShuffleNetv2 network structure with an improved ShuffleBlock, enhancing the feature extraction capability for strip-shaped targets and the perception of the saliency of infrared targets. As seen in the table, the detection performance of the model is significantly improved by 3.1% compared with the original model, but the number of parameters, computation, and size of the model increased due to the addition of the new module. The final fifth set of experiments uses a combination of the three improvement points proposed in this paper, and from the results, a larger performance improvement is obtained at the cost of fewer computational and storage resources compared with the original model.
Figure 7 below shows the P–R curves of each class of different improvement strategies applied to the base model and the complete Edge-YOLO. The figure shows that compared with the base model, the APs of all three target categories are improved with different improvement strategies, and the AP of the bicycle category is improved most significantly.

4.5. Comparison Experiments

To further verify the detection performance of the Edge-YOLO algorithm, this section compares Edge-YOLO with Faster R-CNN, SSD, YOLOv5m, YOLOv7 [22], and other mainstream target detection algorithms for comparison experiments, and the results are shown in Table 3 below.
From the table, we can first see that Faster R-CNN, as a two-stage algorithm, lags far behind the single-stage algorithm in detection speed, and the current detection accuracy does not have an advantage over the single-stage algorithm; while the SSD algorithm in the single-stage algorithm speeds up the detection speed compared with Faster R-CNNN, but the detection accuracy is also reduced accordingly. Both algorithms are not comparable to the current YOLO series algorithms. Second, the detection accuracy of the algorithm in this paper is basically the same compared with the YOLOv5m algorithm, but it has obvious advantages in detection speed and consumption of computational and spatial resources. Again, compared with the latest YOLOv7 algorithm, the detection accuracy of the Edge-YOLO algorithm is slightly behind that of the YOLOv7 algorithm, but the resources consumed by the YOLOv7 algorithm and its detection speed are completely inferior to that of this paper. Finally, compared with the YOLO-FIRI target detection algorithm proposed by Li et al. also for IR scenes, this paper achieves higher detection accuracy with less resource consumption and better results in real-time.

4.6. Comparison of Test Results

Figure 8 below shows the detection results of some images in the dataset under YOLO-FIRI, YOLOv5m, YOLOv7, and Edge-YOLO of this paper.
Since the Faster R-CNN and SSD in the previous subsection are lagging in detection accuracy and detection speed, only YOLO-FIRI, YOLOv5m, and YOLOv7 are used in the visualization effect comparison with the algorithms in this paper.
From the figure, it can be seen that compared with YOLO-FIRI, the target detection algorithm for infrared road scenes, the algorithm in this paper has a certain lead in accuracy and a higher confidence level in detecting targets. In addition, observing the fourth figure, we can see that the YOLO-FIRI algorithm misdetects some pedestrian legs as bicycles, which has some defects. After comparing this algorithm with the YOLOv5m algorithm and YOLOv7 algorithm, we can see that the three algorithms basically maintain the same detection results, and they can detect cars, pedestrians, and a small number of bicycles in road scenes well. Because the algorithm in this paper is a lightweight network model, it is better than the other two algorithms in terms of the number of parameters, computation, and model size, so this algorithm has a more practical application value.

4.7. Actual Edge Device Deployment Testing

This paper uses the RK3588 embedded development board of Rockchip as the verification platform, as shown in Figure 9 below. RK3588 platform is equipped with quad-core A76+ quad-core A55, an octa-core CPU, and NPU with 6TOPs computation power. Its high computation power NPU supports INT4, INT8, INT16, and FP16 mixed computing, which can accelerate the inference of network models. The photo of RK3588 is shown below.
The algorithm model in this paper and the comparison algorithm model are first exported to the compatible ONNX format, and then converted to the RKNN model supported by the NPU of the RK3588 platform using the RKNN-Toolkit2 and rknpu2 tools with inference acceleration such as asymmetric hybrid quantization, and these models are used to infer the test set images, and the performance comparison is obtained as shown in the following table. In addition to inference using NPU, the performance of only CPU inference is also tested in this paper and is shown together in Table 4 below.
As can be seen from the table, the accuracy of all four models on the RK3588 platform has a slight decrease due to model quantization. In addition, if only the ARM CPU is used for inference, the FPS of algorithms such as YOLO-FIRI is less than 1, i.e., the number of images that can be inferred is less than one per second, and the algorithm in this paper only has an FPS of 1.1, which cannot be deployed in practical application scenarios. After using NPU to accelerate, we can see that the inference speed of each algorithm was improved by tens of times. However, the FPS of YOLOv5m and YOLOv7 are only 14.5 and 8.8, respectively, which are more obvious to notice lags in real-world applications, while the algorithm in this paper can achieve 31.9 FPS, which can meet the performance requirements of practical scenarios.

5. Conclusions

The proposed method in this paper, Edge-YOLO, is a lightweight IR target detection approach that aims to ensure good performance in road scenes and is suitable for edge-embedded devices. The algorithm utilizes an optimized bounding box loss function, the improved EX-IoU, to enhance the regression accuracy of the bounding box. Moreover, to improve the up-sampling effect, the algorithm adopts the improved CAU-Lite up-sampling operator, which perceives the contextual content. Lastly, the lightweight ShuffleBlock replaces the backbone feature extraction part of the network, and the strip depthwise convolutional attention module is used to enhance the extraction capability of strip-shaped targets and other salient features present in the IR feature map for the ShuffleBlock, thus further enhancing the detection accuracy of the model. The experimental results on the FLIR dataset demonstrate that Edge-YOLO is essentially equivalent to YOLOv5m in terms of accuracy, while reducing the number of network parameters, computation, and model size by 72.2%, 70.3%, and 71.6%, respectively. Additionally, the detection speed is increased by 44.4%, making the algorithm more suitable for embedded device applications.

Author Contributions

Conceptualization, J.Y.; methodology, J.L.; software, J.L.; validation, J.L.; writing—original draft preparation, J.L.; writing—review and editing, J.L. and J.Y.; supervision, J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, C.P.; Li, H.; Wei, Y.; Xia, T.; Tang, Y.Y. A local contrast method for small infrared target detection. IEEE Trans. Geosci. Remote Sens. 2013, 52, 574–581. [Google Scholar] [CrossRef]
  2. Liu, R.; Lu, Y.; Gong, C.; Liu, Y. Infrared point target detection with improved template matching. Infrared Phys. Technol. 2012, 55, 380–387. [Google Scholar] [CrossRef]
  3. Teutsch, M.; Muller, T.; Huber, M.; Beyerer, J. Low resolution person detection with a moving thermal infrared camera by hot spot classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Columbus, OH, USA, 23–28 June 2014; pp. 209–216. [Google Scholar]
  4. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
  5. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the 14th European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  6. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Amsterdam, The Netherlands, 11–14 October 2016; pp. 779–788. [Google Scholar]
  7. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  8. Wang, C.Y.; Liao HY, M.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H. CSPNet: A new backbone that can enhance the learning capability of CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 390–391. [Google Scholar]
  9. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [Green Version]
  10. Li, S.; Li, Y.; Li, Y.; Li, M.; Xu, X. YOLO-FIRI: Improved YOLOv5 for Infrared Image Object Detection. IEEE Access 2021, 9, 141861–141875. [Google Scholar] [CrossRef]
  11. Fan, Y.; Qiu, Q.; Hou, S.; Li, Y.; Xie, J.; Qin, M.; Chu, F. Application of Improved YOLOv5 in Aerial Photographing Infrared Vehicle Detection. Electronics 2022, 11, 2344. [Google Scholar] [CrossRef]
  12. Dai, X.; Yuan, X.; Wei, X. TIRNet: Object detection in thermal infrared images for autonomous driving. Appl. Intell. 2021, 51, 1244–1261. [Google Scholar] [CrossRef]
  13. Li, B.; Xiao, C.; Wang, L.; Wang, Y.; Lin, Z.; Li, M.; An, W.; Guo, Y. Dense nested attention network for infrared small target detection. IEEE Trans. Image Process. 2022, 32, 1745–1758. [Google Scholar] [CrossRef] [PubMed]
  14. You, S.; Ji, Y.; Liu, S.; Mei, C.; Yao, X.; Feng, Y. A thermal infrared pedestrian-detection method for edge computing devices. Sensors 2022, 22, 6710. [Google Scholar] [CrossRef] [PubMed]
  15. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. Proc. AAAI Conf. Artif. Intell. 2020, 34, 12993–13000. [Google Scholar] [CrossRef]
  16. Zhang, Y.-F.; Ren, W.; Zhang, Z.; Jia, Z.; Wang, L.; Tan, T. Focal and efficient IOU loss for accurate bounding box regression. Neurocomputing 2022, 506, 146–157. [Google Scholar] [CrossRef]
  17. He, J.; Erfani, S.; Ma, X.; Bailey, J.; Chi, Y.; Hua, X.S. Alpha-IoU: A Family of Power Intersection over Union Losses for Bounding Box Regression. Adv. Neural Inf. Process. Syst. 2021, 34, 20230–20242. [Google Scholar]
  18. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  19. Wang, J.; Chen, K.; Xu, R.; Liu, Z.; Loy, C.C.; Lin, D. Carafe: Content-aware reassembly of features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3007–3016. [Google Scholar]
  20. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
  21. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  22. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
Figure 1. The network structure of YOLOv5 series (including n, s, m, l, and x, depending on the number of duplicates of module C3).
Figure 1. The network structure of YOLOv5 series (including n, s, m, l, and x, depending on the number of duplicates of module C3).
Applsci 13 04402 g001
Figure 2. The network structure of Edge-YOLO.
Figure 2. The network structure of Edge-YOLO.
Applsci 13 04402 g002
Figure 3. Structure of CAU-Lite.
Figure 3. Structure of CAU-Lite.
Applsci 13 04402 g003
Figure 4. Strip depthwise convolutional attention module.
Figure 4. Strip depthwise convolutional attention module.
Applsci 13 04402 g004
Figure 5. Improved ShuffleBlock. (a): the base unit with SDCA applied; (b): the base unit for down-sampling (2×).
Figure 5. Improved ShuffleBlock. (a): the base unit with SDCA applied; (b): the base unit for down-sampling (2×).
Applsci 13 04402 g005
Figure 6. Variation in accuracy for different values of α on the algorithm.
Figure 6. Variation in accuracy for different values of α on the algorithm.
Applsci 13 04402 g006
Figure 7. P–R curves for each class. (a): P-R curves of the class ‘Person’; (b): P-R curves of the class ‘Bicycle’; (c): P-R curves of the class ‘Car’; (d): P-R curves of all classes.
Figure 7. P–R curves for each class. (a): P-R curves of the class ‘Person’; (b): P-R curves of the class ‘Bicycle’; (c): P-R curves of the class ‘Car’; (d): P-R curves of all classes.
Applsci 13 04402 g007
Figure 8. Part of the test results.
Figure 8. Part of the test results.
Applsci 13 04402 g008
Figure 9. The RK3588 embedded development board.
Figure 9. The RK3588 embedded development board.
Applsci 13 04402 g009
Table 1. Comparison of lightweight improvement effects.
Table 1. Comparison of lightweight improvement effects.
ModelParams/MFlops/GSize/MB
YOLOv5m20.947.942.2
Edge-YOLO5.814.212.0
Table 2. Ablation experiments.
Table 2. Ablation experiments.
+EX-IoU+CAU-Lite+SDCA[email protected]/%FPSParams/MFlops/GSize/MB
73.185.35.412.911.3
74.385.15.412.911.3
74.784.05.413.111.4
76.281.55.814.112.0
78.880.75.814.212.0
Table 3. Comparison of mainstream target detection algorithms.
Table 3. Comparison of mainstream target detection algorithms.
Model[email protected]/%[email protected]/%FPSParams/MFlops/GSize/MB
PersonBicycleCar
Faster R-CNN76.650.485.770.917.666.1152.7133.5
SSD63.543.579.362.131.946.9117.696.1
YOLOv5m83.864.190.679.555.920.947.942.2
YOLOv786.467.391.481.727.937.2105.174.8
YOLO-FIRI79.552.488.673.571.47.220.415.0
Edge-YOLO83.263.090.278.880.75.814.212.0
Table 4. Testing results on the RK3588 platform.
Table 4. Testing results on the RK3588 platform.
Model[email protected]/%FPS(CPU)FPS(NPU)
YOLO-FIRI73.00.821.6
YOLOv5m79.10.514.5
YOLOv781.20.38.8
Edge-YOLO78.61.131.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, J.; Ye, J. Edge-YOLO: Lightweight Infrared Object Detection Method Deployed on Edge Devices. Appl. Sci. 2023, 13, 4402. https://doi.org/10.3390/app13074402

AMA Style

Li J, Ye J. Edge-YOLO: Lightweight Infrared Object Detection Method Deployed on Edge Devices. Applied Sciences. 2023; 13(7):4402. https://doi.org/10.3390/app13074402

Chicago/Turabian Style

Li, Junqing, and Jiongyao Ye. 2023. "Edge-YOLO: Lightweight Infrared Object Detection Method Deployed on Edge Devices" Applied Sciences 13, no. 7: 4402. https://doi.org/10.3390/app13074402

APA Style

Li, J., & Ye, J. (2023). Edge-YOLO: Lightweight Infrared Object Detection Method Deployed on Edge Devices. Applied Sciences, 13(7), 4402. https://doi.org/10.3390/app13074402

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop