Next Article in Journal
MSGL+: Fast and Reliable Model Selection-Inspired Graph Metric Learning
Previous Article in Journal
Current Source Strategy for Energy Injection from a CapMix Cell
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

YOLOv8-CGRNet: A Lightweight Object Detection Network Leveraging Context Guidance and Deep Residual Learning

School of Electronic and Information Engineering, University of Science and Technology Liaoning, Anshan 114051, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(1), 43; https://doi.org/10.3390/electronics13010043
Submission received: 16 November 2023 / Revised: 12 December 2023 / Accepted: 18 December 2023 / Published: 20 December 2023

Abstract

:
The growing need for effective object detection models on mobile devices makes it essential to design models that are both accurate and have fewer parameters. In this paper, we introduce a YOLOv8 Res2Net Extended Network (YOLOv8-CGRNet) approach that achieves enhanced precision under standards suitable for lightweight mobile devices. Firstly, we merge YOLOv8 with the Context GuidedNet (CGNet) and Residual Network with multiple branches (Res2Net) structures, augmenting the model’s ability to learn deep Res2Net features without adding to its complexity or computational demands. CGNet effectively captures local features and contextual surroundings, utilizing spatial dependencies and context information to improve accuracy. By reducing the number of parameters and saving on memory usage, it adheres to a ‘deep yet slim’ principle, lessening channel numbers between stages. Secondly, we explore an improved pyramid network (FPN) combination and employ the Stage Partial Spatial Pyramid Pooling Fast (SimPPFCSPC) structure to further strengthen the network’s capability in processing the FPN. Using a dynamic non-monotonic focusing mechanism (FM) gradient gain distribution strategy based on Wise-IoU (WIoU) in an anchor-free context, this method effectively manages low-quality examples. It enhances the overall performance of the detector. Thirdly, we introduce Unifying Object Detection Heads with Attention, adapting to various input scenarios and increasing the model’s flexibility. Experimental datasets include the commonly used detection datasets: VOC2007, VOC2012, and VisDrone. The experimental results demonstrate a 4.3% improvement in detection performance by the proposed framework, affirming superior performance over the original YOLOv8 model in terms of accuracy and robustness and providing insights for future practical applications.

1. Introduction

With the evolution of computer technology and the widespread application of principles of computer vision, research on target detection and tracking using computer image processing technology is gaining popularity. Object detection [1] plays a pivotal role in applications such as autonomous driving and unmanned vehicles [2], security and surveillance [3], medical imaging [4], robotics [5], and agriculture [6], with image segmentation [7] and object tracking [8]. Pedestrian re-identification [9] often relies on it. Object detection typically involves two primary steps: locating an object’s position and classifying its type. Early methods employed the sliding window technique, sliding windows of various sizes across the image and running a classifier at each window position to identify targets. With technological advancements, region-based Convolutional Neural Networks (R-CNN) [10] gained popularity, initially using selective search to extract candidate regions and then classifying these regions using CNNs. The introduction of a faster r-cnn [11] containing a region suggestion network (RPN) [11], which automatically suggests areas of an image that may contain objects, can improve efficiency. However, YOLO [12] further streamlined the process by treating object detection as a regression problem, predicting bounding boxes and class scores for all categories simultaneously, thereby circumventing multiple forward propagations. These advancements have facilitated faster and more accurate real-world applications of object detection.
Object detection methods can be combined with evaluation algorithms. Deep architectures with a region proposal network (DeepRPN-BIQA) [13] proposed a deep architecture incorporating a region proposal network (RPN) for blind image quality assessment (BIQA) of natural-scene and screen-content images. The RPN extracts important regions that affect image quality by computing visual saliency. These regions are then fed into a Convolutional Neural Network (CNN) to predict the quality score. Object detection can be enhanced by integrating it with the DeepRPN-BIQA approach proposed in this paper, thereby improving the performance and efficiency of object detection. For instance, DeepRPN-BIQA could be used as a preprocessing step to perform quality assessment on the input image. Based on the quality score, the best image can be selected, or image enhancement can be performed before feeding it into the object detection model. Alternatively, DeepRPN-BIQA could serve as a postprocessing step to assess the quality of the output results from the object detection model. The best detection results can be selected, or result optimization can be performed based on the quality score.
Object detection can be integrated with other algorithms. In the fields of pedestrian re-recognition [14] and computer vision in medical image analysis [15], object detection is used in smart video surveillance systems to provide deeper insights and automated responses. In robotics, object detection is combined with path planning [16] and obstacle avoidance algorithms [17], aiding robots in better navigation and interaction with their environment. In multimodal learning systems, object detection can be integrated with NLP [18] (natural language processing) technologies to process complex data containing both visual and textual information, such as extracting information from social media posts.
Despite significant progress in identifying and locating objects in images, challenges remain concerning computational resources, small object detection, and real-time requirements. The rapid increase in demand for object detection models on mobile devices faces the limitations of processors and GPUs, which are generally less powerful than those on desktops or servers. This limitation affects the complexity of models that can run on devices in real time. Moreover, the typically lower RAM on mobile devices implies stringent constraints on model size and runtime memory usage. Thus, designing models that are both memory-efficient and highly accurate is imperative and challenging. Although improved versions like YOLOv8 provide higher accuracy and detection outcomes, considering the model’s generalization and adaptability to ever-changing application scenarios and dataset characteristics remains crucial for ensuring stability and efficacy under diverse conditions.
Inspired by these considerations, this research introduces a framework for object detection, the YOLOv8 Res2Net Extended Network (YOLOv8-CGRNet). YOLOv8 represents a popular model in the object detection domain, exemplifying one of the latest advancements in the YOLO [12] series, showcasing performance on par or better than other models in public datasets while maintaining rapid inference speed. In its latest iteration, an attempt to amalgamate the backbone network of the YOLO series with the Residual Network with multiple branches (Res2Net) [19] structure has been made, introducing the learning capability of ResNet’s deeper features for object detection tasks, thereby enhancing the model’s recognition of complex scenes and multi-scale objects. The integration of the downsampling module from Context GuidedNet (CGNet) [20] has fortified the learning of local contextual information, offering improved performance in detail and context understanding.
YOLOv8-CGRNet retains the downsampling module from CGNet while incorporating efficient modules like Stage Partial Spatial Pyramid Pooling Fast (SimPPFCSPC) [21] and Wise-IoU (WIoU) [22]. It merges various modules such as skip-path, Fused Convolution, and CIOU loss to optimize model performance and enhance recognition, further refining the model’s feature extraction and representation capabilities. With CGblock downsampling and deep Res2block structures alternating within the network, they work collaboratively, ensuring that while deep features are extracted, context information is effectively preserved and utilized, thereby enhancing performance on object detection tasks.
The experimental datasets comprise two widely utilized object detection sets, Pascal VOC2007 [23], VOC2012 [24], and VisDrone [25], provided by the Visual Geometry Group (VGG) of the University of Oxford, including 20 object categories encompassing everyday items from vehicles to domestic goods, animals to humans. VOC2007 contains approximately 9600 images with about 27,000 object annotations. VOC2012 includes around 11,500 images with roughly 35,000 object annotations. Pascal VOC sets the standard evaluation metrics for various tasks, employing mean Average Precision at IoU threshold 0.5 (mAP50) [26] for object detection assessment. The contributions of this research are outlined as follows: First, an exploration of the combination of the YOLOv8 series with Context GuidedNet and the Res2Net architecture has been conducted, which leverages the depth feature learning capabilities of Res2Net. This fusion facilitates enhanced object detection against complex backgrounds and a wide range of object scales, delivering superior recognition capabilities. Second, the incorporation of the SimPPFCSPC structure further augments the network’s proficiency in handling the Feature Pyramid Network (FPN) [27]. This model employs FPN and integrates contextually enhanced feature extraction techniques with adaptive strategies for depth and width tailored to varying computational and performance demands. Third, this network amalgamates the advantages of various modules and, notably, introduces the novel Detect DyHead (dynamic head) [28], which possesses the aptitude to adapt to different input scenarios, significantly increasing the model’s versatility. To further advance detection effectiveness, a clever mechanism utilizing dynamic non-monotonic focusing mechanism (FM) [22] gradient boosting allocation has been adopted. This mechanism effectively processes low-quality examples and elevates the overall performance of the detector.
As demonstrated in Figure 1, through a series of innovations and enhancements, the proposed model manifests substantial improvements over the baseline YOLOv8, especially in intricate scenarios and multi-scale object detection tasks. These modifications not only significantly boost the model’s performance but also manage to maintain a balance between the number of parameters and inference speed. Testing across multiple datasets has shown that, in comparison to the original YOLOv8, our model achieves an approximate 4% increase in accuracy.
The remainder of this document is organized as follows: Section 2 delineates the methodology, Section 3 discusses the experimental results, and Section 4 concludes by summarizing the findings and suggesting potential avenues for future research. Section 5 deliberates on the model’s strengths, weaknesses, and application discussions.

2. Materials and Methods

The framework utilized in this study is depicted in Figure 2. A noted limitation of lightweight models is the potential compromise between model accuracy and generalization capability. Hence, it is challenging to strike an appropriate balance between performance, speed, and accuracy when designing and selecting lightweight models.
As shown in Figure 2, the model employs a Feature Pyramid Network (FPN) [27]. Initially, the integration of a Context-Guided Block (CG) is undertaken: The Context GuidedNet (CGNet) [20] efficiently captures local features and the surrounding context, leveraging spatial dependencies and contextual information to enhance accuracy. It reduces the number of parameters and saves on memory usage, adhering to the “deep and thin” principle to diminish channel count across stages. Secondly, the fusion of a Res2Net block is implemented, constructing hierarchical-like residual connections within a single residual block. This approach augments multi-scale representation at a finer granularity and expands the receptive field for each network layer. Thirdly, the employment of a SimPPFCSPC structure further amplifies the network’s capacity to process the feature pyramid effectively. Fourthly, the integration of a dynamic attention mechanism from DyHead, which encompasses scale awareness, spatial awareness, and task awareness, is reported. By dynamically combining information from multiple scales, the scale-aware attention mechanism enhances the ability to recognize things of various sizes. In addition to attending to attention at each spatial location, the spatial-awareness attention module adaptively combines various feature levels in order to acquire a representation that is more discriminative. In order to adaptably handle a variety of tasks, including classification, box regression, and center-keypoint learning, task awareness distributes attention across many channels.

2.1. Overall Structure

As depicted in Figure 3, the network’s head starts with a Conv layer for initial feature extraction. The Conv part includes a 2D convolutional layer, 2D batch normalization, and a Sigmoid activation function. The CGD Block and Res2Block modules then work in concert to capture and enhance multi-scale features. The integration of SimPPFCSPC is finalized to achieve enhanced performance with fewer parameters. At the detection head, a decoupled head structure prevalent in current research is utilized, separating classification and detection tasks using an anchor-free approach. Initially, the dynamic head (DyHead) is employed for unifying scale, spatial, and task awareness within the detection head. Subsequently, the features extracted from the DyHead are refined and augmented through successive convolutional operations. After feature extraction by the Conv layer, a two-dimensional convolution operation generates an output of appropriate size. The difference between the predicted bounding boxes is quantified by the Bounding Box Loss, and the discrepancy between the predicted class probabilities and the actual class labels is measured by the Class Loss.
FPN serves as a feature fusion mechanism that enhances the detection of multi-scale targets with minimal computational cost. The main issue addressed by the FPN is the need for more handling of multi-scale variations in object detection. The downsampling ratios of the CGD blocks, numbered (1, 3, 5, and 7), are {4, 8, 16, and 32}, respectively, with each subsequent layer outputting a smaller feature map than the preceding one. These feature maps are employed within the FPN to construct a richer multi-scale feature representation. The top-down pathway begins at the last (and most profound) layer of the network, where feature maps are upsampled to increase resolution. Each upsampled feature map is laterally connected with the corresponding resolution feature map from the bottom-up pathway to merge high-level semantic information with low-level detail information. This process continues until it reaches the original image resolution. Anchor-free object detection provides an efficient alternative to tackle the complexity and limitations associated with traditional anchor-based methods by predicting key points or bounding boxes of objects directly on feature maps, simplifying the detection process while enhancing flexibility and efficiency. Center-based methods identify the center point and dimensions of each object. This framework predicts the distances to the left, top, right, and bottom edges of the target box from its center point.

2.2. Main Blocks

As shown in Figure 4, it is mainly divided into three blocks. First, the CGDBlock learns the joint features of local features and surrounding context and further improves the joint features with global context while maintaining low memory usage and improving accuracy. Secondly, Res2block can effectively utilize the hierarchical residual connection within a residual block to improve the multi-scale feature representation ability and thus improve the performance of various visual tasks. Finally, SimPPFCSPC is used to extract and process features at different scales to adapt to various tasks and data.
Res2block [19] processes detection in a multi-scale manner, which helps to extract global and local information. To better integrate information from different scales, we split the input features into 4 features and passed them through a 1 × 1 convolution. Each subset is processed by a 3 × 3 convolution group to obtain the output feature map. To reduce the number of parameters, the first split convolution is omitted, which can also be viewed as a form of feature reuse. A hierarchical residual connection is established between each convolution group, y i = c i ( y i - 1 ) + y i , y 0 = x 1 . Finally, all the output feature maps are concatenated together to form the final output feature map.
The CGDblock for downsampling and the simulation of spatial dependencies and semantic context information are improved by CGNet [20]. It consists of four sub-modules: local feature extractor f l o c (), surrounding context extractor fsur(), joint feature extractor f j o i (), and global context extractor f g l o (). It first learns the joint features of the local features and surrounding context and then uses the global context for the channel-level weighting of joint features. It also adopts residual learning to enhance information flow.
Downsampling: the conv1x1 layer initially reduces the spatial dimension of the input by half and adjusts the number of channels.
Feature integration: Local (F_loc) and surrounding (F_sur) features are connected and further processed to effectively merge these features. Then, the spatial dimension of the input is reduced by half through downsampling; the conv1x1 layer and the number of channels are adjusted. Finally, the F_glo layer is used to refine these combined features.
f j o i = f j o i ( f l o c , f s u r ) = B N ( P R e L U ( [ f l o c , f s u r ] ) )
Herein, f l o c and f s u r , respectively, represent local features and surrounding context features, and f j o i  represents joint features. [ f l o c , f s u r ] represents the connection operation between local features and surrounding context features. PReLU represents the parameterized linear rectification unit, and BN represents batch normalization.
f g l o = f g l o ( f j o i ) = F C ( F C ( G A P ( f j o i ) ) )
This formula describes how to obtain global context features from joint features. Here, f j o i represents joint features, and f g l o represents global context features. GAP represents global average pooling, and FC represents a fully connected layer.
f o u t = f g l o f j o i
This formula describes how to weigh the joint features at the channel level with global context features to obtain the output features. Here, f g l o  represents global context features,  f j o i represents joint features, and f o u t represents output features. represents element-wise multiplication. This design reflects an intention to balance detailed local feature extraction with more global contextual information, which is usually beneficial for object detection and for enhancing spatial hierarchy structures in visual deep learning models.
In SimPPFCSPC, different input features are first extracted through conv1, conv3, and conv4. Then, dimensionality reduction is performed through the max-pooling layer. Next, the dimensionality-reduced features are further processed by conv5 and conv6 to extract features. Meanwhile, the original input is also processed by conv2 to extract input features. Finally, the outputs of these two parts are concatenated together and mapped to the target space through conv7. Speed improvement is achieved while maintaining the same receptive field.

2.3. FusionDetect–ModuleHead

The design principle of the DyHead [28] is to combine multiple attention mechanisms, each of which focuses on a different dimension: scale, space, and task. The detailed design and working principles of scale-aware attention, spatial-aware attention, and task-aware attention.
By cooperatively combining multiple self-attention mechanisms between feature levels, spatial positions, and output channels, the proposed method significantly improves the representation ability of the object detection head without increasing any computational overhead. Formula (4) represents the general form of applying the self-attention mechanism to the feature tensor, where π(·) is an attention function.
W ( F ) = π ( F ) F
Formula (5) decomposes the attention function into three sequentially applied attention functions, each acting on different dimensions of the feature tensor.
W ( F ) = π C ( π S ( π L ( F ) ) F ) F F
Formula (6) represents the scale-aware attention module, which uses a 1 × 1 convolutional layer to learn the relative importance of different level features and normalizes the weights using a hard Sigmoid function.
π L ( F ) F = σ ( f ( S , C F ) ) F
Formula (7) represents the spatial-aware attention module, which uses deformable convolution to sparsely sample and aggregate features at different levels and spatial positions and adjusts the sampling position and importance using self-learned offsets and weights.
π S ( F ) F = 1 L l = 1 L k = 1 K w l , k F ( l ; p k + Δ p k ; c ) Δ m k
Formula (8) represents the task-aware attention module, which uses a dynamic ReLU function to activate features on different channels and learns the activation threshold using two fully connected layers and a normalization layer.
π C ( F ) F = max ( α 1 ( F ) F c + β 1 ( F ) , α 2 ( F ) F c + β 2 ( F ) )
The meanings of the symbols are as follows:
F is the input feature tensor with dimensions L × S × C, where L represents the number of levels, S represents the number of spatial positions, and C represents the number of channels.
W is the output feature tensor with the same dimensions as F .
π is the attention function, which can have different forms.
σ is the hard Sigmoid function, defined as σ(x) = max(0, min(1, (x + 1)/2)).
f is a linear function implemented by a 1 × 1 convolutional layer.
w l , k are the weight parameters of the deformable convolution learned from the input features.
p k is the original sampling position, and Δ p k is the self-learned offset.
F (l; p k   +   Δ p k ; c) represents the feature value at the position p k   +   Δ p k on the l-th level and the c-th channel.
Δ m k is the self-learned importance scalar, learned from the input features.
α 1 , α 2 , β 1 , and β 2 are the parameters of the dynamic ReLU function, learned from the super function θ(·). θ(·) consists of a global average pooling layer, two fully connected layers, and a normalization layer, and it applies a shifted Sigmoid function to normalize the output to the interval [−1, 1].
Fc represents the feature slice on the c-th channel.
The proposed method solves the problem of improving the performance of various object detection heads without providing a unified perspective, which was attempted by previous works. By cooperatively combining multiple self-attention mechanisms between feature levels, spatial positions, and output channels, the representation ability of the object detection head is significantly improved. Although this method enhances the representation ability of the object detection head, it does not bring a large computational overhead.

3. Experiments

In this section, we demonstrate the superiority of YOLOv8-CGRNet by evaluating its effectiveness in recognizing voc2012 and voc2007 and show the improvement of each step in recognition performance through ablation experiments.

3.1. Introduction to the Dataset

As shown in Table 1, the research utilized the VOC2007 and VOC2012 datasets, released by the PASCAL (Pattern Analysis, Statistical Modeling, and Computational Learning) Network Organisation, funded by the European Union. The VOC2007 dataset comprises 9963 images taken in diverse environments, including both indoor and outdoor scenes. Annotations for the 24,640 objects depicted within these images are provided, with each object delineated by a bounding box. The dataset encompasses 20 categories, such as persons, animals, vehicles, and household furniture. The VOC2012 dataset includes 17,125 images, with a total of 27,450 annotated objects, and features similar environmental conditions and object annotations as VOC2007.
VisDrone was collected by the AISKYEYE team at the Lab of Machine Learning and Data Mining, Tianjin University, China. It includes 288 video clips formed by 261,908 frames and 10,209 static images. The data were captured using various drone-mounted cameras. Each image has a corresponding annotation file that contains the location, category, and occlusion degree of each object.

3.2. Detailed Implementation

The full VOC2012 dataset, alongside the VOC2007 training and validation subsets, was employed for training, with evaluations conducted on the VOC2007 test set. The model was configured for mobile application compatibility, with depth and width multipliers set at 0.33 and 0.25, respectively, and a channel limit of 1024. A consistent training regime of 200 epochs was maintained across all experiments. Ablation studies indicated that, while certain frameworks may enhance efficiency individually, their performance can diminish when integrated. Various fusion methods were trialed before establishing a synergy of YOLOv8 with Context GuidedNet, Res2Net, SimPPFCSPC, DyHead, and WIoU structures that yielded improved compatibility. An NVIDIA 3080 GPU with 10 GB of memory facilitated the computational process.
The first step entailed configuring the baseline model, YOLOv8n, optimized for mobile devices, and training it on the combined VOC datasets for 200 epochs. Evaluating this model on the VOC2007 test set provided a benchmark for subsequent refinements. In the second step, Res2Net structure integration augmented the baseline model’s expressive capacity for feature representation. This modified model underwent identical training processes, validating Res2Net’s contributions through performance comparisons pre- and post-integration. Upon confirming Res2Net’s effectiveness, the third step incorporated the SimPPFCSPC structure to enhance complex feature learning capabilities. The fourth step involved integrating Context GuidedNet and DyHead structures to bolster contextual comprehension and detail resolution in target detection, respectively. The inclusion of the WIoU structure aimed to refine the model’s locational accuracy, culminating in a comprehensive model fusion. Each integration step’s impact was meticulously documented through key performance metrics such as detection precision, model size, and operational speed, ensuring a thorough multidimensional performance evaluation. The experimental methodology prioritized repeatability and the stability of structural combinations. This progression of experimental steps aimed to provide researchers with a definitive framework for assessing the specific impact of various architectures on model performance. Comparisons were also drawn with other models, including YOLOv5 and YOLOv6 variants scaled to equivalent depth and width, with a channel limit of 1024.

3.3. Experimental Results

In the PASCAL VOC dataset, object categories exhibit distinct shapes, sizes, and contextual variances. For instance, the “cat” category tends to occupy larger image areas with rich texture information, facilitating effective learning and recognition by the model. Conversely, objects such as “bottles” and “plants” are typically smaller and more challenging to discern against complex backgrounds, highlighting the significance of structural enhancements for such categories.
As demonstrated by Table 2, Fast R-CNN reported an overall mean Average Precision at the IoU threshold of 0.5 (mAP50) [26] of 68.4%, with exceptional performance in the “aeroplane” category at 82.3%. Faster R-CNN achieved a higher overall mAP50 of 70.4%, with an “aeroplane” category precision of 84.9%. YOLO models registered lower performance in detecting “bottle” category objects, with YOLOv3-tiny achieving a notable mAP of 72.3% and displaying superior performance in the “motorbike” and “train” categories. The YOLOv5n and YOLOv6n iterations exhibited consistent improvements across all categories, with mAP50s of 76.5% and 78.7%, respectively.
The YOLOv8n’s enhanced understanding of small objects and large scene contexts translated to over 70% mAP50 across multiple categories, attesting to its robustness as a general-purpose object detection model. The performance in categories such as “aeroplane”, “bicycle”, “boat”, and “bottle” was particularly notable, likely due to the optimized deep learning architecture and effective training methodologies.
Through multi-scale feature extraction, the Res2Net structure provided the model with refined feature representations, as evidenced by improved performance in the “vehicle” and “aeroplane” categories. This scalability and the hierarchical connection approach of Res2Net may also facilitate better local detail capture in object detection, benefiting categories with intricate details such as “cats” and “dogs”.
The introduction of SimPPFCSPC, with its enhanced feature pyramid and pooling strategy, supplied the model with richer scale information.
Figure 5 illustrates that the numerals along the main diagonal (extending from the upper left to the lower right corner) represent accurate predictions, signifying instances where the model has correctly identified each category as such. Conversely, figures situated off the main diagonal denote incorrect predictions where the model has erroneously assigned a category to another.
The diagonal values within the confusion matrices of each model, as depicted in Figure 5, correspond to the count of instances accurately classified. Higher values on this diagonal indicate superior performance. By comparing these figures, one may determine which model performs optimally for specific categories. Values off the diagonal reflect the number of predictive errors a model makes within a particular category. A model exhibiting lower numbers on the diagonal is deemed to yield more precise predictions for that specific category. Within these confusion matrices, several observations can be made as follows: Firstly, the yolo-tiny has an overall lower prediction count. Secondly, in distinguishing between ‘chair’ and ‘sofa’, the yolov8n-Simppfcspc-Res2 emerges as the most accurate. Thirdly, YOLOv8-CGRNet demonstrates a more balanced performance distribution across various categories, rendering it robust for overall effectiveness. Furthermore, YOLOv8-CGRNet registers minimal misclassifications in the ‘dog’ and ‘cat’ categories and exhibits no misclassifications for ‘train’ and ‘aeroplane’.
As depicted in Figure 6, the model’s detection accuracy is evident through its overall and category-specific F1 score performance. Within each subplot, lines of different colors represent various categories, such as “person”, “car”, “dog”, and others. These curves illustrate the balance between the probability of correct predictions and the probability of false predictions at various confidence thresholds, which serves to assess model performance.
Upon analysis of the F1 confidence curves, it is observed that the majority of the models reach their peak between the confidence thresholds of 0.3 and 0.6. This peak may represent the optimal balance point between accuracy in detection and the reduction of false positives. Each model demonstrates significant variation in detection accuracy across different categories. Lightweight models, specifically YOLOv3-tiny and YOLOv5n, score relatively lower on the overall F1 score, highlighting the trade-off between speed and accuracy inherent in lightweight models. YOLOv8-CGRNet, in particular, showcases that each ablation study exhibits higher F1 scores, particularly in the medium confidence threshold range. It suggests that structural improvements can significantly enhance model detection performance.
Table 3 illustrates the progression of mean Average Precision (mAP50) for different models with the increase in training epochs. mAP50, which employs an Intersection over Union (IoU) threshold of 0.5 to calculate precision, shows a rapid ascent in the initial epochs for all model variants, indicating an expedited learning process and improvement in target detection accuracy. Variance is observed in the rate at which different model variants converge. During the early training stages, the performance enhancements for these variants are closely matched. Over time, however, models with more complex structures may exhibit superior performance. As the number of epochs rises, the rate of performance improvement for all models begins to plateau, suggesting a stabilization in learning efficacy. The YOLOv8-CGRNet model’s performance in later epochs exceeds that of other models, which substantiates the assertion that YOLOv8-CGRNet’s strategic enhancements indeed bolster the model’s generalization abilities in complex scenarios.
Table 4 showcases the performance of various YOLOv6n and YOLOv8n model versions and variants in object detection on the VisDrone2019-DET dataset across different categories. These models incorporate several technological enhancements like Res2Net, CGNet, SimPPFCSPC, DyHead, and WIoU, contributing to their performance. A general trend of improvement is observed from YOLOv6n to YOLOv8-CGRNet in overall Average Precision across all categories, with YOLOv8-CGRNet performing the best at 29.9%. In specific categories, all models tend to perform relatively well in detecting cars and trucks while showing lower effectiveness in categories like bicycles and people. The introduction of technologies like Res2Net and CGNet generally correlates with performance enhancement, especially in complex or dynamic scenes. The combination of multiple technologies, as seen in models like YOLOv8n-R-S-C-D, often leads to further improvements. Compared to YOLOv6n, the YOLOv8n series demonstrates superior performance in most categories, likely due to differences in architecture and training methodologies. This data illustrates that by employing various architectural enhancements and combined techniques, the YOLO models significantly enhance their ability to detect objects in complex drone imagery, particularly in common categories such as vehicles and pedestrians.
The combined preprocess and postprocess time on the RTX 3080 is only 2 ms. In contrast, the combined preprocess and postprocess time on the ARM Cortex-A57 is significantly higher at 24 ms.
The data in Table 5 and Table 6 display tests conducted on the NVIDIA GeForce RTX 3080, showing that CGR-Net experiences negligible differences in terms of time, influenced by preprocess and postprocess times.
The data from Table 7 and Table 8, showcasing tests on the ARM Cortex-A57 CPU, a proxy for mobile device performance, illustrate CGR-Net’s operational feasibility in a mobile environment. Although there is a noticeable drop in FPS and an increase in inference time compared to high-end GPUs, CGR-Net still manages to function, suggesting its adaptability to less powerful hardware.

4. Discussion

According to the data in Table 2, the YOLOv8 series of models surpasses other models in overall mAP50 performance on the PASCAL VOC2007 test set, particularly after the integration of Res2Net (R), Context GuidedNet (C), SimPPFCSPC (S), DyHead (D), and WIoU (W). YOLOv8-CGRNet, with its ensemble of sophisticated network modules and mechanisms, exhibits increased precision and robustness.
The framework, however, is not without drawbacks. First, higher accuracy often comes with an increase in computational costs, as indicated in Table 5, Table 6, Table 7 and Table 8. Balancing accuracy with computational expense remains a challenge in resource-constrained application scenarios. Second, certain categories, like ‘bottle’ or ‘plant’, show lower detection accuracy, which suggests that deficiencies remain in detecting small or irregularly shaped objects. Third, the incorporation of new technologies significantly heightens model complexity, which leads to prolonged training durations and heightened computational resource demands.
Based on the experimental comparisons shown in Table 1, Table 2, Table 3 and Table 4, this study explores various application scenarios, demonstrating how the adapted models perform across different datasets. It is observed that while some models may excel in one dataset, their effectiveness diminishes when applied to another. Our integrated model, although increasing computational demands, enhances adaptability and accuracy across diverse datasets. This feature is particularly advantageous for mobile deployment, where processing capabilities are limited.
The augmented model maintains efficacy on mobile platforms, as the camera capture rate limits render the performance of the improved model nearly equivalent to the original on standard computers or compute-capable boards. However, on lower-powered ARM chips, while the speed may decrease, the accuracy significantly improves, making it suitable for scenarios where precision is paramount.
This flexibility is crucial in a range of applications. For instance, in autonomous vehicle systems where hardware limitations constrain processing power and speed, the model’s high-precision object detection capabilities are vital for safety and reliability. It accelerates the response to road conditions, obstacles, and traffic signs, promoting safer navigation and decision making. Additionally, the model’s efficiency makes it ideal for real-time detection applications such as Augmented Reality (AR), autonomous drone navigation, surveillance, access control, and interactive marketing.
In summary, our fusion model balances increased computational demands with improved adaptability and accuracy, making it a versatile solution for both power-constrained mobile devices and higher-capability computing platforms, catering to a wide range of applications that require varying levels of accuracy and processing speed.

5. Conclusions

The advent of deep learning and computer vision has ushered in a new era of analytical capabilities, with model optimization and improvement taking center stage. This progression necessitates a deeper exploration of algorithmic architectures and parameter optimization, particularly in the context of object detection models on mobile devices. The demand for models that balance a small memory footprint with high accuracy is rapidly increasing, presenting a unique set of challenges given the parameter-heavy nature of current leading-edge networks, which are ill-suited for mobile environments.
In this paper, we have proposed the YOLOv8-CGRNet method, which represents a significant step forward in this domain. Our approach synergizes YOLOv8 with the CGNet and the Res2Net structure, enhancing the model’s ability to learn deep features from Res2Net. This integration provides a multi-scale representation at a more granular level without adding to the model’s complexity or computational demands. The CGNet is particularly adept at capturing local features and contextual information, leveraging spatial dependencies to bolster accuracy.
Furthermore, we have delved into an improved pyramid network combination utilizing the SimPPFCSPC structure, which augments the network’s proficiency in managing FPN. The innovative application of a dynamic, non-monotonic FM gradient gain distribution strategy, which operates on an anchor-free basis, effectively addresses the challenge of low-quality samples, thereby enhancing the detector’s overall efficacy.
This head network amalgamates the strengths of various modules, with a particular emphasis on our Unifying Object Detection Heads with Attentions. This module is designed to be versatile across a range of input scenarios, improving the model’s adaptability. By integrating multiple self-attention mechanisms in a coordinated fashion across feature levels, spatial positions, and output channels, we have succeeded in elevating the representational power of the object detection heads without incurring a substantial computational burden.
Our experimental evaluation of the VOC2007 and VOC2012 datasets has provided a robust model for demonstrating the efficacy of YOLOv8-CGRNet. The results have been promising, showcasing the model’s capability to achieve top-tier performance on these well-established benchmarks. However, it is important to acknowledge the potential limitations and avenues for future research. While our model excels in memory efficiency and accuracy, the quest for an even smaller model footprint without compromising performance continues. Additionally, the real-time application of YOLOv8-CGRNet in diverse environments remains an area ripe for exploration.
In conclusion, YOLOv8-CGRNet stands as a testament to the potential of innovative model design in the realm of mobile object detection. It paves the way for future research aimed at refining and deploying lightweight, high-performance models across a spectrum of real-world applications.
Future optimizations will focus on the following aspects: First, the exploration of new lightweight model designs that do not compromise applicability in edge computing and mobile devices. Second, the development of models capable of operating across multiple domains and the enhancement of their adaptability to different data distributions. Third, the employment of integrated learning approaches to enhance model robustness and more research on dynamic adaptive networks that allow the model to adjust its structure dynamically based on different inputs.
Performance can also be improved by expanding the training dataset. Recognizing that a more diverse dataset, such as Pascal 2011, which includes the “Pascal Visual Object Classes (VOC) Challenge”, can significantly improve the model’s robustness. Moreover, integrating various databases, such as ImageNet and COCO (Common Objects in Context), will further refine our model’s detection capabilities. This expansion not only provides young researchers with the opportunity to delve deeper into a broader range of data processing and analysis techniques but also ensures the adaptability of the model in different real-world scenarios. Simultaneously, we can explore integration with assessment-type algorithms like DeepRPN-BIQA [13] or apply the structure to segmentation algorithms similar to SegR-Net [31]. On the other hand, future iterations of YOLOv8-CGRNet can benefit from exploring advanced machine learning techniques and cross-disciplinary applications, particularly in medical imaging, where increased accuracy and precision are crucial. Therefore, future work on our model is not only aimed at enhancing its performance but also at expanding its application spectrum in the fields of computer vision and medical image analysis.

Author Contributions

Methodology, Y.N. and W.C.; Software, Y.N.; Validation, C.S.; Writing—original draft, Y.N.; Writing—review & editing, S.F.; Supervision, S.F. All authors have read and agreed to the published version of the manuscript.

Funding

This project is funded by Liaoning University of Science and Technology, under the 2023 Graduate Education Reform, Technological Innovation, and Entrepreneurship Project of Liaoning University of Science and Technology, with the funding number LKDYC202208.

Data Availability Statement

The data presented in this study are available in the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cheng, G.; Yuan, X.; Yao, X.; Yan, K.; Zeng, Q.; Xie, X.; Han, J. Towards large-scale small object detection: Survey and benchmarks. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 23821153. [Google Scholar] [CrossRef] [PubMed]
  2. Wang, H.; Xu, Y.; Wang, Z.; Cai, Y.; Chen, L.; Li, Y. Centernet-auto: A multi-object visual detection algorithm for autonomous driving scenes based on improved centernet. IEEE Trans. Emerg. Top. Comput. Intell. 2023, 7, 742–752. [Google Scholar] [CrossRef]
  3. Jahangir, H.; Lakshminarayana, S.; Maple, C.; Epiphaniou, G. A Deep Learning-Based Solution for Securing the Power Grid against Load Altering Threats by IoT-Enabled Devices. IEEE Internet Things J. 2023, 10, 23205575. [Google Scholar] [CrossRef]
  4. Shamshad, F.; Khan, S.; Zamir, S.W.; Khan, M.H.; Hayat, M.; Khan, F.S.; Fu, H. Transformers in medical imaging: A survey. Med. Image Anal. 2023, 88, 102802. [Google Scholar] [CrossRef] [PubMed]
  5. You, K.; Zhou, C.; Ding, L. Deep learning technology for construction machinery and robotics. Autom. Constr. 2023, 150, 104852. [Google Scholar] [CrossRef]
  6. Ragu, N.; Teo, J. Object detection and classification using few-shot learning in smart agriculture: A scoping mini review. Front. Sustain. Food Syst. 2023, 6, 1039299. [Google Scholar] [CrossRef]
  7. Zhang, S.; Zhang, C. Modified U-Net for plant diseased leaf image segmentation. Comput. Electron. Agric. 2023, 204, 107511. [Google Scholar] [CrossRef]
  8. Fu, C.; Lu, K.; Zheng, G.; Ye, J.; Cao, Z.; Li, B.; Lu, G. Siamese object tracking for unmanned aerial vehicle: A review and comprehensive analysis. Artif. Intell. Rev. 2023, 56, 1417–1477. [Google Scholar] [CrossRef]
  9. Zeng, Z.; Li, Z.; Cheng, D.; Zhang, H.; Zhan, K.; Yang, Y. Two-stream multirate recurrent neural network for video-based pedestrian reidentification. IEEE Trans. Ind. Inform. 2017, 14, 3179–3186. [Google Scholar] [CrossRef]
  10. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  11. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  12. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  13. ur Rehman, M.; Nizami, I.F.; Majid, M. DeepRPN-BIQA: Deep architectures with region proposal network for natural-scene and screen-content blind image quality assessment. Displays 2022, 71, 102101. [Google Scholar] [CrossRef]
  14. Liu, H.; Zheng, T.; Sun, F.; Wang, C.; Deng, L. ER-DeepSORT: Pedestrian Multiobject Tracking with Enhanced Reidentification. IEEJ Trans. Electr. Electron. Eng. 2023, 18, 427–435. [Google Scholar] [CrossRef]
  15. Li, X.; Jiang, Y.; Liu, Y.; Zhang, J.; Yin, S.; Luo, H. RAGCN: Region aggregation graph convolutional network for bone age assessment from X-ray images. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  16. Ji, Y.; Ni, L.; Zhao, C.; Lei, C.; Du, Y.; Wang, W. TriPField: A 3D potential field model and its applications to local path planning of autonomous vehicles. IEEE Trans. Intell. Transp. Syst. 2023, 24, 3541–3554. [Google Scholar] [CrossRef]
  17. Müller, H.; Niculescu, V.; Polonelli, T.; Magno, M.; Benini, L. Robust and efficient depth-based obstacle avoidance for autonomous miniaturized uavs. IEEE Trans. Robot. 2023, 39, 4935–4951. [Google Scholar] [CrossRef]
  18. Bayer, M.; Kaufhold, M.-A.; Buchhold, B.; Keller, M.; Dallmeyer, J.; Reuter, C. Data augmentation in natural language processing: A novel text generation approach for long and short text classifiers. Int. J. Mach. Learn. Cybern. 2023, 14, 135–150. [Google Scholar] [CrossRef]
  19. Gao, S.H.; Cheng, M.M.; Zhao, K.; Zhang, X.Y.; Yang, M.H.; Torr, P. Res2net: A new multi-scale backbone architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 652–662. [Google Scholar] [CrossRef]
  20. Wu, T.; Tang, S.; Zhang, R.; Cao, J.; Zhang, Y. Cgnet: A light-weight context guided network for semantic segmentation. IEEE Trans. Image Process. 2020, 30, 1169–1179. [Google Scholar] [CrossRef]
  21. Li, C.; Li, L.; Geng, Y.; Jiang, H.; Cheng, M.; Zhang, B.; Ke, Z.; Xu, X.; Chu, X. Yolov6 v3. 0: A full-scale reloading. arXiv 2023, arXiv:2301.05586. [Google Scholar]
  22. Tong, Z.; Chen, Y.; Xu, Z.; Yu, R. Wise-IoU: Bounding Box Regression Loss with Dynamic Focusing Mechanism. arXiv 2023, arXiv:2301.10051. [Google Scholar]
  23. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  24. Everingham, M.; Eslami, S.A.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes challenge: A retrospective. Int. J. Comput. Vis. 2015, 111, 98–136. [Google Scholar] [CrossRef]
  25. Zhu, P.; Wen, L.; Du, D.; Bian, X.; Fan, H.; Hu, Q.; Ling, H. Detection and tracking meet drones challenge. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 7380–7399. [Google Scholar] [CrossRef] [PubMed]
  26. Martin, D.R.; Fowlkes, C.C.; Malik, J. Learning to detect natural image boundaries using local brightness, color, and texture cues. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 530–549. [Google Scholar] [CrossRef] [PubMed]
  27. Lin, T.-Y.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  28. Dai, X.; Chen, Y.; Xiao, B.; Chen, D.; Liu, M.; Yuan, L.; Zhang, L. Dynamic head: Unifying object detection heads with attentions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7373–7382. [Google Scholar]
  29. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
  30. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer International Publishing: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  31. Ryu, J.; Rehman, M.U.; Nizami, I.F.; Chong, K.T. SegR-Net: A deep learning framework with multi-scale feature fusion for robust retinal vessel segmentation. Comput. Biol. Med. 2023, 163, 107132. [Google Scholar] [CrossRef]
Figure 1. Comparison between the number of parameters (M) and mAP50.
Figure 1. Comparison between the number of parameters (M) and mAP50.
Electronics 13 00043 g001
Figure 2. Overall framework.
Figure 2. Overall framework.
Electronics 13 00043 g002
Figure 3. The specific network architecture characterized. In the image, “k” stands for Kernel size. “s” stands for Stride. “p” stands for Padding. Depth: 0.33, width: 0.25; and MaxChannels: 1024.
Figure 3. The specific network architecture characterized. In the image, “k” stands for Kernel size. “s” stands for Stride. “p” stands for Padding. Depth: 0.33, width: 0.25; and MaxChannels: 1024.
Electronics 13 00043 g003
Figure 4. The main blocks of the backbone network and the neck network.
Figure 4. The main blocks of the backbone network and the neck network.
Electronics 13 00043 g004
Figure 5. Confusion matrix.
Figure 5. Confusion matrix.
Electronics 13 00043 g005
Figure 6. The F1 score performance of the model across multiple categories at varying confidence thresholds.
Figure 6. The F1 score performance of the model across multiple categories at varying confidence thresholds.
Electronics 13 00043 g006
Table 1. Dataset attributes.
Table 1. Dataset attributes.
DatasetImagesNumber of ClassObjects
Voc2012 [24]11,5402027,450
Voc2007 [23]99632024,640
VisDrone [25]10,209102,645,719
Table 2. The mAP50 for the PASCAL VOC2007 test dataset, utilizing a union of VOC2007 and VOC2012 trainval data. Herein, “R” stands for Res2Net, “C” for CGNet, “S” for SimPPFCSPC, “D” for DyHead, and “W” for WIoU. In the table, the bolded values represent the highest mAP among these methods.
Table 2. The mAP50 for the PASCAL VOC2007 test dataset, utilizing a union of VOC2007 and VOC2012 trainval data. Herein, “R” stands for Res2Net, “C” for CGNet, “S” for SimPPFCSPC, “D” for DyHead, and “W” for WIoU. In the table, the bolded values represent the highest mAP among these methods.
MethodAllAeroBikeBirdBoatBottleBusCarCatChairCowTableDogHorseMbikePersonPlantSheepSofaTrainTv
Fast [11]68.482.378.470.852.338.777.871.689.344.2735587.580.580.87235.168.365.780.464.2
Faster [29]70.484.979.874.353.949.877.575.988.545.677.155.386.981.780.979.640.172.660.981.261.5
YOLO [12]57.97767.257.738.322.768.355.981.436.260.848.577.272.371.363.528.952.254.873.950.8
SSD300 [30]72.485.680.170.557.646.279.476.189.2537760.88783.182.379.445.975.969.581.967.5
SSD512 [30]74.987.482.375.85952.681.781.59055.47959.888.484.384.783.350.27866.386.372
YOLOv3-tiny72.372.284.765.665.657.481.485.176.753.175.466.173.484.384.183.945.873.86577.375
YOLOv5n76.584.387.172.569.262.384.889.181.158.276.775.879.686.783.885.845.374.372.48476.3
YOLOv6n78.783.689.174.368.966.386.490.285.563.177.777.983.789.486.98747.774.977.286.876.5
YOLOv8n78.885.888.773.869.465.285.790.384.361.481.276.88188.385.887.150.878.576.189.377.1
YOLOv8n-W79.186.288.874.669.266.185.390.88661.880.978.582.489.986.187.153.17874.586.475.5
YOLOv8n-C79.98789.877.270.767.885.990.885.362.483.879.181.389.985.987.949.379.57688.778.9
YOLOv8n-S8087.689.676.771.666.186.190.686.562.780.776.784.990.187.387.650.67977.988.479.2
YOLOv8n-R80.186.388.176.970.466.786.891.687.862.781.878.883.590.587.38853.876.677.789.977.2
YOLOv8n-R-C-W80.588.39076.173.76886.791.186.363.980.278.884.689.588.488.452.777.5798878.3
YOLOv8n-D80.688.488.976.471.968.388.691.787.162.780.979.381.990.58788.452.181.178.289.878.3
YOLOv8n-R-S-W81.388.589.878.67467.988.29288.264.580.881.68691.38988.752.279.378.288.578.7
YOLOv8n-R-S-C-D81.78989.977.575.769.889.791.988.665.880.777.985.390.587.889.352.479.781.590.980.3
YOLOv8- CGRNet81.989.389.977.376.268.78991.789.166.281.47986.591.58889.153.878.780.890.280.9
Table 3. The progress of mAP50 per 10 epochs.
Table 3. The progress of mAP50 per 10 epochs.
EpochsYolov5nYolov6nYolov8nYolov8n-Wyolov8n-CYolov8n-SYolov8n-RYOLOv8n-R-C-WYOLOv8n-R-S-WYOLOv8n-R-S-C-DYolov8n-CGRNet
10.2580.0990.5920.5920.1270.3380.3850.1160.0890.3810.333
1139.10434.85447.92147.92144.18847.44946.33345.99345.21541.16741.449
2154.42155.09461.49861.49859.79562.9562.20261.87462.11160.09562.101
3162.42763.84767.64867.64867.67269.51469.12369.13170.0968.60970.142
4166.52268.48771.57371.57372.25973.32873.17173.35174.20573.21174.006
5168.90871.12773.73273.73274.3475.61175.03675.85976.01275.57676.195
6170.36572.66274.84874.84875.64577.03476.17377.00177.3776.96877.253
7171.3373.59975.5775.5776.36377.67576.97477.58878.12577.94878.136
8172.01974.11876.07376.07376.90478.02777.47178.03278.64578.57878.617
9172.57174.776.43476.43477.31478.44577.87478.41279.10879.00979.044
10173.02675.25176.83176.83177.71178.82978.21878.74379.53179.45379.484
11173.48375.84477.15577.15578.02279.14478.50779.14479.83479.73979.875
12173.90576.35577.49877.49878.25679.41878.85179.46680.1680.10180.193
13174.2876.74577.77577.77578.53179.59779.09679.65780.36480.38980.553
14174.70477.21578.04878.04878.79779.72279.36279.89380.5980.68880.865
15175.04477.50178.2978.2978.94479.81279.59680.11380.80480.98381.139
16175.39177.76878.46978.46979.15479.96679.79680.20681.06781.21481.308
17175.71878.0578.68678.68679.31980.01979.96280.29281.1481.34481.527
18176.00278.20878.89378.89379.41180.01680.11980.38381.18681.49281.68
19176.29178.47878.95278.95279.6379.97980.12580.48281.21381.63781.824
20076.4678.64578.98678.98679.8180.00180.13980.47681.25881.70181.91
Table 4. The mAP50 for the VisDrone2019-DET. Herein, “R” stands for Res2Net, “C” for CGNet, “S” for SimPPFCSPC, “D” for DyHead, and “W” for WIoU. In the table, the bolded values represent the highest mAP among these methods.
Table 4. The mAP50 for the VisDrone2019-DET. Herein, “R” stands for Res2Net, “C” for CGNet, “S” for SimPPFCSPC, “D” for DyHead, and “W” for WIoU. In the table, the bolded values represent the highest mAP among these methods.
MethodAllPedestrianPeopleBicycleCarVanTruckTricycleAwning TricycleBusMotor
yolov6n25.32010.34.12663129.19.7812.450.520.2
yolov5n26.521.311.36.0466.530.831.311.513.150.223.2
yolov8n26.922.111.65.8966.730.530.912.714.750.823.1
Yolov8n-S27.122.911.47.3666.830.829.912.415.150.924
yolov8n-W27.52312.76.4967.231.331.611.614.952.623.8
yolov8n-C27.822.712.25.5667.532.532.213.316.65223
yolov8n-R-S28.223.212.87.276832.133.213.115.752.524.3
yolov8n-R28.223.412.46.3168.132.532.91316.352.724.2
yolov8n-R-C28.523.813.37.568.132.331.91316.653.325
yolov8n-D28.723.512.57.4468.332.933.814.415.753.224.7
YOLOv8n-R-S-C-D28.822.712.46.9868.633.935.612.717.653.124.4
yolov8- CGRNet29.924.213.67.7269.333.336.814.91855.226.1
Table 5. The test results of various object detection models when evaluated on the VisDrone dataset, utilizing a NVIDIA GeForce RTX 3080 graphics card.
Table 5. The test results of various object detection models when evaluated on the VisDrone dataset, utilizing a NVIDIA GeForce RTX 3080 graphics card.
Yolov6nYolov5nYolov8nYolov8n-SYolov8n-WYolov8n-CYolov8n-R-CYolov8n-DYOLOv8n-R-SYolov8n-RYOLOv8n-R-S-C-DYolov8-CGRNet
FPS175.44153.85161.29151.52153.85119.05112.3699.01109.89131.5785.4784.03
Inference (ms)3.74.54.24.64.56.46.98.17.15.69.79.9
GFLOPs11.87.18.19.48.19.19.99.711.28.812.812.8
Table 6. The VOC2007 test results of various object detection models when evaluated on the VisDrone dataset, utilizing a NVIDIA GeForce RTX 3080 graphics card.
Table 6. The VOC2007 test results of various object detection models when evaluated on the VisDrone dataset, utilizing a NVIDIA GeForce RTX 3080 graphics card.
Yolov6nYolov5nYolov8nYolov8n-SYolov8n-WYolov8n-CYolov8n-R-CYolov8n-DYOLOv8n-R-SYolov8n-RYOLOv8n-R-S-C-DYolov8-CGRNet
FPS72.46 109.89 99.01 87.72 99.01 90.09 84.03 85.47 75.76 92.59 67.57 67.57
Inference (ms)1.82.12.32.31.93.03.54.63.52.66.46.5
GFLOPs11.87.18.19.48.19.19.99.711.28.812.812.8
Table 7. The test results of various object detection models when evaluated on the VisDrone dataset, utilizing an ARM Cortex-A57cpu.
Table 7. The test results of various object detection models when evaluated on the VisDrone dataset, utilizing an ARM Cortex-A57cpu.
Yolov6nYolov5nYolov8nYolov8n-SYolov8n-WYolov8n-CYolov8n-R-CYolov8n-DYOLOv8n-R-SYolov8n-RYOLOv8n-R-S-C-DYolov8-CGRNet
FPS1.071.021.101.001.010.760.620.520.610.990.420.41
Inference (ms)9299749079999841310162419311631101124042458
GFLOPs11.87.18.19.48.19.19.99.711.28.812.812.8
Table 8. The VOC2007 test results of various object detection models when evaluated on the VisDrone dataset, utilizing an ARM Cortex-A57 graphics card.
Table 8. The VOC2007 test results of various object detection models when evaluated on the VisDrone dataset, utilizing an ARM Cortex-A57 graphics card.
Yolov6nYolov5nYolov8nYolov8n-SYolov8n-WYolov8n-CYolov8n-R-CYolov8n-DYOLOv8n-R-SYolov8n-RYOLOv8n-R-S-C-DYolov8-CGRNet
FPS0.440.400.470.430.450.320.290.220.270.390.190.19
Inference (ms)222524402086225421633083345745063705251053145321
GFLOPs11.87.18.19.48.19.19.99.79.48.812.812.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Niu, Y.; Cheng, W.; Shi, C.; Fan, S. YOLOv8-CGRNet: A Lightweight Object Detection Network Leveraging Context Guidance and Deep Residual Learning. Electronics 2024, 13, 43. https://doi.org/10.3390/electronics13010043

AMA Style

Niu Y, Cheng W, Shi C, Fan S. YOLOv8-CGRNet: A Lightweight Object Detection Network Leveraging Context Guidance and Deep Residual Learning. Electronics. 2024; 13(1):43. https://doi.org/10.3390/electronics13010043

Chicago/Turabian Style

Niu, Yixing, Wansheng Cheng, Chunni Shi, and Song Fan. 2024. "YOLOv8-CGRNet: A Lightweight Object Detection Network Leveraging Context Guidance and Deep Residual Learning" Electronics 13, no. 1: 43. https://doi.org/10.3390/electronics13010043

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop