Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (11)

Search Parameters:
Keywords = YOLOv8-FC

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
30 pages, 19525 KB  
Article
Disease Monitoring and Characterization of Feeder Road Network Based on Improved YOLOv11
by Ying Fan, Kun Zhi, Haichao An, Runyin Gu, Xiaobing Ding and Jianhua Tang
Electronics 2025, 14(9), 1818; https://doi.org/10.3390/electronics14091818 - 29 Apr 2025
Viewed by 770
Abstract
In response to the challenges of the low accuracy and high misdetection and omission rate of disease detection on feeder roads, an improved Rural-YOLO (SAConv-C2f+C2PSA_CAA+MCSAttention+WIOU) disease detection algorithm is proposed in this paper, which is an enhanced target detection framework based on the [...] Read more.
In response to the challenges of the low accuracy and high misdetection and omission rate of disease detection on feeder roads, an improved Rural-YOLO (SAConv-C2f+C2PSA_CAA+MCSAttention+WIOU) disease detection algorithm is proposed in this paper, which is an enhanced target detection framework based on the YOLOv11 architecture, for the identification of common diseases in the complex feeder road environment. The proposed methodology introduces four key innovations: (1) Switchable Atrous Convolution (SAConv) is introduced into the backbone network to enhance multiscale disease feature extraction under occlusion conditions; (2) Multi-Channel and Spatial Attention (MCSAttention) is constructed in the feature fusion process, and the weight distribution of multiscale diseases is adjusted through adaptive weight redistribution. By adjusting the weight distribution, the model’s sensitivity to subtle disease features is improved. To enhance its ability to discriminate between different disease types, Cross Stage Partial with Parallel Spatial Attention and Channel Adaptive Aggregation (C2PSA_CAA) is constructed at the end of the backbone network. (3) To mitigate category imbalance issues, Weighted Intersection over Union loss (WIoU_loss) is introduced, which helps optimize the bounding box regression process in disease detection and improve the detection of relevant diseases. Based on experimental validation, Rural-YOLO demonstrated superior performance with minimal computational overhead. Only 0.7 M additional parameters is required, and an 8.4% improvement in recall and a 7.8% increase in mAP50 were achieved compared to the initial models. The optimized architecture also reduced the model size by 21%. The test results showed that the proposed model achieved 3.28 M parameters with a computational complexity of 5.0 GFLOPs, meeting the requirements for lightweight deployment scenarios. Cross-validation on multi-scenario public datasets was carried out, and the model’s robustness across diverse road conditions. In the quantitative experiments, the center skeleton method and the maximum internal tangent circle method were used to calculate crack width, and the pixel occupancy ratio method was used to assess the area damage degree of potholes and other diseases. The measurements were converted to actual physical dimensions using a calibrated scale of 0.081:1. Full article
Show Figures

Figure 1

32 pages, 9587 KB  
Article
A Layered Framework for Universal Extraction and Recognition of Electrical Diagrams
by Weiguo Cao, Zhong Chen, Congying Wu and Tiecheng Li
Electronics 2025, 14(5), 833; https://doi.org/10.3390/electronics14050833 - 20 Feb 2025
Cited by 1 | Viewed by 1210
Abstract
Secondary systems in electrical engineering often rely on traditional CAD software (AutoCAD v2024.1.6) or non-structured, paper-based diagrams for fieldwork, posing challenges for digital transformation. Electrical diagram recognition technology bridges this gap by converting traditional diagram operations into a “digital” model, playing a critical [...] Read more.
Secondary systems in electrical engineering often rely on traditional CAD software (AutoCAD v2024.1.6) or non-structured, paper-based diagrams for fieldwork, posing challenges for digital transformation. Electrical diagram recognition technology bridges this gap by converting traditional diagram operations into a “digital” model, playing a critical role in power system scheduling, operation, and maintenance. However, conventional recognition methods, which primarily rely on partition detection, face significant limitations such as poor adaptability to diverse diagram styles, interference among recognition objects, and reduced accuracy in handling complex and varied electrical diagrams. This paper introduces a novel layered framework for electrical diagram recognition that sequentially extracts the element layer, text layer, and connection relationship layer to address these challenges. First, an improved YOLOv7 model, combined with a multi-scale sliding window strategy, is employed to accurately segment large and small diagram objects. Next, PaddleOCR, trained with electrical-specific terminology, and PaddleClas, using multi-angle classification, are utilized for robust text recognition, effectively mitigating interference from diagram elements. Finally, clustering and adaptive FcF-inpainting algorithms are applied to repair the connection relationship layer, resolving local occlusion issues and enhancing the overall coupling of the diagram. Experimental results demonstrate that the proposed method outperforms existing approaches in robustness and universality, particularly for complex diagrams, providing technical support for intelligent power grid construction and operation. Full article
Show Figures

Figure 1

26 pages, 13643 KB  
Article
An Approach to Multiclass Industrial Heat Source Detection Using Optical Remote Sensing Images
by Yi Zeng, Ruilin Liao, Caihong Ma, Dacheng Wang and Yongze Lv
Energies 2025, 18(4), 865; https://doi.org/10.3390/en18040865 - 12 Feb 2025
Viewed by 1032
Abstract
Industrial heat sources (IHSs) are major contributors to energy consumption and environmental pollution, making their accurate detection crucial for supporting industrial restructuring and emission reduction strategies. However, existing models either focus on single-class detection under complex backgrounds or handle multiclass tasks for simple [...] Read more.
Industrial heat sources (IHSs) are major contributors to energy consumption and environmental pollution, making their accurate detection crucial for supporting industrial restructuring and emission reduction strategies. However, existing models either focus on single-class detection under complex backgrounds or handle multiclass tasks for simple targets, leaving a gap in effective multiclass detection for complex scenarios. To address this, we propose a novel multiclass IHS detection model based on the YOLOv8-FC framework, underpinned by the multiclass IHS training dataset constructed from optical remote sensing images and point-of-interest (POI) data firstly. This dataset incorporates five categories: cement plants, coke plants, coal mining areas, oil and gas refineries, and steel plants. The proposed YOLOv8-FC model integrates the FasterNet backbone and a Coordinate Attention (CA) module, significantly enhancing feature extraction, detection precision, and operational speed. Experimental results demonstrate the model’s robust performance, achieving a precision rate of 92.3% and a recall rate of 95.6% in detecting IHS objects across diverse backgrounds. When applied in the Beijing–Tianjin–Hebei (BTH) region, YOLOv8-FC successfully identified 429 IHS objects, with detailed category-specific results providing valuable insights into industrial distribution. It shows that our proposed multiclass IHS detection model with the novel YOLOv8-FC approach could effectively and simultaneously detect IHS categories under complex backgrounds. The IHS datasets derived from the BTH region can support regional industrial restructuring and optimization schemes. Full article
(This article belongs to the Section J: Thermal Management)
Show Figures

Figure 1

22 pages, 10092 KB  
Article
Sewage and Location Detection with Improved Cycle Generative Adversarial Network-Based Augmented Datasets and YOLOv5-BiFC
by Minkai Wang, Chenglin Li, Yunzhong Jiang and Mingxiang Yang
Appl. Sci. 2024, 14(21), 9932; https://doi.org/10.3390/app14219932 - 30 Oct 2024
Cited by 1 | Viewed by 1146
Abstract
Sewage discharge from outfalls significantly contaminates the environment. However, due to the unique characteristics of environmental policy, challenges such as data acquisition difficulties arise. This study introduces an enhanced approach by utilizing an improved Cycle GAN, the core function of which involves extrapolating [...] Read more.
Sewage discharge from outfalls significantly contaminates the environment. However, due to the unique characteristics of environmental policy, challenges such as data acquisition difficulties arise. This study introduces an enhanced approach by utilizing an improved Cycle GAN, the core function of which involves extrapolating a small sample to a large sample. An enhanced YOLOv5 model is used to focus on lightweight model construction and model performance enhancement. The proposed Cycle GAN incorporating Self-Attention and residual modules is suggested to tackle the problem of limited data at the outfall. Additionally, a refined YOLOv5 model (YOLOV5-BIFC) is proposed, integrating the C3Ghost module, BiFPN module, and CBAM attention mechanism to overcome low model recognition efficiency and large model size concerns. The research employs an augmented dataset for training and evaluates model performance using metrics such as mAP, F1 score, model size, and FPS. The results indicate that the YOLOV5-BIFC model has a size of 7.8 MB, representing a 44.2% reduction compared to the original YOLOv5 model. The FPS of this model is 26.7, enabling real-time discharge monitoring. The F1 score and mAP achieved by YOLOv5-BiFC are 89.8% and 87.3%. Comparative analysis with other neural networks demonstrates the superior accuracy and efficiency of the enhanced model. The YOLOv5-BiFC model facilitates precise, rapid, and intelligent discharge inspection. Full article
Show Figures

Figure 1

26 pages, 10106 KB  
Article
DFLM-YOLO: A Lightweight YOLO Model with Multiscale Feature Fusion Capabilities for Open Water Aerial Imagery
by Chen Sun, Yihong Zhang and Shuai Ma
Drones 2024, 8(8), 400; https://doi.org/10.3390/drones8080400 - 16 Aug 2024
Cited by 8 | Viewed by 2603
Abstract
Object detection algorithms for open water aerial images present challenges such as small object size, unsatisfactory detection accuracy, numerous network parameters, and enormous computational demands. Current detection algorithms struggle to meet the accuracy and speed requirements while being deployable on small mobile devices. [...] Read more.
Object detection algorithms for open water aerial images present challenges such as small object size, unsatisfactory detection accuracy, numerous network parameters, and enormous computational demands. Current detection algorithms struggle to meet the accuracy and speed requirements while being deployable on small mobile devices. This paper proposes DFLM-YOLO, a lightweight small-object detection network based on the YOLOv8 algorithm with multiscale feature fusion. Firstly, to solve the class imbalance problem of the SeaDroneSee dataset, we propose a data augmentation algorithm called Small Object Multiplication (SOM). SOM enhances dataset balance by increasing the number of objects in specific categories, thereby improving model accuracy and generalization capabilities. Secondly, we optimize the backbone network structure by implementing Depthwise Separable Convolution (DSConv) and the newly designed FasterBlock-CGLU-C2f (FC-C2f), which reduces the model’s parameters and inference time. Finally, we design the Lightweight Multiscale Feature Fusion Network (LMFN) to address the challenges of multiscale variations by gradually fusing the four feature layers extracted from the backbone network in three stages. In addition, LMFN incorporates the Dilated Re-param Block structure to increase the effective receptive field and improve the model’s classification ability and detection accuracy. The experimental results on the SeaDroneSee dataset indicate that DFLM-YOLO improves the mean average precision (mAP) by 12.4% compared to the original YOLOv8s, while reducing parameters by 67.2%. This achievement provides a new solution for Unmanned Aerial Vehicles (UAVs) to conduct object detection missions in open water efficiently. Full article
Show Figures

Figure 1

17 pages, 4756 KB  
Article
CFE-YOLOv8s: Improved YOLOv8s for Steel Surface Defect Detection
by Shuxin Yang, Yang Xie, Jianqing Wu, Weidong Huang, Hongsheng Yan, Jingyong Wang, Bi Wang, Xiangchun Yu, Qiang Wu and Fei Xie
Electronics 2024, 13(14), 2771; https://doi.org/10.3390/electronics13142771 - 15 Jul 2024
Cited by 9 | Viewed by 3076
Abstract
Due to the low detection accuracy in steel surface defect detection and the constraints of limited hardware resources, we propose an improved model for steel surface defect detection, named CBiF-FC-EFC-YOLOv8s (CFE-YOLOv8s), including CBS-BiFormer (CBiF) modules, Faster-C2f (FC) modules, and EMA-Faster-C2f (EFC) modules. Firstly, [...] Read more.
Due to the low detection accuracy in steel surface defect detection and the constraints of limited hardware resources, we propose an improved model for steel surface defect detection, named CBiF-FC-EFC-YOLOv8s (CFE-YOLOv8s), including CBS-BiFormer (CBiF) modules, Faster-C2f (FC) modules, and EMA-Faster-C2f (EFC) modules. Firstly, because of the potential information loss that convolutional neural networks (CNN) may encounter when dealing with miniature targets, the CBiF combines CNN with Transformer to optimize local and global features. Secondly, to address the increased computational complexity caused by the extensive use of convolutional layers, the FC uses the FasterNet block to reduce redundant computations and memory access. Lastly, the EMA is incorporated into the FC to design the EFC module and enhance feature fusion capability while ensuring the light weight of the model. CFE-YOLOv8s achieves mAP@0.5 values of 77.8% and 69.5% on the NEU-DET and GC10-DET datasets, respectively, representing enhancements of 3.1% and 2.8% over YOLOv8s, with reductions of 22% and 18% in model parameters and FLOPS. The CFE-YOLOv8s demonstrates superior overall performance and balance compared to other advanced models. Full article
(This article belongs to the Special Issue Machine Learning and Deep Learning Based Pattern Recognition)
Show Figures

Figure 1

19 pages, 7940 KB  
Article
Light-FC-YOLO: A Lightweight Method for Flower Counting Based on Enhanced Feature Fusion with a New Efficient Detection Head
by Xiaomei Yi, Hanyu Chen, Peng Wu, Guoying Wang, Lufeng Mo, Bowei Wu, Yutong Yi, Xinyun Fu and Pengxiang Qian
Agronomy 2024, 14(6), 1285; https://doi.org/10.3390/agronomy14061285 - 14 Jun 2024
Cited by 6 | Viewed by 1939
Abstract
Fast and accurate counting and positioning of flowers is the foundation of automated flower cultivation production. However, it remains a challenge to complete the counting and positioning of high-density flowers against a complex background. Therefore, this paper proposes a lightweight flower counting and [...] Read more.
Fast and accurate counting and positioning of flowers is the foundation of automated flower cultivation production. However, it remains a challenge to complete the counting and positioning of high-density flowers against a complex background. Therefore, this paper proposes a lightweight flower counting and positioning model, Light-FC-YOLO, based on YOLOv8s. By integrating lightweight convolution, the model is more portable and deployable. At the same time, a new efficient detection head, Efficient head, and the integration of the LSKA large kernel attention mechanism are proposed to enhance the model’s feature detail extraction capability and change the weight ratio of the shallow edge and key point information in the network. Finally, the SIoU loss function with target angle deviation calculation is introduced to improve the model’s detection accuracy and target positioning ability. Experimental results show that Light-FC-YOLO, with a model size reduction of 27.2% and a parameter reduction of 39.0%, has a Mean Average Precision (mAP) and recall that are 0.8% and 1.4% higher than YOLOv8s, respectively. In the counting comparison experiment, the coefficient of determination (R2) and Root Mean Squared Error (RMSE) of Light-FC-YOLO reached 0.9577 and 8.69, respectively, both superior to lightweight models such as YOLOv8s. The lightweight flower detection method proposed in this paper can efficiently complete flower positioning and counting tasks, providing technical support and reference solutions for automated flower production management. Full article
Show Figures

Figure 1

24 pages, 3319 KB  
Article
YOLOv5-SA-FC: A Novel Pig Detection and Counting Method Based on Shuffle Attention and Focal Complete Intersection over Union
by Wangli Hao, Li Zhang, Meng Han, Kai Zhang, Fuzhong Li, Guoqiang Yang and Zhenyu Liu
Animals 2023, 13(20), 3201; https://doi.org/10.3390/ani13203201 - 13 Oct 2023
Cited by 11 | Viewed by 2348
Abstract
The efficient detection and counting of pig populations is critical for the promotion of intelligent breeding. Traditional methods for pig detection and counting mainly rely on manual labor, which is either time-consuming and inefficient or lacks sufficient detection accuracy. To address these issues, [...] Read more.
The efficient detection and counting of pig populations is critical for the promotion of intelligent breeding. Traditional methods for pig detection and counting mainly rely on manual labor, which is either time-consuming and inefficient or lacks sufficient detection accuracy. To address these issues, a novel model for pig detection and counting based on YOLOv5 enhanced with shuffle attention (SA) and Focal-CIoU (FC) is proposed in this paper, which we call YOLOv5-SA-FC. The SA attention module in this model enables multi-channel information fusion with almost no additional parameters, enhancing the richness and robustness of feature extraction. Furthermore, the Focal-CIoU localization loss helps to reduce the impact of sample imbalance on the detection results, improving the overall performance of the model. From the experimental results, the proposed YOLOv5-SA-FC model achieved a mean average precision (mAP) and count accuracy of 93.8% and 95.6%, outperforming other methods in terms of pig detection and counting by 10.2% and 15.8%, respectively. These findings verify the effectiveness of the proposed YOLOv5-SA-FC model for pig population detection and counting in the context of intelligent pig breeding. Full article
(This article belongs to the Section Pigs)
Show Figures

Figure 1

24 pages, 8705 KB  
Article
A Semi-Supervised Learning Approach for Automatic Detection and Fashion Product Category Prediction with Small Training Dataset Using FC-YOLOv4
by Yamin Thwe, Nipat Jongsawat and Anucha Tungkasthan
Appl. Sci. 2022, 12(16), 8068; https://doi.org/10.3390/app12168068 - 12 Aug 2022
Cited by 5 | Viewed by 3770
Abstract
Over the past few decades, research on object detection has developed rapidly, one of which can be seen in the fashion industry. Fast and accurate detection of an E-commerce fashion product is crucial to choosing the appropriate category. Nowadays, both new and second-hand [...] Read more.
Over the past few decades, research on object detection has developed rapidly, one of which can be seen in the fashion industry. Fast and accurate detection of an E-commerce fashion product is crucial to choosing the appropriate category. Nowadays, both new and second-hand clothing is provided by E-commerce sites for purchase. Therefore, when categorizing fashion clothing, it is essential to categorize it precisely, regardless of the cluttered background. We present recently acquired tiny product images with various resolutions, sizes, and positions datasets from the Shopee E-commerce (Thailand) website. This paper also proposes the Fashion Category—You Only Look Once version 4 model called FC-YOLOv4 for detecting multiclass fashion products. We used the semi-supervised learning approach to reduce image labeling time, and the number of resulting images is then increased through image augmentation. This approach results in reasonable Average Precision (AP), Mean Average Precision (mAP), True or False Positive (TP/FP), Recall, Intersection over Union (IoU), and reliable object detection. According to experimental findings, our model increases the mAP by 0.07 percent and 40.2 percent increment compared to the original YOLOv4 and YOLOv3. Experimental findings from our FC-YOLOv4 model demonstrate that it can effectively provide accurate fashion category detection for properly captured and clutter images compared to the YOLOv4 and YOLOv3 models. Full article
(This article belongs to the Special Issue Applications of Deep Learning and Artificial Intelligence Methods)
Show Figures

Figure 1

16 pages, 5018 KB  
Article
Contextual Detection of Pedestrians and Vehicles in Orthophotography by Fusion of Deep Learning Algorithms
by Masoomeh Shireen Ansarnia, Etienne Tisserand, Patrick Schweitzer, Mohamed Amine Zidane and Yves Berviller
Sensors 2022, 22(4), 1381; https://doi.org/10.3390/s22041381 - 11 Feb 2022
Cited by 6 | Viewed by 3355
Abstract
In the context of smart cities, monitoring pedestrian and vehicle movements is essential to recognize abnormal events and prevent accidents. The proposed method in this work focuses on analyzing video streams captured from a vertically installed camera, and performing contextual road user detection. [...] Read more.
In the context of smart cities, monitoring pedestrian and vehicle movements is essential to recognize abnormal events and prevent accidents. The proposed method in this work focuses on analyzing video streams captured from a vertically installed camera, and performing contextual road user detection. The final detection is based on the fusion of the outputs of three different convolutional neural networks. We are simultaneously interested in detecting road users, their motion, and their location respecting the static environment. We use YOLOv4 for object detection, FC-HarDNet for background semantic segmentation, and FlowNet 2.0 for motion detection. FC-HarDNet and YOLOv4 were retrained with our orthophotographs dataset. The last step involves a data fusion module. The presented results show that the method allows one to detect road users, identify the surfaces on which they move, quantify their apparent velocity, and estimate their actual velocity. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

18 pages, 11193 KB  
Article
Ship Detection in SAR Images Based on Multi-Scale Feature Extraction and Adaptive Feature Fusion
by Kexue Zhou, Min Zhang, Hai Wang and Jinlin Tan
Remote Sens. 2022, 14(3), 755; https://doi.org/10.3390/rs14030755 - 6 Feb 2022
Cited by 90 | Viewed by 7969
Abstract
Deep learning has attracted increasing attention across a number of disciplines in recent years. In the field of remote sensing, ship detection based on deep learning for synthetic aperture radar (SAR) imagery is replacing traditional methods as a mainstream research method. The multiple [...] Read more.
Deep learning has attracted increasing attention across a number of disciplines in recent years. In the field of remote sensing, ship detection based on deep learning for synthetic aperture radar (SAR) imagery is replacing traditional methods as a mainstream research method. The multiple scales of ship objects make the detection of ship targets a challenging task in SAR images. This paper proposes a new methodology for better detection of multi-scale ship objects in SAR images, which is based on YOLOv5 with a small model size (YOLOv5s), namely the multi-scale ship detection network (MSSDNet). We construct two modules in MSSDNet: the CSPMRes2 (Cross Stage Partial network with Modified Res2Net) module for improving feature representation capability and the FC-FPN (Feature Pyramid Network with Fusion Coefficients) module for fusing feature maps adaptively. Firstly, the CSPMRes2 module introduces modified Res2Net (MRes2) with a coordinate attention module (CAM) for multi-scale features extraction in scale dimension, then the CSPMRes2 module will be used as a basic module in the depth dimension of the MSSDNet backbone. Thus, our backbone of MSSDNet has the capabilities of features extraction in both depth and scale dimensions. In the FC-FPN module, we set a learnable fusion coefficient for each feature map participating in fusion, which helps the FC-FPN module choose the best features to fuse for multi-scale objects detection tasks. After the feature fusion, we pass the output through the CSPMRes2 module for better feature representation. The performance evaluation for this study is conducted using an RTX2080Ti GPU, and two different datasets: SSDD and SARShip are used. These experiments on SSDD and SARShip datasets confirm that MSSDNet leads to superior multi-scale ship detection compared with the state-of-the-art methods. Moreover, in comparisons of network model size and inference time, our MSSDNet also has huge advantages with related methods. Full article
(This article belongs to the Special Issue Target Detection and Information Extraction in Radar Images)
Show Figures

Figure 1

Back to TopTop