Next Article in Journal
Concrete Surface Crack Detection Algorithm Based on Improved YOLOv8
Previous Article in Journal
Human Activity Recognition, Monitoring, and Analysis Facilitated by Novel and Widespread Applications of Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Fast and High-Accuracy Foreign Object Detection Method for Belt Conveyor Coal Flow Images with Target Occlusion

1
School of Mechanical Engineering, Xi’an University of Science and Technology, Xi’an 710054, China
2
Shaanxi Key Laboratory of Mine Electromechanical Equipment Intelligent Detection and Control, Xi’an University of Science and Technology, Xi’an 710054, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(16), 5251; https://doi.org/10.3390/s24165251
Submission received: 14 June 2024 / Revised: 16 July 2024 / Accepted: 29 July 2024 / Published: 14 August 2024
(This article belongs to the Topic Applications in Image Analysis and Pattern Recognition)

Abstract

:
Foreign objects in coal flow easily cause damage to conveyor belts, and most foreign objects are often occluded, making them difficult to detect. Aiming at solving the problems of low accuracy and efficiency in the detection of occluded targets in a low-illumination and dust fog environment, an image detection method for foreign objects is proposed. Firstly, YOLOv5s back-end processing is optimized by soft non-maximum suppression to reduce the influence of dense objects. Secondly, SimOTA label allocation is used to reduce the influence of ambiguous samples under dense occlusion. Then, Slide Loss is used to excavate difficult samples, and Inner–SIoU is used to optimize the bounding box regression loss. Finally, Group–Taylor pruning is used to compress the model. The experimental results show that the proposed method has only 4.20 × 105 parameters, a computational amount of 1.00 × 109, a model size of 1.20 MB, and an mAP0.5 of up to 91.30% on the self-built dataset. The detection speed on the different computing devices is as high as 66.31, 41.90, and 33.03 FPS. This proves that the proposed method achieves fast and high-accuracy detection of multi-layer occluded coal flow foreign objects.

1. Introduction

The intelligentization of coal mines is the core technical requirement to support the high-quality development of the coal industry. As the key equipment of the main coal transportation system, the intelligentization of belt conveyors is required to guarantee the safe and efficient transportation of coal, which is of great significance to improving the safety level of coal mines and ensuring the stable supply of coal [1,2]. In the process of the continuous high-strength operation of coal transportation in the complex and harsh environment of coal mines, the materials conveyed by conveyor belts are often mixed with sharp foreign objects, such as gangues and iron wares, which easily cause local damage to conveyor belts, resulting in safety accidents. If foreign objects such as gangues and iron wares can be identified and sorted before the appearance of major safety hazards, the continuous, stable, green, and efficient production of coal mines will undoubtedly be guaranteed, thus advancing the intelligent development of coal mines [3,4].
Common methods for foreign object detection on the belt conveyors of coal mines include manual sorting, iron removal, spectral detection, and image recognition [5]. With the development of intelligent coal mines, inefficient and labor-intensive manual sorting methods cannot meet the requirements of intelligent and unmanned operation. The iron removal method cannot detect non-metallic foreign objects such as gangues and wood, and the efficiency of spectral detection is low. Compared to traditional image recognition algorithms, the target detection algorithm based on deep learning has received widespread attention due to its strong feature extraction ability, good generalization capability, and high recognition efficiency. At present, many scholars use machine vision and deep learning methods to conduct in-depth research on the intelligent identification of foreign objects such as gangues and iron wares on the conveyor belts of coal mines.
In an investigation of the influence of coal mine illumination on the recognition of foreign objects, Wang et al. [6] studied the changing rules of different illumination and gangue surface features. Li et al. [7] studied the influence of light, water, and dust on the image characteristics of gangues through gray and texture features. Cao et al. [8] studied a high-quality image acquisition method under multiple factors such as light source distribution and illumination, and based on this research, the recognition rate of gangues was improved. Li et al. [9] used the Retinex algorithm to adjust the overall brightness of the image and reduce the interference of dust, water mist, etc., and used the least squares support vector machine as a classifier to prove that enhancing the illumination can improve the recognition rate of gangues.
In the dust fog environment, Hao et al. [3] investigated dust interference and the high-speed operation of conveyors by introducing the Convolutional Block Attention Module (CBAM) into YOLOv5, replacing the ordinary convolution and loss function to improve the detection accuracy and speed of foreign objects. Mao et al. [10] examined the low foreign object detection accuracy caused by environmental factors such as dust fog and blurring by using a dark channel prior defogging algorithm to reduce the impact of dust fog on image clarity. Using 3 × 3 convolution to refine the image to reduce the blurring under high-speed conditions, the detection accuracy of foreign objects on the enhanced YOLOv5 algorithm was improved. In view of damage and interference to the conveyor belts’ surface in a low-illumination and dust fog environment, Fan et al. [11] enhanced the clarity of the image by using limited contrast adaptive histogram equalization and then improved YOLOv5, which enhances the detection effect of foreign objects. In an environment with a high concentration of dust and water mist, Fan et al. [12] proposed a method of the dust removal and illumination enhancement of the images based on dark channel segmentation and fusion; the enhanced image details were more obvious, and the detection effect of foreign objects was improved.
In terms of improving the performance of the target detection model, Gao et al. [13] used the CBAM to improve the MobileNetV3 network, reduced the overfitting of the model through data enhancement, and improved the accuracy of gangue detection. Xu et al. [14] identified gangue images based on the Convolutional Neural Network (CNN) and optimized the model with the model pruning method, which improved the detection accuracy and reduced the model size. Guo et al. [15] used the optimization method to improve the classical neural network and constructed a gangue detection model for small samples, which reduced the model size and improved the detection accuracy. Wu et al. [16] proposed a foreign object detection model based on the Faster R–CNN and bidirectional feature pyramid network, which improved the detection accuracy. Lv [17] used the VGG16 to improve the Faster R–CNN network to complete the intelligent identification of gangues and irons with a high recall rate. Hu et al. [18] used Focal Loss to replace the cross-entropy loss function of the YOLOv3 model and optimized its hyperparameters, identifying the anchor rods, irons, and nuts with high confidence. Ren et al. [19] proposed a foreign object detection method based on the improved CenterNet, which reduced false and missed detections. Wang et al. [20] optimized the SSD model with the deep separable convolution and generalized intersection over union (GIoU) loss function, which improved the detection accuracy and speed. Cheng et al. [21] proposed a lightweight network that integrated residual information, which improved the detection accuracy of foreign objects and detection speed. Chen [22] improved YOLOv5 and designed a foreign object detection system for conveyor belts, which realized the real-time detection of foreign objects. Based on the above studies, intelligent detection based on deep learning is the main technical means of foreign object detection, and measurable progress has been achieved.
However, in actual coal mines, foreign objects in real-time coal flow are often in a state of mutual occlusion. Ding [23] used a near-infrared camera and a visible light camera to form a binocular system, established a real-time multi-level fusion detection system for sensors, samples, and features, the SVM was combined with YOLOv3, and the length suppression algorithm was introduced to solve the problem of occlusion. Li et al. [24] used SKNet to replace MobileNetv3 as a backbone network and added a shallow detection structure and optimized loss with CIoU, which solved the problems of large memory occupation and low detection speed and accuracy for small and occluded targets, but the occlusion degree was low, and the categories of identified targets were few. Zhang et al. [25] designed a lightweight YOLOv4 model to solve the problems of image blurring and poor real-time performance at high speed; the dataset constructed by six categories of non-coal foreign objects contained a small number of occluded samples, and the occlusion degree was also low. In the process of using KinD++ and YOLOv4 to achieve the efficient detection of foreign objects in low illumination, Chen et al. [26] used the GridMask method to partially mask foreign object images, thereby expanding the image dataset. Mao et al. [10] conducted a test on YOLOv5, and the missed detection of foreign objects was reduced after model improvement, but there was also a low-occlusion case. Yao et al. [27] used iron, wood, and gangue targets to imitate foreign objects, and the mutual occlusion among foreign objects was considered, and data enhancement as random occlusion was used to expand the number of samples; high accuracy, recall, and mAP0.5 were obtained on the enhanced YOLO model, but the environmental factors were not considered.
In view of the above work, it is found that the detection of foreign objects in coal flow mainly focuses on the recognition of non-occluded foreign objects. Research on partially occluded foreign objects does not fully consider the influence of low illumination and dust fog. Constructed occluded samples are simple, the occlusion degree is low, and occluded object proportion is limited. Occlusion will cause a low detection accuracy of foreign objects. Therefore, considering the actual situation of coal mines, the foreign object target detection dataset with different occlusion degrees and categories in a low-illumination and dust fog environment will be constructed, and a fast and high-accuracy intelligent detection method for foreign objects considering complex occlusion cases will be proposed to solve the problem of the poor detection effect and weak real-time performance of occluded foreign objects.

2. Image Acquisition of Multi-Layer Occluded Foreign Objects

2.1. Experimental Platform and Parameters

As shown in Figure 1, the image acquisition platform of coal flow foreign objects of belt conveyors is composed of a belt conveyor, belt speed control computer, industrial camera, camera controller, light source, and so on. The parameters are shown in Table 1. The conveyor belt is 11 m long and 0.8 m wide. The belt speed control computer controls the motor to realize the speed regulation of the belt conveyor. An industrial area array camera is selected, and its image resolution is 1440 × 1080 and acquisition frame rate is 160~190 FPS. The vertical distance from the camera mirror to the belt is 0.81 m. The camera controller is connected through USB. The light source and camera are arranged directly above the detection area.
In the experiment, the basic model of target detection is YOLOv5s, the input image size is 1440 × 1080, the batch size is 16, the epochs are 300, the initial learning rate is 0.01, and the SGD is used to optimize the loss function. The momentum factor and weight attenuation factor are 0.937 and 0.0005. The hardware and software configuration used for the model training is shown in Table 2. In order to verify the effectiveness of the model compression method, the GPU and CPU are used to compare the model detection speed, and their corresponding parameters are shown in Table 3.

2.2. Image Data Construction

Taking gangues, anchor rods, boasts, wires, nuts, and sticks as the detection objects, the different degrees of occlusion are set up, and the occluded images considering illuminance and belt speed are collected. Due to the difficulty of creating occluded images and the inability to simulate all occlusion conditions, the non-occluded images are combined with black rectangles to form a multi-layer occluded dataset with 690 real occluded images, for a total of 1134 images. The number of images used for the training, verification, and test is 827, 207, and 100, respectively. The occlusion degree includes no occlusion (0%), general occlusion (1~35%), severe occlusion (35~80%), and complete occlusion (≥80%). The occlusion type includes intra-class occlusion and inter-class occlusion.
Table 4 shows clear images of real and synthetic occluded images, with illuminations of about 80, 185, and 280 Lux. In order to more clearly display the target in the 80-Lux image, after it is transferred from RGB space to HSV space, the brightness of the V channel is enhanced by 3 times, which is represented by the 80–V5 case in Table 4.
It can be seen from Table 4 that under 80 Lux, the small nuts, wires, gangues, and other targets above the coal in real occluded images have low visibility and are all interfered with by the background. The anchor rods in the middle area are blocked by gangues and sticks. The visual area of the anchor rods is smaller, which is divided into three sections, resulting in information loss and positioning difficulty. The three irregular wires blocked by gangues in the lower right corner are blocked by each other, which is difficult to detect. The six categories of targets in the synthetic occlusion case are also disturbed by the background. Adding black rectangular boxes increases the loss of target information, maintains the difficulty of occlusion detection, and enriches the occlusion type. For 185 Lux, the six categories of targets are similar. The two sticks on the right are mixed in the gangue pile, making their features unclear. About half of the information on the nut target in the middle is lost, and the learning feature is lower. The objects in the synthesized image are truncated by long rectangular boxes, which increases positioning difficulty. The images under 280 Lux are less disturbed by the background; it can be clearly seen that the occluded objects are cut into several segments, but most of the occluded targets also lose a lot of information.
Table 5 shows real and synthetic obscured images considering blur and dust fog. The belt speed for the blurred images is about 2 m/s. In order to more intuitively obtain the dust fog concentration, the brightness of the V channel of the dust fog images is enhanced by 3 times, which is represented by the Dust fog–V3 case in Table 5.
Figure 2 is a distribution map of all label sizes in the training set and verification set of multi-layer occluded coal flow foreign object images. The horizontal and vertical coordinates represent the absolute width and height of the label box. Figure 3 shows the number bar graph of the labeled categories in the training set and verification set. The horizontal and vertical coordinates represent the labeled categories and the corresponding number.
According to Figure 2 and Figure 3, it can be seen that a small number of anchor rods, sticks, gangues, boasts, wires, and all nuts in the occluded case are small targets; some nuts reach the minimum target, accounting for 12.21%. Combined with Table 4 and Table 5, it can be seen that due to occlusion, the actual width and height of most targets are far from the width and height of the 100~400-pixel size shown in Figure 2. The actual visualization area is small, but it is still larger than the visualization area of the nut.
In summary, the multi-layer occluded coal flow foreign object dataset constructed by using real and synthetic occlusion simulates a variety of cases. The number of data for each type of target is sufficient. It not only considers the factors such as illuminance, dust fog, and belt speed, but it also considers the occlusion degree and type, taking into account the dense degree at the same time. It can be seen that the constructed dataset has high complexity, a certain detection difficulty, and a more comprehensive design.

3. Detection Model of Multi-Layer Occluded Foreign Objects

3.1. YOLOv5s Network Model

Considering model complexity, detection accuracy, and speed, the 7.0 version of YOLOv5s is used as a benchmark model, and it is optimized from the following two aspects: accuracy improvement for occluded object detection and speed improvement for an embedded system’s deployment. The former is mainly improved from the perspective of the model back-end processing algorithm, label allocation strategy, and loss function, while the latter uses model pruning to speed up the detection process. YOLOv5s consists of the Input, Backbone, Neck, and Head, as shown in Figure 4.
Input: This consists of mosaic data enhancement, adaptive image scaling, and adaptive anchor frame calculation. The objective of mosaic data enhancement is to randomly crop and scale images and then randomly splice them into one image. Adaptive image scaling consists of scaling images to a standard size that is accepted by the network. Adaptive anchor frame calculation refers to a process where the network outputs the predicted frame based on an initial anchor frame and compares it with the real frame, calculates the difference between the two and updates it in reverse, and then adjusts the network parameters.
Backbone: The backbone network of YOLOv5s mainly includes the Conv, C3, and SPPF modules, which are used to extract input image features. Conv consists of a convolution layer (Conv2d), a batch normalization layer (BN), and a SiLU activation function. The convolution layer is used to extract local spatial information, the normalization layer is used to normalize feature distribution, and the activation function is used to achieve nonlinear mapping. C3 consists of three standard convolutions and multiple bottleneck structures (Bottleneck), which are used to increase the depth and receptive field of the network and improve feature extraction capability. The SPPF passes the input through multiple maximum pooling layers (MaxPool) with different kernel sizes in parallel, and then concatenation (Concat) is performed to fuse different features, which solves the multi–scale issue of detected targets.
Neck: This is composed of a feature pyramid network (FPN) and path aggregation network (PAN). The fine-grained spatial information of the shallow feature map and the semantic information of the deep feature map are fused from top to bottom through the FPN, and the low-level positioning information is easily transmitted upward from the bottom to the top through the PAN. The combination of the two strengthens the network feature fusion ability. Upsample is used to increase the size of the feature map.
Head: This is used to output the detection results of the targets. The CIoU loss function is used to optimize the accuracy of the bounding box. The regression of the target box is described from three aspects: overlapping area, center point distance, and aspect ratio.

3.2. Soft Non-Maximum Suppression

The non-maximum suppression (NMS) is an important part of the object detection model, which is used to suppress a large number of redundant candidate detection boxes. In the case of target aggregation, occlusion, etc., in the model, it is easy to set the confidence score of the candidate box whose overlap degree is greater than the predetermined threshold to 0, resulting in the missed detection of adjacent targets. For the detection of occluded foreign objects, Soft–NMS [28] is used to optimize the YOLOv5s model. The flow of the Soft–NMS algorithm is shown in Algorithm 1.
Algorithm 1: Soft–NMS (Non-maximum suppression)
Sensors 24 05251 i015
The Soft–NMS algorithm attenuates the confidence score of candidate boxes with an overlap greater than the threshold by the attenuation function instead of the direct suppression of the NMS algorithm, which can reduce the missed detection cases of the model. When the overlap of two boxes is greater, the degree of the confidence score attenuation of the corresponding candidate box is greater. The attenuation function includes a linear and a Gaussian weighted function, and the formulas are shown below.
s i = s i , i o u ( M , b i ) < N t s i ( 1 i o u ( M , b i ) ) , i o u ( M , b i ) N t
s i = s i e i o u ( M , b i ) σ 1 ,   b i D
where s i is the confidence score of the i–th candidate box, M is the candidate box with the current maximum confidence, b i is the i–th candidate box, i o u ( M , b i ) is the ratio of the intersection and union of the area of the candidate box M and b i , N t is the overlap threshold, D is the retained high-score detection box set, and the NMS and Soft–NMS overlap threshold uses a default value of 0.6. The value of the standard deviation σ 1 of the Gaussian function defaults to 0.5. The attenuation function in the Soft–NMS algorithm is the Gaussian weighted function.

3.3. SimOTA Label Allocation Optimization

Appropriate sample allocation can improve the model’s detection performance. YOLOv5 adopts a static sample allocation strategy based on shape matching. In an image with serious occlusion and dense targets, it is difficult to meet different detection scenarios by artificially setting hyperparameters, and it is easy to cause an imbalance between positive and negative samples. The optimal transport assignment (OTA) [29] strategy analyzes the sample allocation from a global perspective and converts it into an optimal transport (OT) problem, which reduces the negative impact of ambiguous samples and can adapt to multi-target overlaps and occlusion, small targets, etc. However, using the Sinkhorn–Knopp iterative method to solve the OT problem will lead to extra training time. The simplifying OTA (SimOTA) [30] method obtains an approximate solution of the OT problem through a dynamic top–k strategy, which can reduce the model training time.
The goal of the OT problem is to find an optimal transportation plan π * = { π i , j   | i = 1 , 2 , m ;   j = 1 , 2 , n } so that all goods from the provider can be transported to the demander at the minimum cost.
For a target detection task, the cost matrix C is composed of classification and regression loss. If there are m ground truth boxes (GT) and n anchor boxes (Anchor) in the image, each GT is regarded as a provider with k ( p i = k ,   i = 1 , 2 , m ) positive samples, and each Anchor that needs a unit sample is regarded as a demander. The cost of transporting a unit of positive samples from g t i to Anchor a j is defined as the weighted sum of classification and regression loss.
c i j f g = L c l s ( P j c l s , G i c l s ) + α L r e g ( P j b o x , G i b o x )
where P j c l s and P j b o x are the predicted classification score and bounding box, G i c l s and G i b o x are the GT classification score and bounding box, L c l s and L r e g are the cross-entropy loss and IoU loss, and α is the value of a balance coefficient; the default value is 3.
Because the Anchor can be assigned not only as a positive sample but also as a negative sample, another provider (background) is introduced to provide n m × k negative samples. The cost of transporting a unit of negative samples from the background to the Anchor is c i j b g .
c i j b g = L c l s ( P j c l s , )
where is the background.
The last row of C f g R m × n is spliced with C b g R 1 × n to form a cost matrix C of ( m + 1 ) × n . The updating provider vector formula is shown below.
p i = k , i f   i m n m × k , i f   i = m + 1
After obtaining the cost matrix C , it is necessary to determine the number dynamic_k of positive samples for each GT. The calculation method for the dynamic_k value consists of calculating the IoU between the current GT and all prediction boxes generated by the Anchor, selecting the top 10 largest samples, and then rounding down the sum of the 10 IoU values, which is the dynamic_k; the minimum number of positive samples is guaranteed to be 1. After obtaining the positive samples, the cost matrix is traversed, and the first dynamic_k samples with the smallest loss for each GT are taken as positive samples. Finally, the situation where one prediction frame is assigned to multiple GTs is eliminated, and the division of positive and negative samples is completed. A logic diagram of the SimOTA algorithm is shown in Figure 5.
Through the SimOTA label allocation strategy, more reasonable samples are allocated for dense, occluded, and small targets from a global perspective, reducing missed detection caused by ambiguous samples and improving the detection performance of the model.

3.4. Improvement of Loss Function

3.4.1. Slide Loss

For a single-stage detector, many bounding boxes are not iteratively filtered out, resulting in a large number of simple samples and sparse difficult samples, which leads to sample imbalance. During the training process, the cumulative contributions of simple samples dominate the updating of the model, resulting in model overfitting. YOLOv5 uses the cross-entropy loss function to calculate classification and target loss and uses CIoU to calculate the positioning loss. In order to excavate difficult samples and assign higher weights to difficult samples, Slide Loss [31] is used to optimize the cross-entropy function. Setting high weights near μ will increase the relative loss of difficult samples and focus more on difficult and misclassified samples. The formula for the Slide Loss is as follows.
f ( x ) = 1 , x μ 0.1 e 1 μ , μ 0.1 < x < μ e 1 x , x μ
where μ is the average value of IoU calculated by all bounding boxes, those less than μ are negative samples, and those greater than μ are positive samples.
A schematic diagram of Slide Loss is shown in Figure 6. The samples near the boundary often suffer large loss due to unclear classification. If the network can make full use of the difficult samples near the μ value, the detection performance of the model can be improved. However, the number of such samples is rare, so assigning high weights to such difficult samples increases the relative loss, making the network pay more attention to difficult and misclassified samples.

3.4.2. Inner–SIoU

The traditional IoU loss function depends on geometric factors such as the overlap area, distance, and aspect ratio between the prediction box and GT, but it does not consider the direction between them, resulting in a poor detection effect and it being unable to adaptively adjust. The original YOLOv5s adopts the CIoU loss function without considering the angle factor. This article uses SIoU with Inner–IoU [32] to build Inner–IoU to calculate IoU loss. Compared to the traditional loss function, Inner–SIoU not only considers the factors such as the overlapping area, distance, angle, and direction, but it also controls the scale of the auxiliary frame by adjusting the scale factor ratio to adapt to different IoU samples, improving the detection accuracy of small and occluded targets.
The formula of Inner–SIoU is shown in Equation (7).
L I n n e r S I o U = L S I o U + I o U I o U i n n e r = 1 I o U + Δ + Ω 2 + I o U I o U i n n e r = 1 I o U i n n e r + Δ + Ω 2
where I o U i n n e r is the Inner–IoU value, L S I o U is the SIoU loss, Δ and Ω are the distance and shape loss in SIoU, and Δ and Ω are shown in Equations (15) and (24). A schematic diagram of Inner–IoU is shown in Figure 7, the formula of Inner–IoU is shown in Equation (8), and a schematic diagram of SIoU is shown in Figure 8.
I o U i n n e r = I n n e r i n t e r I n n e r u n i o n
I n n e r i n t e r = min b r g t , b r max b l g t , b l × min b r g t , b r max b l g t , b l
I n n e r u n i o n = w g t × h g t × r a t i o 2 + w × h × r a t i o 2 I n n e r i n t e r
b l g t = x c g t w g t × r a t i o 2 ,   b r g t = x c g t + w g t × r a t i o 2
b t g t = y c g t h g t × r a t i o 2 ,   b b g t = y c g t + h g t × r a t i o 2
b l = x c w × r a t i o 2 ,   b r = x c + w × r a t i o 2
b t = y c h × r a t i o 2 ,   b b = y c + h × r a t i o 2
where I n n e r i n t e r and I n n e r u n i o n are the intersection and union of the auxiliary bounding box of the target ground truth box (Inner ground truth box) and the auxiliary bounding box of the anchor box (Inner anchor box), b l g t , b r g t , b t g t , b b g t , and b l , b r , b t , and b b are the horizontal and vertical coordinates of the left, right, upper, and lower sides of the Inner ground truth box and Inner anchor box. ( x c g t , y c g t ), w g t , h g t , and ( x c , y c ), w , and h are the midpoint coordinates, width, and height of the GT and Anchor. The r a t i o is the scale factor that controls the width and height of the auxiliary bounding box, and its value range is [0.5, 1.5]. When r a t i o < 1 , the auxiliary bounding box is shown on the left side of Figure 4. When r a t i o > 1 , the auxiliary bounding box is shown on the right side of Figure 4. When r a t i o = 1 , the auxiliary bounding box is consistent with the GT and Anchor. When r a t i o = 1.10 , a better balance can be achieved between the different categories of targets, and the detection effect is the best. The larger auxiliary frame can not only include more information, facilitating the utilization of contextual information, but also reduce error fluctuations caused by pixel deviation and improve model detection accuracy.
Δ = t = x , y 1 e γ ρ t
ρ x = d w c w 2 , ρ y = d h c h 2 , γ = 2 Λ
d w = max x c g t , x c min x c g t , x c
d h = max y c g t , y c min y c g t , y c
c w = max x c g t + w g t 2 , x c + w 2 min x c g t w g t 2 , x c w 2
c h = max y c g t + h g t 2 , y c + h 2 min y c g t h g t 2 , y c h 2
Λ = 1 2 × sin 2 arcsin x π 4 = sin 2 α
x = d h σ 2 = sin α
σ 2 = d h 2 + d w 2
where d w and d h are the horizontal and vertical distances between the centers of GT and Anchor, c w and c h are the width and height of the minimum external rectangular frame of GT and Anchor, Λ is the angle cost, and its value range is [0, 1]; α and β are the angles between the line connecting the center point of GT and Anchor and the horizontal axis and plumb axis, and σ 2 is the distance between GT and the Anchor center point.
Ω = t = w , h 1 e w t θ
ω w = w w g t max w , w g t ,   ω h = h h g t max h , h g t
where θ is the degree of the attention to shape loss, its value range is [2, 6], and the default value is 4.
In the case of limited samples, Slide Loss can fully excavate difficult sample information and improve information utilization. Inner–SIoU not only considers the angle factor but also uses auxiliary border training to adapt to IoU samples of different scale sizes, which can improve the detection accuracy of small and occluded targets.

3.5. Model Pruning

The model compression method can be used to remove redundant parameters in a network, reduce the model complexity, and improve the detection speed. Group–Taylor [33] channel pruning in the structured pruning method can be used to prune the model. The square of the final loss change caused by deleting a single parameter is used as the parameter importance evaluation criterion, and the first-order Taylor expansion is used to approximately estimate the importance of a single parameter. The least important parameters are deleted until the preset global pruning rate is satisfied. Finally, fine-tuning is performed to partially restore the model accuracy. The formula for calculating the importance of a single parameter is shown in Equation (26).
I z ( 1 ) ( W ) = g z w z 2
g z = E w z
where W is the neural network parameter, W = { w 0 , w 1 , , w z } , g z is the gradient of the parameter, z is a positive integer with a value greater than 0, and E is the loss of the network.
The formula of the joint importance of a set of structural parameters W S P is shown in Equation (28).
I S P ( 1 ) ( W S P ) s p S P I S P ( 1 ) ( W S P ) = s p S P ( g s p w s p ) 2
where S P is a non-empty set composed of all parameters in the convolution kernel and other structures, and s p is the element in the set S P .
By using the first-order Taylor expansion to approximate the importance of parameters, the model is pruned without the layer-by-layer parameter sensitivity analysis, and the gradient in the back-propagation process can be used in parallel to reduce the computational complexity and improve the detection speed.

4. Experimental Results and Analysis

4.1. Model Evaluation Indexes

The mean Average Precision (mAP) with IOU thresholds of 0.5, 0.75, and 0.5:0.95 is used to evaluate the detection accuracy of the model, which is recorded as mAP0.5, mAP0.75, and mAP0.5:0.95, respectively. The model complexity is evaluated by parameters, the model size, and floating point operations per second (FLOPs). The detection speed is measured by frames per second (FPS). The ratio of the number of parameters before and after model pruning is used to measure the model compression degree. The calculation of the mAP is as follows.
P = TP TP + FP × 100 %
R = TP TP + FN × 100 %
AP = 0 1 P ( R ) d ( R )
m AP = i N A P i N × 100 %
C D = P a r a m e t e r s a P a r a m e t e r s b
where TP, FP, and FN are the number of correct detection boxes, false detection boxes, and missed detection boxes, N is the number of categories, P and R are Precision and Recall, and AP is Average Precision (AP). P a r a m e t e r s a and P a r a m e t e r s b are the parameters of the model before and after pruning, and C D is the compression degree of the model.

4.2. Ablation Experiment of Accuracy Improvement Model

In order to verify the effectiveness of the Soft–NMS algorithm, SimOTA label allocation strategy, Slide Loss, and Inner–SIoU function, the ablation experiment is carried out under the same conditions on the above self-made occluded dataset based on the 7.0 version of YOLOv5s, and the influence of each improvement point on the model performance is evaluated. The results of the ablation experiment are shown in Table 6. Since this work mainly solves the problem of target occlusion from the loss function and does not modify the network structure, the accuracy of the model is mainly discussed here.
It can be seen from Table 6 that after the back-end processing algorithm is changed to Soft–NMS in Model 2, mAP0.75 and mAP0.5:0.95 increase greatly, both reaching 3.5%, and mAP0.5 increases by 0.3%, which proves that Soft–NMS plays a large role in dealing with dense and occluded targets. Based on Model 2, Model 3 enhances label the allocation strategy to SimOTA, and mAP0.5, mAP0.75, and mAP0.5:0.95 have great enhancement, with increases of 1.1%, 1.9%, and 2.1%, which shows that SimOTA has a significant effect on dense and occluded targets. On the basis of Model 3, Model 4 uses Slide Loss, mAP0.5 and mAP0.75 are increased by 0.6% and 0.5%, and mAP0.5:0.95 is reduced by 0.1%, which indicates normal fluctuation. On the basis of Model 4, Inner–SIoU is used in Model 5. When the growth potential of the model is about to reach its peak, mAP0.5, mAP0.75, and mAP0.5:0.95 are increased by 0.1%, 0.3%, and 0.3%. In summary, introducing the four modules in Table 6 is effective. The mAP0.5, mAP0.75, and mAP0.5:0.95 of the enhanced model are 2.1%, 6.2%, and 5.8% higher than the original YOLOv5s.
In order to evaluate the effect more objectively, the AP line charts of the six categories of targets are shown in Figure 9.
It can be seen from Figure 9 that the AP0.5, AP0.75, and AP0.5:0.95 of the six categories of targets are generally rising. The AP0.5 polylines of nuts, gangues, and wires are relatively flat, and the changes in the remaining targets are obvious. Aside from the fact that the polylines of the AP0.75 and AP0.5:0.95 of gangues are relatively flat, the AP0.75 and AP0.5:0.95 of other targets are greatly improved. Compared to AP0.75 and AP0.5:0.95, AP0.5 has lower positioning accuracy requirements. When the overlap between the prediction frame and real frame reaches 50%, the target is considered to be detected. The same prediction frame is more likely to be judged as the correct detection frame. The more frames that the model determines to be correctly detected, the higher the AP is, but the corresponding AP0.5 improvement is limited. On the basis of Model 1, Soft–NMS is used. The AP0.75 and AP0.5:0.95 of the six categories of targets change greatly; the increases are above 2% and 2.3%, respectively. Among these, the AP0.75 and AP0.5:0.95 of the nuts increase by 4.8% and 3.6%, which verifies that Soft–NMS has a significant effect on dealing with small and occluded targets. Then, SimOTA is changed in Model 2. The AP0.75 and AP0.5:0.95 of the gangues have the smallest increase, but the stick has the largest increase; both indicators increase by 2.9%. The increases in the remaining targets are still large, which verifies the strength of SimOTA in dealing with small and occluded targets. After adding Slide Loss to Model 3, the improvements in the six categories of targets increase and decrease, but only the nuts are decreased by 0.3% on AP0.5, and the rest are improved, with the highest increase of 2.9%. For AP0.5:0.95, it is mostly decreased, and anchor rods and wires are increased by 0.5% and 0.6%, which still indicates the effectiveness of Slide Loss. Finally, the addition of Inner–SIoU has a negative impact on the detection of anchor rods. Its AP0.5 decreases by 2.3%, and AP0.5:0.95 decreases by 2.1%. However, it has a gain effect on other targets. The gain effect of boasts is the most obvious, reaching 2.4% and 2.2%. The nuts also have an increase of 1.1% and 0.6%, which proves that Inner–SIoU can deal with different scale and shape targets.

4.3. Comparative Experiment of Model Compression Method

Model compression is studied while maintaining mAP0.5 > 90%. Based on the above Model 5, the main pruning methods used for structured pruning are compared based on model accuracy, complexity, and speed. Table 7 shows the results of different pruning methods, ‘+Random’ indicates that the Random pruning method is used on the basis of Model 5, and the compression degree is the numerical calculation result of parameters in the table.
It can be seen from Table 7 that the Group–Taylor and Group–Hessian pruning methods are significantly better than the Random, Lamp, and L1 methods. When the parameters, FLOPs, and model size of the first two methods are smaller than the latter three methods, mAP0.5 is still higher than the latter three methods. The higher the sparsity of the model, the more serious the accuracy loss. After the model parameters are compressed to about 1/17~1/14 of the original, the mAP0.5, mAP0.75, and mAP0.5:0.95 indexes are greatly reduced. After using the Group–Hessian and Group–Taylor methods, the parameters, FLOPs, and model size are reduced by 94.03%, 93.67%, and 91.67%, but mAP0.5 still maintains 91.30%.
In order to more objectively show the difference in the detection accuracy of five pruning methods under different compression degrees, a line chart of mAP0.5, mAP0.75, and mAP0.5:0.95 is drawn, as shown in Figure 10. On the mAP0.5 chart, the accuracy values of the Group–Taylor and Group–Hessian methods are significantly better than those of the other three methods, and in particular, when the compression degree is close to 17, the difference is obvious. In the mAP0.75 and mAP0.5:0.95 chart, the accuracy values of the L1, Group–Taylor, and Group–Hessian methods are basically higher than those of the Random and Lamp methods. From the three charts, the Group–Taylor and Group–Hessian pruning methods are better.
In order to verify the superiority of the Group–Taylor method, the GPU and CPU devices are used to compare the detection speed of different pruning methods in Table 7. The results are shown in Table 8. The numerical value is the mean value of 100 images for five tests, and the batch size is uniformly 1.
It can be seen from Table 8 that on RTX 4080, the detection speed of the model before and after pruning is comparable. The five pruning methods can reach more than 100 FPS, but the difference is not obvious. On GTX 1050Ti, there is a significant difference in the model speed after and before pruning. The speed after pruning is 60 FPS, while the Group–Taylor method is the fastest, up to 66.31 FPS, which is 29.48 FPS higher than Model 5. The speed of Model 5 on GT 1030 is only about 40% of GTX 1050Ti. The model speed after pruning is about 2.58 to 2.90 times that before pruning; Group–Taylor is still the fastest. On Intel (R) Core (TM) i5–11400F, the Group–Taylor method is 4.32 times faster than the speed of Model 5, which meets the real–time processing requirements. From Table 7 and Table 8 and Figure 10, the superiority of the Group–Taylor method is verified.
Figure 11 shows a comparison histogram of the number of convolutional layer channels in the Backbone, Neck, and Head before and after pruning, using the Group–Taylor method with a compression degree of 16.74 in the enhanced Model 5. In Figure 11a,b, the missed 11, 15,12, 16, 19, and 22 in the abscissa represent the Upsample and Concat modules that do not contain the convolution layer. The rest of the numbers represent the number of network layers, which corresponds to the number of layers in Figure 1. The number 24 represents the detection layer structure without pruning, and the number of channels in the three detection layers is 33. The ordinate represents the number of channels. Aside from the detection layer structure represented by 24, the number of channels before and after pruning in Figure 11 is arranged from low to high, i.e., [32, 64, 128, 256, 512], [7, 15, 31, 62, 125], and the numerical values correspond to the height of the histogram. From Figure 11, the number of channels in the 0~23 layers of the convolutional module after pruning is less than one-quarter of that before pruning. For example, the number of channels in the convolutional layer with 512 channels is 125 after pruning. Thus, the number of channels is greatly reduced, and the model is more compact.
The Group–Taylor pruning method with a compression degree of 16.74 is used to prune the improved model after optimizing the post-processing, label allocation strategy, Slide Loss, and Inner–SIoU loss function. For the high-accuracy detection model with only 4.20 × 105 parameters, 1.00 × 109 FLOPs, and 1.20 MB, an mAP0.5 of 91.30% is obtained, which is that of the final model. The model has detection speeds of 66.31, 41.90, and 33.03 FPS on the GTX 1050Ti, GT 1030, and CPU devices, which are far beyond those of the YOLOv5s, and achieves high real–time processing speed.

4.4. Comparative Experiment of Different Detection Algorithms

Under the same experimental conditions, YOLOv5n, YOLOv5n–Group–Taylor, YOLOv5s, YOLOv5m, YOLOv7, YOLOv7–X, YOLOv8n, YOLOv8s, and YOLOv8m are selected for comparison with the proposed method. Among them, YOLOv5n–Group–Taylor represents the pruning of YOLOv5n by the Group–Taylor method, which still needs to meet mAP0.5 > 90%. mAP0.5 is used as the accuracy evaluation index, and the model parameters, FLOPs, and size are used as the complexity evaluation indexes. The FPS calculated on the CPU is used to evaluate the detection speed. The results are shown in Table 9.
From Table 9, it can be seen that YOLOv5s can better balance the detection accuracy and speed in different versions. Using YOLOv7, the accuracy improvement is limited, and the FLOPs are as high as 103.20 × 109, which is much higher than other algorithms. The mAP0.5 of YOLOv7–X is only 0.1% higher than that of YOLOv7, but its parameters, FLOPs, and model size are 1.94, 1.82, and 1.90 times than those of YOLOv7, respectively, and its detection speed is slower. YOLOv8m has the highest accuracy among these algorithms, but its detection speed is very low. Also using Group–Taylor pruning, the accuracy, complexity, and speed of YOLOv5n–Group–Taylor are not as good as the proposed method. Although the mAP0.5 of the proposed method is the lowest, it is still higher than 90%, and the parameters, FLOPs, and model size are far lower than other algorithms. More importantly, its detection speed is the fastest of the above algorithms. It is verified that the proposed method can realize the fast detection of occluded objects with high accuracy and low complexity on PC devices.

4.5. Visual Analysis of Detection Results in Different Scenarios

In order to test the detection effect of the proposed method in a low-illumination environment, real and synthetic images under different illumination cases are tested. The visual results are shown in Figure 12.
From Figure 12, it can be seen that in the 80-Lux real occluded image (a), due to too low illumination, irregular wires blocked by the gangues have low visibility and a small visual area. In addition, due to the influence of nearby wires, the model misses a wire target. In the middle of the image, because the boast is blocked by the gangue and anchor rod, its contour is unclear, and there are multiple detection times and inaccurate positioning. On the left side of the image, two sticks are crossed, one stick is cut off by the coal and nut, which leads to a big positioning error when the model detects the other sticks, and the prediction frame is obviously larger. In the 185-Lux real occluded image (c), the influence of illumination on the occluded targets is significantly reduced. For the right long stick, it is truncated and occluded, resulting in repeated detection, the positioning error of the prediction box is big, and the confidence of the relatively accurate prediction box is only 0.64. In the 280-Lux real occluded image (e), it can be seen that the long stick on the left side is cut into several sections by gangues. The lower part of the stick is similar to the gangue, and the prediction frame is not positioned accurately. Most of the occluded targets in the real occluded images (a), (c), and (e) are effectively detected. Compared with the real occluded images, most targets in the 80-, 185-, and 280-Lux synthetic images (b), (d), and (f) have better detection results and higher confidence. It can be seen that the low-illumination environment has an impact on the detection of occluded targets, and this impact gradually decreases as the illumination increases.
Figure 13 shows the visual results of the proposed method on the real and synthetic occluded images in a low-blur and high-dust fog environment.
Figure 13a,b were collected at a belt speed of 2 m/s. Subjectively, the overall ambiguity of the image is low, and the impact on large targets is limited. In Figure 13a, the small nut target on the right has low discrimination with the gangue contour, and the detection confidence of the nut is only 0.74, which affects the overall detection effect. Due to the occlusion of multiple pieces of gangues, the long stick on the right side is cut into multiple segments, and the model attention to the boundary is relatively insufficient, resulting in the poor positioning of the stick and obvious excess blanks at the upper end of the prediction frame, which does not contain the target component. The targets in Figure 13b are relatively sparse, and the overall detection effect is good. The dust fog images (c) and (d) were collected under a certain concentration. The concentration of dust fog is limited and has relatively little impact on the targets. The anchor rod in the middle of Figure 13c is obscured by the gangues and dust fog, resulting in unclear boundaries. The prediction frame does not fully include the anchor rod, and the detection effect is poor, with a confidence of only 0.52. The overall detection effect of the targets in Figure 13d is good. It can be seen that the factors such as blur and dust fog increase the difficulty of target detection, but the proposed method effectively detects most targets.
For the synthetic occluded images in Figure 12 and Figure 13, the occlusion type is background occlusion, there is a certain difference in color between the black rectangular block and the targets, and the targets are relatively sparse. The model can learn the difference between the two very well. However, occlusion results in there being a small learnable area, which affects the target detection effect. In terms of the sparseness, the detection effect of targets in a dense occlusion is much lower than that of targets in a sparse occlusion, which shows that the density of occluded targets greatly affects the detection effect.
Judging from the detection effect of each category of targets, small nut targets are seriously affected by occlusion, and missed detection, repeated detection, and poor positioning are mainly due to the targets being obscured by other targets, resulting in a large amount of image information being lost.
In summary, it can be seen that the object detection method proposed for low-illumination, blur, dust fog, and dense occluded scenarios can identify targets in these scenarios fast and be highly accurate.

5. Conclusions

Aiming to solve the coal flow foreign object detection problem in a coal mine with a low-illumination and dust fog environment, this article proposes a fast and high-accuracy intelligent detection method capable of dealing with multi-layer body occlusion cases for the monitored images of an operating belt conveyor.
(1)
For the coal flow foreign object occluded image dataset’s construction, the target occlusion degree, occlusion type, and dense degree are comprehensively considered, and the synthetic occlusion is used. The synthetic occluded images and the real occluded images together constitute the coal flow foreign object occluded image dataset in a low-illumination and dust fog environment.
(2)
The YOLOv5 algorithm is improved by the Soft–NMS, SimOTA, Slide Loss, and Inner–SIoU modules. The problem of missed and false detection caused by dense and occluded targets is solved, and the detection accuracy of the model used for small and occluded targets is improved. The mAP0.5 of the accuracy improvement model reaches the detection level of YOLOv8s, which is higher than YOLOv5n, YOLOv5s, YOLOv5m, YOLOv7, YOLOv7–X, and YOLOv8n.
(3)
The accuracy improvement model is pruned through the Group–Taylor method. After pruning, the mAP0.5 of the model is greater than 90%, and its model parameters, FLOPs, and size are also greatly reduced. The detection speed is faster on the GT 1050Ti, GT 1030, and Intel (R) Core (TM) i5–11400F processors, which meets the real-time processing speed requirements.
(4)
The mAP0.5 of the proposed method on the self-built multi-layer occluded coal flow foreign object dataset is 91.30%, and the mAP0.5 is lower than YOLOv5n, YOLOv5s, YOLOv5m, YOLOv7, YOLOv7–X, YOLOv8n, YOLOv8s, and YOLOv8m. However, both the model’s lightweight level and detection speed are superior to the above, and the detection effect of most occluded targets is good in various scenarios.

Author Contributions

Conceptualization, H.F. and J.L.; Methodology, J.L.; Software, J.L. and X.Y.; Validation, J.L. and C.Z.; Formal analysis, J.L.; Investigation, J.L. and H.F.; Resources, H.F., X.C. and Q.M.; Data curation, J.L.; Writing—original draft preparation, J.L.; Writing—review and editing, J.L. and H.F.; Visualization, J.L.; Supervision, H.F.; Project administration, H.F. and Q.M.; Funding acquisition, H.F. and Q.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Special Project of Shaanxi Provincial Department of Education for Local Service under Grant 23JC052, the “Scientist & Engineer” Team Construction Project of Shaanxi Qin Chuang Yuan under Grant 2024QCY–KXJ–164, and the Shaanxi University Youth Innovation Team Project under Grant 23JP096.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in the current study is available from the corresponding author (correspondence: [email protected]) on reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chen, X.J. Current status and development trend of intelligent technology of underground coal mine transportation system. J. Mine Autom. 2022, 48, 6–14+35. [Google Scholar] [CrossRef]
  2. Wang, H.J.; Wang, H.L. Status and prospect of intelligent key technologies of belt conveyor. Coal Sci. Technol. 2022, 50, 225–239. [Google Scholar] [CrossRef]
  3. Hao, S.; Zhang, X.; Ma, X.; Sun, S.Y.; Wen, H.; Wang, J.L. Foreign object detection in coal mine conveyor belt based on CBAM-YOLOv5. J. China Coal Soc. 2022, 47, 4147–4156. [Google Scholar] [CrossRef]
  4. Ge, S.R.; Hao, S.Q.; Zhang, S.H.; Zhang, X.F.; Zhang, L.; Wang, S.B.; Wang, Z.B.; Bao, J.S.; Yang, X.L.; Yang, J.J. Status of intelligent coal mining technology and potential key technologies in China. Coal Sci. Technol. 2020, 48, 28–46. [Google Scholar] [CrossRef]
  5. Wu, S.P. Rescarch on Detcction Method of Foreign Object on Coal Conveyor Belt Based on Computer Vision. Master’s Thesis, China University of Mining and Technology, Xuzhou, China, 2019; pp. 1–60. [Google Scholar]
  6. Wang, J.C.; Li, L.H.; Yang, S.L. Experimental study on gray and texture features extraction of coal and gangue image under different illuminance. J. China Coal Soc. 2018, 43, 3051–3061. [Google Scholar] [CrossRef]
  7. Li, B.; Wang, X.W.; Pang, S.Z.; Gao, X.Y.; Wang, L.Y.; Ding, E.F.; Bao, Q.B. Image characteristics analysis and experimental study of coal and gangue. Coal Sci. Technol. 2022, 50, 236–246. [Google Scholar] [CrossRef]
  8. Cao, X.G.; Hao, P.Y.; Wang, P.; Liu, S.Y.; Wu, X.D.; Huo, X.Q. Research on method of acquiring high quality coal gangue images under multi-factor illumination condition. Coal Sci. Technol. 2023, 51, 455–463. [Google Scholar] [CrossRef]
  9. Li, M.; Yang, M.L.; Liu, C.Y.; He, X.L.; Duan, Y. Illuminance adjustment method for image-based coal and gangue separation. J. China Coal Soc. 2021, 46, 1149–1158. [Google Scholar] [CrossRef]
  10. Mao, Q.H.; Li, S.K.; Hu, X.; Xue, X.S. Coal mine belt conveyor foreign objects recognition method of improved YOLOv5 algorithm with defogging and deblurring. Energies 2022, 15, 9504. [Google Scholar] [CrossRef]
  11. Fan, H.W.; Liu, J.P.; Cao, X.G.; Zhang, C.; Zhang, X.H.; Li, M.; Ma, H.W.; Mao, Q.H. Multi-scale target intelligent detection method for coal, foreign object and early damage of conveyor belt surface under low illumination and dust fog. J. China Coal Soc. 2024, 1–12. [Google Scholar] [CrossRef]
  12. Fan, H.W.; Zhang, C.; Cao, X.G.; Liu, J.P.; Zhang, X.H.; Zhao, H. An image dust removal and enhancement method in low illumination environment based on dark-bright channel segmentation and fusion. J. China Coal Soc. 2024, 49, 2167–2178. [Google Scholar] [CrossRef]
  13. Gao, Y.S.; Zhang, B.Q.; Lang, L.Y. Coal and gangue recognition technology and implementation based on deep learning. Coal Sci. Technol. 2021, 49, 202–208. [Google Scholar] [CrossRef]
  14. Xu, Z.Q.; Lv, Z.Q.; Wang, W.D.; Zhang, K.H.; Lv, H.M. Machine vision recognition method and optimization for intelligent separation of coal and gangue. J. China Coal Soc. 2020, 45, 2207–2216. [Google Scholar] [CrossRef]
  15. Guo, Y.C.; Wang, X.; He, L.; Liu, P.Z. Research on coal and gangue recognition method based on TW-RN optimized CNN. Coal Sci. Technol. 2022, 50, 228–236. [Google Scholar]
  16. Wu, S.P.; Ding, E.J.; Yu, X. Foreign body identification of belt based on improved FPN. Saf. Coal Mines 2019, 50, 127–130. [Google Scholar] [CrossRef]
  17. Lv, Z.Q. Research on Foreign Body Image Recognition of Coal Mine Belt Transport under Complex Environment. Master’s Thesis, China University of Mining and Technology, Xuzhou, China, 2020; pp. 1–60. [Google Scholar]
  18. Hu, J.H.; Gao, Y.; Zhang, H.J.; Jin, B.Q. Research on the identification method of non-coal foreign object of belt conveyor based on deep learning. Ind. Mine Autom. 2021, 47, 57–62, 90. [Google Scholar] [CrossRef]
  19. Ren, Z.L.; Zhu, Y.C. Research on foreign objects recognition of coal mine belt transportation with improved CenterNet algorithm. Control Eng. China 2023, 30, 703–711. [Google Scholar] [CrossRef]
  20. Wang, Y.B.; Wang, Y.J.; Dang, L.F. Video detection of foreign objects on the surface of belt conveyor underground coal mine 12 based on improved SSD. J. Ambient Intell. Humaniz. Comput. 2023, 14, 5507–5516. [Google Scholar] [CrossRef]
  21. Chen, D.Q.; Xu, J.Y.; Kou, Q.Q.; Zhang, H.W.; Han, C.G.; Yu, B.; Qian, J.S. Lightweight network based on residual information for foreign body classification on coal conveyor belt. J. China Coal Soc. 2022, 47, 1361–1369. [Google Scholar] [CrossRef]
  22. Chen, Z. Research on Foreign Objects Detection of Mine Belt Conveyor Based on Deep Learning. Ph.D. Thesis, China University of Mining and Technology, Xuzhou, China, 2023; pp. 1–59. [Google Scholar]
  23. Ding, Z.H.; Chen, G.D.; Wang, Z.; Chi, W.Z.; Wang, Z.H.; Fan, Y.L. A real-time multilevel fusion recognition system for coal and gangue based on near-infrared sensing. IEEE Access 2020, 8, 178722–178732. [Google Scholar] [CrossRef]
  24. Li, D.Y.; Ren, H.W.; Wang, G.F.; Wang, S.; Wang, W.S.; Du, M. Coal gangue detection and recognition method based on multiscale fusion lightweight network SMS-YOLOv3. Energy Sci. Eng. 2023, 11, 1783–1797. [Google Scholar] [CrossRef]
  25. Zhang, M.C.; Cao, Y.H.; Jiang, K.; Li, M.X.; Liu, L.X.; Yu, Y.; Zhou, M.S.; Zhang, Y. Proactive measures to prevent conveyor belt Failures: Deep Learning-based faster foreign object detection. Eng. Fail. Anal. 2022, 141, 106653. [Google Scholar] [CrossRef]
  26. Chen, Y.M.; Sun, X.; Xu, L.; Ma, S.C.; Li, J.; Pang, Y.S.; Chen, G. Application of YOLOv4 algorithm for foreign object detection on a belt conveyor in a low-illumination environment. Sensors 2022, 22, 6851. [Google Scholar] [CrossRef] [PubMed]
  27. Yao, R.B.; Qi, P.; Hua, D.Z.; Zhang, X.; Lu, H.; Liu, X.H. A Foreign Object Detection Method for Belt Conveyors Based on an Improved YOLOX Model. Technologies 2023, 11, 114. [Google Scholar] [CrossRef]
  28. Bodla, N.; Singh, B.; Chellappa, R.; Davis, L.S. Soft-NMS improving object detection with one line of code. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5561–5569. [Google Scholar] [CrossRef]
  29. Ge, Z.; Liu, S.T.; Li, Z.M.; Yoshie, Q.; Sun, J. Ota: Optimal transport assignment for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 303–312. [Google Scholar] [CrossRef]
  30. Ge, Z.; Liu, S.T.; Wang, F.; Li, Z.M.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar] [CrossRef]
  31. Yu, Z.P.; Huang, H.B.; Chen, W.J.; Su, Y.X.; Liu, Y.H.; Wang, X.Y. Yolo-facev2: A scale and occlusion aware face detector. arXiv 2022, arXiv:2208.02019. [Google Scholar] [CrossRef]
  32. Zhang, H.; Xu, C.; Zhang, S.J. Inner-iou: More effective intersection over union loss with auxiliary bounding box. arXiv 2023, arXiv:2311.02877. [Google Scholar] [CrossRef]
  33. Molchanov, P.; Mallya, A.; Tyree, S.; Frosio, I.; Kautz, J. Importance estimation for neural network pruning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11264–11272. [Google Scholar] [CrossRef]
  34. Torch-Pruning. Available online: https://github.com/VainF/Torch-Pruning (accessed on 8 April 2024).
  35. Lee, J.; Park, S.; Mo, S.W.; Ahn, S.S.; Shin, J.W. Layer-adaptive sparsity for the magnitude-based pruning. arXiv 2021, arXiv:2010.07611. [Google Scholar] [CrossRef]
  36. Li, H.; Kadav, A.; Durdanovic, I.; Samet, H.; Graf, H.P. Pruning filters for efficient convnets. arXiv 2017, arXiv:1608.08710. [Google Scholar] [CrossRef]
  37. Yann, L.C.; John, S.D.; Sara, A.S. Optimal brain damage. Adv. Neural Inf. Process. Syst. 1989, 2, 598–605. [Google Scholar]
Figure 1. Image acquisition platform of coal flow foreign objects.
Figure 1. Image acquisition platform of coal flow foreign objects.
Sensors 24 05251 g001
Figure 2. Absolute scale of foreign object.
Figure 2. Absolute scale of foreign object.
Sensors 24 05251 g002
Figure 3. Foreign object number bar chart.
Figure 3. Foreign object number bar chart.
Sensors 24 05251 g003
Figure 4. Foreign object width and height distribution map.
Figure 4. Foreign object width and height distribution map.
Sensors 24 05251 g004
Figure 5. Logic flow of SimOTA.
Figure 5. Logic flow of SimOTA.
Sensors 24 05251 g005
Figure 6. Schematic diagram of Slide Loss.
Figure 6. Schematic diagram of Slide Loss.
Sensors 24 05251 g006
Figure 7. Schematic diagram of Inner–IoU.
Figure 7. Schematic diagram of Inner–IoU.
Sensors 24 05251 g007
Figure 8. Schematic diagram of SIoU.
Figure 8. Schematic diagram of SIoU.
Sensors 24 05251 g008
Figure 9. Enhancement effect of each improvement module on different categories: (a) nut; (b) anchor rod; (c) stick; (d) anchor rod; (e) boast; and (f) wire.
Figure 9. Enhancement effect of each improvement module on different categories: (a) nut; (b) anchor rod; (c) stick; (d) anchor rod; (e) boast; and (f) wire.
Sensors 24 05251 g009aSensors 24 05251 g009b
Figure 10. Model compression–accuracy line chart. (a) mAP0.5; (b) mAP0.75; (c) mAP0.5:0.95.
Figure 10. Model compression–accuracy line chart. (a) mAP0.5; (b) mAP0.75; (c) mAP0.5:0.95.
Sensors 24 05251 g010
Figure 11. Comparison of number of convolution channels before and after pruning. (a) Backbone; (b) Neck + Head.
Figure 11. Comparison of number of convolution channels before and after pruning. (a) Backbone; (b) Neck + Head.
Sensors 24 05251 g011
Figure 12. Image detection results under different illumination values: (a) 80-Lux real occluded image; (b) 80-Lux synthetic occluded image; (c) 185-Lux real occluded image; (d) 185-Lux synthetic occluded image; (e) 280-Lux real occluded image; and (f) 280-Lux synthetic occluded image.
Figure 12. Image detection results under different illumination values: (a) 80-Lux real occluded image; (b) 80-Lux synthetic occluded image; (c) 185-Lux real occluded image; (d) 185-Lux synthetic occluded image; (e) 280-Lux real occluded image; and (f) 280-Lux synthetic occluded image.
Sensors 24 05251 g012
Figure 13. Occluded blur and dust fog image detection results. (a) Real occluded blur image; (b) synthetic occluded blur image; (c) real occluded dust fog image; (d) synthetic occluded dust fog image.
Figure 13. Occluded blur and dust fog image detection results. (a) Real occluded blur image; (b) synthetic occluded blur image; (c) real occluded dust fog image; (d) synthetic occluded dust fog image.
Sensors 24 05251 g013
Table 1. Device parameters of image acquisition platform.
Table 1. Device parameters of image acquisition platform.
DeviceParameters
Belt conveyorBelt length: 11 m, belt width: 0.8 m, motor frequency range: 0~50 Hz
Belt speed control computerOperating system: Windows XP, MotorControl software 1.0
Industry cameraHikvision MV–CA016–10UC–USB3.0 area array camera
ShotHFO420M, focal length: 4 mm
Camera controllerOperating system: Windows 10, Hikvision client software MVS 4.3.1
Light sourceLED strip lamp: 50 cm long, white light source, manually adjusted
Illuminometer31/2-bit LCD display, measuring range: 0~200,000 Lux, manually adjusted
Table 2. Configuration of model training platform.
Table 2. Configuration of model training platform.
PlatformParameters
Operating systemWindows 10
GPUNVIDIA GeForce RTX 4080, Memory size: 16 GB
CPUIntel(R) Core(TM)i7–13700KF 3.40GHz
Deep learning frameworkPytorch 2.0.0, CUDA 11.8.0
Table 3. Configuration of model speed test platform.
Table 3. Configuration of model speed test platform.
PlatformParameters
GPU0NVIDIA GeForce RTX 4080, Pytorch 2.0.0, CUDA 11.8.0
GPU1GTX 1050Ti, Pytorch 1.10.1, CUDA 10.2.89
GPU2GT 1030, Pytorch 1.10.1, CUDA 12.2.89
CPU3Intel(R) Core(TM)i5–11400F 2.60 GHz, Pytorch 1.10.1
Table 4. Real and synthetic occluded images without blur and dust fog.
Table 4. Real and synthetic occluded images without blur and dust fog.
Illumination (Lux)Real Occlusion CaseSynthetic Occlusion Case
80Sensors 24 05251 i001Sensors 24 05251 i002
80–V5Sensors 24 05251 i003Sensors 24 05251 i004
185Sensors 24 05251 i005Sensors 24 05251 i006
280Sensors 24 05251 i007Sensors 24 05251 i008
Table 5. Real and synthetic occluded images with blur and dust fog.
Table 5. Real and synthetic occluded images with blur and dust fog.
Real Occlusion CaseSynthetic Occlusion Case
BlurSensors 24 05251 i009Sensors 24 05251 i010
Dust fogSensors 24 05251 i011Sensors 24 05251 i012
Dust fog–V3Sensors 24 05251 i013Sensors 24 05251 i014
Table 6. Results of ablation experiment.
Table 6. Results of ablation experiment.
ModelSoft–NMSSimOTASlide LossInner–SIoUmAP0.5/%mAP0.75/%mAP0.5:0.95/%
1 93.985.274.0
2 94.288.777.5
3 95.390.679.6
4 95.991.179.5
596.091.479.8
Table 7. Results of model accuracy and complexity for different pruning methods.
Table 7. Results of model accuracy and complexity for different pruning methods.
Pruning MethodmAP0.5/%mAP0.75/%mAP0.5:0.95/%Parameters
/105
FLOPs
/109
Model Size /MBCompression Degree
Model 596.0091.4079.8070.2615.8014.401.00
+Random [34]90.1075.6064.504.871.201.3014.34
+Lamp [35]90.3075.5064.204.871.201.3014.34
+L1 [36]90.0075.9064.604.201.101.2016.74
+Group–Hessian [37]91.3076.6065.304.201.001.2016.74
+Group–Taylor91.3078.2065.904.201.001.2016.74
Table 8. Results of model detection speed for different pruning methods.
Table 8. Results of model detection speed for different pruning methods.
Pruning MethodFPS
RTX 4080GTX 1050TiGT 1030CPU
Model 5108.0336.8314.467.63
+Random105.7161.9337.3229.20
+Lamp106.3662.1038.2729.60
+L1104.4662.7939.4332.59
+Group–Hessian105.8263.1539.8332.61
+Group–Taylor109.3766.3141.9033.03
Table 9. Results of different detection algorithms.
Table 9. Results of different detection algorithms.
MethodmAP0.5/%Parameters/105FLOPs/109Model Size/MBFPS(CPU)
YOLOv5n92.5017.704.203.9018.51
YOLOv5n–Group–Taylor91.006.731.601.7023.60
YOLOv5s93.9070.2615.8014.407.91
YOLOv5m94.30208.7347.9042.203.78
YOLOv794.30365.07103.2074.801.65
YOLOv7–X94.40708.14188.10142.101.12
YOLOv8n95.4030.078.106.2012.67
YOLOv8s96.00111.2728.4022.505.84
YOLOv8m96.70258.4378.7052.002.67
Proposed method91.304.201.001.2033.03
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fan, H.; Liu, J.; Yan, X.; Zhang, C.; Cao, X.; Mao, Q. A Fast and High-Accuracy Foreign Object Detection Method for Belt Conveyor Coal Flow Images with Target Occlusion. Sensors 2024, 24, 5251. https://doi.org/10.3390/s24165251

AMA Style

Fan H, Liu J, Yan X, Zhang C, Cao X, Mao Q. A Fast and High-Accuracy Foreign Object Detection Method for Belt Conveyor Coal Flow Images with Target Occlusion. Sensors. 2024; 24(16):5251. https://doi.org/10.3390/s24165251

Chicago/Turabian Style

Fan, Hongwei, Jinpeng Liu, Xinshan Yan, Chao Zhang, Xiangang Cao, and Qinghua Mao. 2024. "A Fast and High-Accuracy Foreign Object Detection Method for Belt Conveyor Coal Flow Images with Target Occlusion" Sensors 24, no. 16: 5251. https://doi.org/10.3390/s24165251

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop