Next Article in Journal
Joint Aperture and Power Allocation Strategy for a Radar Network Localization System Based on Low Probability of Interception Optimization
Previous Article in Journal
WiTransformer: A Novel Robust Gesture Recognition Sensing Model with WiFi
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cost-Sensitive YOLOv5 for Detecting Surface Defects of Industrial Products

School of Mechanical and Precision Instrument Engineering, Xi’an University of Technology, Xi’an 710048, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(5), 2610; https://doi.org/10.3390/s23052610
Submission received: 28 January 2023 / Revised: 21 February 2023 / Accepted: 21 February 2023 / Published: 27 February 2023
(This article belongs to the Topic AI and Data-Driven Advancements in Industry 4.0)

Abstract

:
Owing to the remarkable development of deep learning algorithms, defect detection techniques based on deep neural networks have been extensively applied in industrial production. Most existing surface defect detection models assign equal costs to the classification errors among different defect categories but do not strictly distinguish them. However, various errors can generate a great discrepancy in decision risk or classification costs and then produce a cost-sensitive issue that is crucial to the manufacturing process. To address this engineering challenge, we propose a novel supervised classification cost-sensitive learning method (SCCS) and apply it to improve YOLOv5 as CS-YOLOv5, where the classification loss function of object detection was reconstructed according to a new cost-sensitive learning criterion explained by a label–cost vector selection method. In this way, the classification risk information from a cost matrix is directly introduced into the detection model and fully exploited in training. As a result, the developed approach can make low-risk classification decisions for defect detection. It is applicable for direct cost-sensitive learning based on a cost matrix to implement detection tasks. Using two datasets of a painting surface and a hot-rolled steel strip surface, our CS-YOLOv5 model outperforms the original version with respect to cost under different positive classes, coefficients, and weight ratios, but also maintains effective detection performance measured by mAP and F1 scores.

1. Introduction

Industrial and manufacturing object detection has been greatly facilitated by deep neural networks [1,2,3,4]. Object detection models are exploited to predict the position of the object in the input visual data and the corresponding class information, which plays a role in accelerating intelligent industry transformation [5,6,7]. The YOLO family [8,9] is one of the most popular approaches in real applications due to their fast inference speed and outstanding accuracy. Moreover, in the industrial sector, product surface defect detection is a general task [10,11,12]. To improve productivity, many manufacturing industries utilize machine vision systems, within which object detection is the core algorithm [13].
After all, there is no object detection model that can be completely accurate. The vast majority of existing surface defect detection approaches assume the same cost for all detection errors and focus on realizing high detection accuracy [11,14]. However, previous studies have confirmed that different types of detections and misclassifications lead to distinct costs [15,16,17], which leads to the cost-sensitive problem of models in practical industrial applications. For example, a defect detection model for automobile parts may judge nondestructive parts as defective parts, resulting in the loss of time and efficiency. Conversely, the misclassification of defective parts as nondestructive components may be detrimental to the safety of vehicle users and cause potential risks. Obviously, the cost (decision risk) of the latter is significantly higher than the former. Detection methods based on deep learning have achieved great success [18,19], but most of them cannot directly deal with the cost-sensitive problem, which is an urgent issue needing further investigation. This work mainly focuses on the cost-sensitive defect detection problem caused by discriminative misclassification errors.
Cost-sensitive learning has attracted much attention in past years [17,20,21,22]. Existing cost-sensitive models can be roughly divided into two categories: external cost-sensitive and internal cost-sensitive methods [23,24,25]. The objective of the external cost-sensitive method is to deal with the problem of model discriminant bias related to imbalanced data such as the long-tailed distribution of training datasets [26,27,28,29]. Differently, the internal cost-sensitive method is aimed at the decision cost (risk) caused by the classification errors among different categories in specific application scenarios [24,30]. Recently, these methods have been applied to scenarios such as face recognition, intelligent decisions, and intelligent healthcare [16,31,32,33,34,35,36]. However, most existing deep learning-based object detection work does not consider the cost-sensitive problem, but treats the misclassification cost between different classes equally, which may lead to high decision risk and is not applicable to cost-sensitive surface defect detection.
To address the above internal cost-sensitive issue in defect detection, a supervised cost-sensitive YOLOv5 detection model (CS-YOLOv5) is proposed. Considering the insufficiency of after-processing cost-sensitive methods, a direct-type cost-sensitive principle is conducted after re-examining the training process of object detection. Such a principle requires the output class prediction results of the defect detection model to be directly cost-sensitive, rather than being computed in an additional decision stage. That is, the parameters of the object detection model should preserve the classification cost information. The main contributions are listed as follows:
  • Therefore, a classification loss function based on a label–cost vector selection method is designed, which can equip YOLOv5 with cost sensitivity after training. The misclassification cost involved might be the labor cost of defect detection, security cost, etc., which can be specified in practice by defining the cost matrix.
  • Compared with the original YOLOv5 model, CS-YOLOv5 can solve the internal cost-sensitive problem, exploiting the classification risks defined by a risk matrix in specific applications.
  • Experiments on our newly constructed painting defect dataset as well as a hot-rolled steel strip defect dataset demonstrate the superiority of our approach.
The rest of the paper is organized as follows. In Section 2, a newly developed principle and model are proposed. In Section 3, experiments and results analysis are reported to verify the effectiveness of the new model while discussing the proposed method. Finally, Section 4 concludes this paper.

2. Methodology

Figure 1 illustrates the CS-YOLOv5 model and its classification loss structure proposed in this paper. To exploit defect misclassification risks described by a cost matrix, a direct-type cost-sensitive principle based on a label–cost vector selection was designed. Under this principle, we propose a new cost-sensitive loss function. Specifically, the input defect images in training go through the forward process of the deep neural network and are encoded to class predictions and region predictions. Then, cost vectors are selected from the cost matrix via supervised class labels, which are utilized for the cost-sensitive loss calculation along with the class prediction. Meanwhile, classification loss and region loss are also derived. Finally, an object detection model with cost-sensitive features is obtained by backpropagating the loss gradient and updating the model parameters.

2.1. Cost-Sensitive Learning Modeling

Internal cost-sensitive learning attempts to reduce the decision risk defined in specific scenarios in training. This can be formulated as follows [37]:
ϕ * ( x ) = arg m i n j loss ( x , j )
where loss ( x , j )   is the expected cost of classifying sample x as j-th class, which is determined by the element Cij in the cost matrix C. Furthermore, the cost matrix C is obtained via analysis from domain experts or data mining, e.g., estimating the cost of leakage and security costs in surface defect detection.
Without loss of generality, in deep learning, the classification loss is expressed as a function of the difference between the supervised class labels and the output prediction distribution. Assuming X and D are the input space and the output space, respectively, then the expectation of classification error can be expressed as:
E = E X , D l [ φ ( X ) , D ]
where φ ( X ) is the prediction result of the model for input space X, and l ( ) is the measurement method of the classification error. Therefore, the classification error for a single input sample x is described as follows:
e x , d = l ( p , d )
where p = φ ( x ) = [ p 1 , , p m ] is the category probability vector estimated by the model for x, and the element pj denotes the probability of the input sample x being recognized as the j-th class. Similarly, d is the one-hot label vector of x. For a sampled batch b with size B, Formula (2) can be expressed as:
E b = 1 B i B l ( p ( i ) , d ( i ) )
In terms of a defect detection model, p = g ( x , θ ) , where g ( , θ ) is the deep neural network with a parameter set of θ. The general classification loss function can measure the difference between the input space and the output space. However, there is no specific design for the output space. Classification errors between different classes are treated equally, without cost sensitivity. Generally, the total loss function includes the classification loss and a loss related to the region predictions for targets. Therefore, the general form can be expressed as follows:
L = L c l s + β L r e g
where Lcls is the classification loss designed in a cross-entropy form, Lreg is the location region loss, and β is a trade-off parameter.
According to the calculation of the ordinary classification loss described by Equation (4), the following conclusions can be drawn: (1) The classification loss between different categories is not assigned specifically. (2) Moreover, only the difference between class predictions and the correspondence labels is taken into account. In essence, it is still a generic non-cost-sensitive loss function that cannot process misclassification risks in defect detection.
The classification cost matrix C (decision risk matrix) is formulated as:
[ 0 C 12 C 1 N C 21 0 C 2 N C i 2 0 C N 1 C N ( N 1 ) 0 ]
The number of rows and columns of C are equal to the classes’ quantity N, where C i j 0 and C i 0 . Without loss of generality, C i i = 0 means that the classification is correct without cost. Row vector Ci denotes the cost of classifying the samples with the class i into each class [30,38]. In multiclass detection or classification tasks, if all classes are divided into positive classes and negative ones, the misclassification coefficient can be divided into 4 types [16]:
  • False acceptance coefficient λNP: the risk of misclassifying a target (imposter) that belongs to the negative category into the positive category or positive class (gallery).
  • False rejection coefficient λPN: the risk of misclassifying a gallery to an imposter.
  • Two types of misidentification coefficients λPP and λNN: that is, misclassification between classes of samples with the same nature (the positive class or negative class).
The relationship between the magnitudes of the different coefficients is generally expressed as follows:
λ N P > λ P N > λ N N > λ P P
Moreover, the elements of the cost matrix can be expressed as:
C i j = { λ N P ,   if   i ϵ N , j ϵ P λ P N ,   if   i ϵ P , j ϵ N λ N N ,   if   i ϵ N , j ϵ N λ P P ,   others  

2.2. Direct-Type Cost-Sensitive Learning

The goal of cost-sensitive classification is to learn the probability prediction vectors formulated in Equation (4), which not only has high accuracy defined by supervised labels but also low cost with respect to the cost matrix C. In this work, we argue that the cost matrix is additional supervised information for object detection. The rationale behind this is that the cost matrix reflects requirements for the model in a specific scenario. Therefore, we describe how to train a cost-sensitive defect detection model using both kinds of supervised information in the following part.
As the label exploitation is already formulated in Equation (4), we construct the cost-sensitive constraint in a label-cost vector selection method. Specifically, the cost vector is selected from the cost matrix C via the label d of the input sample x. This vector also constrains the output of posterior probability p, in which the maximum probability in p determines the output class. Obviously, our approach incorporates cost information directly into training, which is different from the post-processing cost-sensitive learning based on Bayes minimum risk methods [31,39].
In the case of supervised learning, the ground-truth class index of sample x with label d is:
t = arg m a x 1 i N d i
According to Equation (1), the target of direct-type cost-sensitive learning is formulated as:
{ p = arg m i n p N p C t ϕ * ( x ) = arg m a x 1 i N p
where p is the output probability vector, and Ct is the aforementioned selected risk (cost) vector. Let Rtp(x) represent the misclassification risk of samples of class t under a current probability vector p. The risk brought by p can be expressed as:
{ p = arg m i n p N R t p ( x ) = arg m i n p N j = 1 N p j C t j ϕ * ( x ) = arg m a x 1 i N ( p i )
Equation (11) indicates that in the training of the direct-type cost-sensitive model: (1) output vector p is to minimize the overall misclassification cost, i.e., minimizing the risk of classifying the sample belonging to class t into each class. (2) The same as before, the maximum index in the probability vector determines the class output. This is the direct-type cost-sensitive learning principle of the deep learning model with supervised label and cost information. The new principle specifies the calculation form of risk under supervised learning and provides linear constraints for the estimation of probability directly in N-dimensional space.

2.3. CS-YOLOv5 Model and Method Analysis

In order to equip the defect object detection model with cost sensitivity, the modified loss function is proposed and described as follows:
Loss = λ i o u i = 0 S 2 j = 0 B 1 i j o b j L CIoU λ c l s i = 0 S 2 j = 0 B 1 i j o b j c ϵ c l a s s e s λ c p ^ t i ( c ) log p i ( c ) + 1 p ^ t i ( c ) log 1 p i ( c ) + η λ cost i = 0 S 2 j = 0 B 1 i j o b j c ϵ llasses λ c S p i ( c ) C [ t ( i j ) ] c + λ o b j i = 0 S 2 j = 0 B 1 i j o b j λ p C ^ i log C i + 1 C ^ i log 1 C i + λ o b j i = 0 S 2 j = 0 B 1 i j noobj C ^ i log C i + 1 C ^ i log 1 C i
where C [ t ( i j ) ] c is the cost of that the ij-th sample of the class t ( i j ) is classified into C class, and is the c-th element of cost vector C [ t ( i j ) ] from the risk coefficient matrix C. η, λ cost   , λ c l s , λ i o u , and λ o b j are trade-off parameters that control the importance of distinct parts of losses. 1 i j o b j returns the value of 1 if the ij-th region contains an object. S ( y ) = ( 1 + e y ) 1 is the sigmoid logistic regression function. Individually, the new classification loss function is expressed as:
L c l s = η λ cost i = 0 S 2 j = 0 B 1 i j o b j λ c S ( p i ) C [ t ( i j ) ] λ c l s i = 0 S 2 j = 0 B 1 i j o b j c ϵ   lasses   λ c [ p ^ t i ( c ) log ( p i ( c ) ) + ( 1 p ^ t i ( c ) ) log ( 1 p i ( c ) ) ]
where the first term is the cost-sensitive classification loss. When the parameter λ cost = 0 and λ c l s 0 , the loss degenerates to a non-cost-sensitive loss. When λ cost 0 and λ c l s = 0 , the loss degradation calculation is similar to the loss function in [25]. However, the intuitive meaning of improving the loss function in [25] is that a O 2 item is sufficiently close to the label value, which does not follow the cost-sensitive principle in Equation (10). The cost-sensitive principle in Equation (10) is further formulated as:
L cost = ccclasses   λ c S ( p i ( c ) ) C t c  
The use of a sigmoid function and normalization operation for the risk matrix in practical calculation can guarantee the boundary, thereby preventing the vanishing gradient and gradient explosion in training. In Equation (13), the classification loss is constructed as minimizing classification risk under class labels and the cost matrix. In other words, the optimization of the cross-entropy loss ensures the accuracy of the output in the label sense, that is, the estimate of the posterior probability is close enough to the label. The existence of cost-sensitive classification loss ensures that the risk corresponding to the estimated posterior probability is small enough. The loss weights of both realize the balancing effect of classification loss for different parts.

2.3.1. Cost-Sensitive Gradients

During the supervised training of the object detection model, the impact of the modified classification loss function on the implementation of the backpropagation algorithm needs to be taken into consideration [40]. Since the cross-entropy loss in the classification loss function remains unchanged and is superimposed with the cost-sensitive loss function, it will not affect the original backpropagation process. For a sample x of class t, the estimate of its output probability is set as p. For the output layer, the gradient with respect to output neuron 0n is directly associated with the n-th dimension output pn and can be expressed as:
L cost   o n = λ c n C t n S ( p n ) p n
Since S ( y ) y = S ( y ) ( 1 S ( y ) ) = e y ( 1 + e y ) 2 , then:
L c o s t o n = λ c n C t n e p n ( 1 + e p n ) 2
The gradient of the cost-sensitive loss function directly with respect to the output probability is:
L c o s t = n = 1 N λ c n C t n e p n ( 1 + e p n ) 2
Since e p n > 0 and C i 0 , then L cos t   0 .

2.3.2. Cost-Sensitive YOLOv5 Algorithm

The role of the cost-sensitive classification loss function proposed in this paper is to establish a cost-sensitive parameter optimization space by combining various loss functions. The optimization algorithm is executed on this parameter space, thereby continuously updating the weight parameters of the model. The forward propagation of the established CS-YOLOv5 model has cost sensitivity. The implementation process is expressed in Algorithm 1:
Algorithm 1: Training direct-type cost-sensitive YOLOv5
Input: training space ( X ,   D ) , validation set ( X v , D v ) , minibatch size b, training epochs e , learning rate lr, as well as loss weight ( λ c o s t , λ c l s , λ i o u , λ c ) , and model g(θ)
Output: trained model g(θ*)
1: Initialize parameters randomly of the model g(θ0)
2: If the current epoch i e , loop:
3:   If the current batch is j, loop:
4:    Forward propagation: O j = g ( X j , D j θ j 1 )
5:    Calculate the loss L j = ( O j , D j ) based on Equation (12)
6:    Calculate the gradient L j and backpropagation, execute the SGD algorithm
7:    Update parameter θ j θ j 1
8:   Evaluate the model: g ( X v , D v θ i )
9: Return g ( θ * )
End
Through the proposed cost-sensitive loss method, the misclassification cost information is directly involved in the training of the deep neural network model. With the classification loss weight ( λ cost , λ c l s ) set within an appropriate range, we have the following observations:
(Case 1): When the risk coefficient of the cost vector is Cii = 0 or Cii is the smallest element in vector Ci, classifying the sample into its ground-truth has the lowest risk. If the risk coefficient Cin is corresponding to the large value pn in the output probability, the classification cost will be large. The model optimization process results in a delineation of the ground-truth and drives away this costly situation.
(Case 2): When the risk coefficient Cii of the cost vector is the element with a larger value in the vector Ci and the risk coefficient corresponding to the larger value pn in the posterior probability estimation is exactly Cii, the correct classification will cause a large classification cost. After parameter updating, the model optimization process will lead to the division of clear truth value label pointing.
The essence of the classification loss function is to move away or strengthen the decision boundary towards the direction of less risk. Then, the boundary ambiguity is reduced compared with non-cost-sensitive models. From another perspective, the method proposed in this paper uses the prior knowledge of the cost matrix to modify the constraints of the model optimization. Thus, the optimization of model parameters can be regarded as a regularization method. The cost-sensitive learning with the original Bayes minimum risk decision type is to simply concatenate the output of estimating probability through the process of the minimum-risk decision. Instead, our new cost-sensitive learning method proposed in this paper is to obtain the optimal posterior probability vector in the sense of the label and cost matrix in one training process, which makes the model output directly cost-sensitive.

3. Experiments and Results Analysis

3.1. Experimental Dataset

NEU surface defect database: This industrial image dataset of hot-rolled steel strip surface defects is constructed in [41], which contains six types of defects, i.e., rolled-in scale (RS), patches (Pa), crazing (Cr), pitted surface (PS), inclusion (In), and scratches (Sc). There are 300 images in each class for a total of 1800 images. In the experiment, 1500 images were divided into a training set, and the remaining 300 images were divided into a test set. The original data size was 200 × 200 pixels, which was converted to size 192 × 192 in the experiments. This dataset is referred to as the NEU dataset for simplicity in this paper.
Paint surface defect dataset: This is a manufacturing image dataset involving paint spraying surface defects. The author participated in the collection and collation work of the inspection site of construction machinery structural parts. The dataset contains four classes of target defects, including dust, pit, sag, and scratch. There were 1417 images in total, including 1246 training set images and 171 test set images. The original size was 1024 × 1024, which was set to size 640 × 640 in the experiments. For simplicity, this dataset is referred to as the Paint dataset in the remainder of this paper.
The class distributions and the size distribution of ground-truth bounding box of these two datasets are counted and shown in Figure 2a,b, respectively. In the bar figures, the values denote the statistics of the sample numbers for each class. For the NEU dataset, the number of sample classes is more uniform than Paint. The Paint dataset has obvious uneven sample distribution, in which sag and scratch are less than the other two classes, and most of the two classes are more than 75% of the total number of samples. This is because the frequency of sag and scratch is so low at the collection site. Thus, it is a long-tail problem, which is very common in application.
The scatter figures exhibit the relative size distribution of ground-truth bounding box for the datasets, that is, the distribution of length and width. The darker color indicates more bounding boxes are concentrated here. It can be seen that the bounding box width and height of the NEU dataset are widely distributed. The densest is where the relative width is about 0.2 and the relative height is about 0.1. In the Paint dataset, the bounding box is clearly concentrated in the lower left corner of the scatterplot. This is because the actual bounding boxes of dust samples in the dataset are relatively small, and the corresponding number is much larger than other classes. Based on the above analysis, the Paint dataset has obvious sample imbalance and most sample bounding boxes are too small. Therefore, the object detection problem of the Paint dataset is more difficult than that of the NEU dataset.

3.2. YOLOv5 Model Settings

YOLOv5 was selected as the baseline model in experiments, whose code is provided by [7]. We experimented with several Tesla T4 GPUs and implemented the code with the Pytorch 1.7.1 platform. The initial learning rate was set to 0.002. The cosine annealing strategy was utilized during training. The weight decay coefficient was 0.0005 for the adopted SGD optimizer. The balance parameter β was set to 0.375, and η was set to 1/10,000. For the NEU dataset, the batch size was set to 230, with 90 epochs of training. For the Paint dataset, the batch size was set to 24, with 55 epochs of training. The experiments all used the Mosaic data augmentation method. As the examples show in Figure 3a,b, this augmentation method stitches images together in an overlapping manner to improve the robustness of the models.

3.3. Experimental Metrics

3.3.1. Classification Cost Evaluation Metric

The proposed cost-sensitive learning method concentrates on minimizing the decision risk within the acceptable range of accuracy. To quantitatively measure the classification cost, we construct a metric formulated as:
Cost   = 1 F i = 1 F c   cesses   p i c C [ t ( i ) ] c = 1 F i = 1 F P i · C [ t ( i ) ]
where C [ t ( i ) ] c is the c-th element of the cost vector C [ t ( i ) ] from matrix C and denotes the cost of classifying the j-th sample belonging to the class t ( i ) into the c -th class. F is the number of positive samples used to calculate the loss. Intuitively, such a cost metric is consistent with the cost-sensitive target to be optimized in the loss function.

3.3.2. Comprehensive Metrics

The recall ratio in object detection refers to the proportion of data samples correctly detected by the model, which is one of the basic performance measurements. TP, FP, TN, and FN are true positive, false positive, true negative, and false negative examples, respectively. Then, the recall rate is:
Recall   = TP TP + FN
To comprehensively evaluate the trained models, we also adopted the common metric, mean average precision (mAP) and F1-score in the experiments. They are described as follows:
mAP = 0 1 Precision ( r ) d r
where Precison   = TP / ( TP + FP ) is the accuracy rate, and r represents Recall.
F 1 = 2 ×   Precision     Recall     Precision   +   Recall  
For the multiclass data (more than two), the calculation method is given in the literature [42].

3.4. Risk Coefficient Experiment

The premise of cost-sensitive learning methods is to construct a classification cost matrix for the current data. Since the cost matrix is directly involved in the calculation of the classification cost of Equation (16), we performed experiments with different classification cost matrices. Different risk coefficient ratios simulate different setting conditions of the cost matrix in real applications to verify the effectiveness of the cost-sensitive learning method.
Each element of the cost matrix is the classification risk coefficient. Moreover, objects in the dataset were divided into a positive class and a negative class. The settings of the four risk coefficients follow Equation (7). In this experiment, we tested the cost-sensitive learning method by adjusting the rate of risk coefficient under fixed positive classes.

3.4.1. Risk Coefficient Setting

For the NEU dataset, positive classes were selected as Pa, PS, In, and Sc, while for Paint, dust and scratch were positive classes. Figure 4a,b shows the examples of positive and negative classes in NEU and Paint dataset, respectively.
According to Equation (7), the base coefficient is selected as λ N P : λ P N : λ N N : λ P P = 200 : 20 : 2 : 1 . The ratios between adjacent coefficients are linearly scaled by equal proportions to obtain four sets of classification risk coefficients. Table 1 shows the risk coefficient grouping.

3.4.2. Experimental Results and Analysis

The experimental results of classification cost of four groups (a)–(d) with different risk coefficients are shown in Figure 5. For the four groups (a), (b), (c), and (d), the classification costs related to CS-YOLOv5 are 0.000470, 0.000337, 0.000340, and 0.000341, while those for YOLOv5 are 0.00368, 0.00378, 0.00381, and 0.00422. Compared with YOLOv5, the cost of group (a) for CS-YOLOv5 decreases by 87.16%, and the most significant decrease in group (d) is 91.93%. This is because the risk coefficient ratio of different groups directly affects the cost matrix, thus giving CS-YOLOv5 different cost sensitivities during training. According to the cost metric given by Equation (17), different risk matrices also make models have different cost performances. Similar conclusions can be obtained in Figure 5b. According to the results of the NEU and Paint datasets, the classification cost of the cost-sensitive YOLOv5 model is also significantly lower than that of the original YOLOv5. Under different risk coefficients, the output of the cost-sensitive YOLOv5 model itself has lower classification costs, which proves the effectiveness of our method.
Moreover, the results of mAP and F1-score are shown in Table 2. The improved model has cost-sensitive capabilities while performing comparably to the original model under two metrics. Even the performance of the proposed approach is better than the original YOLOv5 under some risk coefficients (bold in the table). It can be concluded that when other cost matrix settings such as positive and negative classes are fixed, cost-sensitive YOLOv5 significantly improves the cost-sensitive performance without losing the comprehensive performance.

3.5. Experiment of Positive Classes

In real scenarios, positive and negative classes are often determined in accordance with different production requirements. The division of the positive class and negative class plays a crucial role in cost-sensitive learning. According to Section 3.2, the division is directly related to the form of the cost matrix. Therefore, we performed this experiment by setting the number of positive classes as a variable to verify the effectiveness of CS-YOLOv5.

3.5.1. Number Setting of Positive Classes

In this part, the risk coefficient was set to a fixed λ N P : λ P N : λ N N : λ P P = 200 : 20 : 2 : 1 . For the NEU dataset, the number of positive classes was increased from 1 to 4. For the Paint (with a total of 4 classes), the number of positive classes was increased from 1 to 3. Considering the preciseness, the positive classes were randomly selected and tested several times.

3.5.2. Experimental Results and Analysis

For the division, samples belonging to positive classes were classified into the positive set, while samples of other classes are classified into the negative set. The cost matrix conducted based on this division and risk coefficients was exploited in training. The classification cost results under different positive classes are shown in Figure 6.
As can be seen from Figure 6, when the number of positive classes in the NEU surface defect database dataset is 1, 2, 3, and 4, the classification cost of CS-YOLOv5 is 67.58%, 75.36%, 44.56%, and 89.27% lower than that of YOLOv5. Different numbers of positive classes will affect the construction of the cost matrix, so the parameters of CS-YOLOv5 have different cost sensitivity. According to the cost metric defined by Formula (17), the change in the cost matrix will lead to different classification cost of YOLOv5. Similarly, the analysis of Figure 6b can also reach a similar conclusion. Therefore, the classification cost of the CS-YOLOv5 model is lower than the original model under various partitions on both datasets, which demonstrates the superiority of the direct-type cost-sensitive method. In addition, the performance under mAP and F1-score is shown in Table 3.
From these results, we have some observations: (1) CS-YOLOv5 has performance which is comparable to or better than the original model under the metrics. (2) When other settings such as the misclassification risk coefficient are fixed, the cost sensitivity of CS-YOLOv5 is significantly improved.

3.6. Weight Ratio Experiment

As stated in Section 3.4, the cost-sensitive classification loss (Equation (14)) conforms to the direct-type cost-sensitive principle proposed for supervised learning. Furthermore, the optimization of the cross-entropy loss in the classification loss function enforces the estimation of the output probability close to the corresponding ground-truth. On the other hand, the role of the cost loss function (Equation (16)) is the total classification risk of the output probability. The proportion and balance of the two targets are determined by the weight ratio r = λ cos t   / λ c l s . Therefore, it is reasonable and necessary to implement experiments on this weight ratio in the classification loss.

3.6.1. Settings for Weight Ratio Experiment

The risk coefficient is set to λ N P : λ P N : λ N N : λ P P = 200 : 20 : 2 : 1 . For the NEU dataset, positive classes are selected as Pa, PS, In, and Sc, and those are dust and scratch for Paint. Moreover, the weight ratio r = λ cos t   / λ c l s is set to vary from 0 to 1.5 with an increment of 0.25. A total of 7 groups of linearly changing ratios are considered. According to the analysis in Section 2.3, when r = 0 , the classification loss used in the experiment degenerates into a non-cost-sensitive loss function. In this case, the model is non-cost-sensitive.

3.6.2. Experimental Results and Analysis

The experimental results under different loss weight ratios are developed into coefficient weight–classification cost curves in Figure 7. When r = 0 , the results denote the original non-cost-sensitive model and the other 6 groups of CS-YOLOv5.
As shown in Figure 7a, r varies from 0.25 to 1.5 on NEU, representing 6 groups of CS-YOLOv5 models with different loss weight ratios. r = 0 indicates that the model degenerates to the original model YOLOv5. From these results, we can see that the classification cost of 6 groups of CS-YOLOv5 models is lower than that of YOLOv5, which indicates the effectiveness of the cost-sensitive method for model improvement. Too-large r can negatively affect the performance of CS-YOLOv5. Thus, in constructing CS-YOLOv5, an appropriate loss weight ratio can equip the model with cost sensitivity and maintain good performance. Similar conclusions can be drawn from the experiments on the Paint dataset in Figure 7b. The cost of the CS-YOLOv5 model on both datasets is lower than the results of the ordinary YOLOv5 model represented by r = 0. Consequently, under the cost-sensitive classification functions of different weights in experiments, the effectiveness of the direct-type cost-sensitive method in reducing the classification cost has been demonstrated. In addition, the comprehensive performance measured by mAP and F1-score is shown in Table 4.
From the results on the Paint dataset in Table 4, it can be observed that the improved model with appropriate weight ratio performs better than the original model. When the weight ratios are 0.75 and 0.64, the mAP and F1 metrics reach their optimal values, respectively. As the weight of the cross-entropy loss function decreases, the performance of the model degrades. When the weight ratio is set to 1.5, both mAP and F1 drop to the lowest value of 0.59. This is because a small proportion of cross-entropy loss makes the model parameter more inclined to be cost-sensitive, with insufficient accuracy optimization. Similar conclusions can be drawn from the results of the NEU dataset. Therefore, it can be concluded that the two weights λ c o s t and λ c l s in the classification loss need to be properly set, so that YOLOv5 can have cost sensitivity and comprehensive performance.

3.7. Ablations

As clarified in Section 3.1, the resolution of these two datasets is quite different, which is 192 × 192 for NEU and 640 × 640 for Paint. Thus, the batch sizes on them have a gap under same the computing resources. On another, we implemented a smaller batch size on NEU and found that they had few differences, and the results are shown in Table 5.
Moreover, since the number of images in these two datasets is not large (1800 and 1417, respectively) and the categories are fewer than 10, many epochs are not necessary. We present the training classification loss and cost curves in Figure 8, and we can observe that the models tend to converge under the adopted settings.

3.8. Discussions and Comparisons

In this paper, we design a strategy to address the cost-sensitive issue of defect detection by presenting the CS-YOLOv5. Besides the widely adopted YOLO family in manufacturing, another single-stage SSD method [43] is also exploited in the previous literature [44]. Such methods are comparable with YOLO in detection speed, but the settings of the prior box are sensitive and may lead to instability [45]. On the other hand, the two-stage detector Faster-RCNN [4] is also utilized in defect detection [46], which especially perfroms better in denser detection. However, it is slower than single-stage designs due to their higher complexity. More importantly, both of these vanilla versions of the detector cannot process the internal cost-sensitive problem without a specific design, while our CS-YOLOv5 can not only reduce the misclassification cost but also inherits the speed and accuracy advantages.

4. Conclusions

To tackle the challenge of cost-sensitive defect detection in manufacturing, in this work, we propose a novel general cost-sensitive learning method called the supervised classification cost-sensitive learning method. Specifically, the misclassification risks represented by a cost matrix are directly integrated into the model optimization. Cost vectors are selected with labels as additional supervised information to train. Based on this approach, we enrich the YOLO family with a cost-sensitive YOLOv5 (CS-YOLOv5). It is a method that reduces the misclassification risk while maintaining the original model structure. We also construct a defect detection dataset from an industrial site. Extensive experiments demonstrate the effectiveness of the proposed CS-YOLOv5 for cost-sensitive defect detection. In the future, we will pay attention to some denser detection tasks which are more sensitive to location information. Hence, future research work can focus on feasible principles and methods for location cost in manufacturing.

Author Contributions

B.L. and Y.L. provided the conception and methodology of this study and are responsible for the draft text writing. B.L. and F.G. designed the experiments and completed the data analysis. Y.L. assisted with the validation and guided the writing and revision of the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Key Industrial Chain Project of Shaanxi Province (No. 2021ZDLGY 10-07) and the Scientific-Technological Innovation Projects in Strategic Emerging Industries of Shandong Province (No. 2017TSCYCX-24).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.

Acknowledgments

Thanks to the advanced manufacturing lab of Xi’an University of Technology for providing this research with open experiment data and resources.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Qi, S.; Yang, J.; Zhong, Z. A Review on Industrial Surface Defect Detection Based on Deep Learning Technology. In Proceedings of the 3rd International Conference on Machine Learning and Machine Intelligence, Hangzhou, China, 18–20 September 2020; pp. 24–30. [Google Scholar] [CrossRef]
  2. Mahto, P.; Garg, P.; Seth, P.K.; Panda, J. Refining Yolov4 for Vehicle Detection. Int. J. Adv. Res. Eng. Technol. (IJARET) 2020, 11, 409–419. [Google Scholar]
  3. He, Y.; He, N.; Zhang, R.; Yan, K.; Yu, H. Multi-scale feature balance enhancement network for pedestrian detection. Multimed. Syst. 2022, 28, 1135–1145. [Google Scholar] [CrossRef]
  4. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar] [CrossRef] [Green Version]
  6. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar] [CrossRef]
  7. Jocher, G.; Stoken, A.; Borovec, J.; NanoCode012; Chaurasia, A.; Liu, C.; Xie, T.; Abhiram, V.; Laughing; Tkianai; et al. Ul-tralytics/Yolov5: V5.0—YOLOv5-P6 1280 Models, AWS, Supervise.ly and YouTube Integrations. Zenodo 2021, 11. [Google Scholar] [CrossRef]
  8. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef] [Green Version]
  9. Yang, G.; Feng, W.; Jin, J.; Lei, Q.; Li, X.; Gui, G.; Wang, W. Face mask recognition system with YOLOV5 based on image recognition. In Proceedings of the 2020 IEEE 6th International Conference on Computer and Communications (ICCC), Chengdu, China, 11–14 December 2020; pp. 1398–1404. [Google Scholar] [CrossRef]
  10. Chen, X.; Lv, J.; Fang, Y.; Du, S. Online Detection of Surface Defects Based on Improved YOLOV3. Sensors 2022, 22, 817. [Google Scholar] [CrossRef] [PubMed]
  11. Xu, R.; Hao, R.; Huang, B. Efficient surface defect detection using self-supervised learning strategy and segmentation network. Adv. Eng. Inform. 2022, 52, 101566. [Google Scholar] [CrossRef]
  12. Tabernik, D.; Šela, S.; Skvarč, J.; Skočaj, D. Segmentation-based deep-learning approach for surface-defect detection. J. Intell. Manuf. 2019, 31, 759–776. [Google Scholar] [CrossRef] [Green Version]
  13. Chen, Y.; Ding, Y.; Zhao, F.; Zhang, E.; Wu, Z.; Shao, L. Surface Defect Detection Methods for Industrial Products: A Review. Appl. Sci. 2021, 11, 7657. [Google Scholar] [CrossRef]
  14. Da, Y.; Dong, G.; Wang, B.; Liu, D.; Qian, Z. A novel approach to surface defect detection. Int. J. Eng. Sci. 2018, 133, 181–195. [Google Scholar] [CrossRef] [Green Version]
  15. Elkan, C. The foundations of cost-sensitive learning. In Proceedings of the 17th International Joint Conference on Artificial Intelligence, Seattle, WA, USA, 4–10 August 2001; Morgan Kaufmann Publishers Inc.: Seattle, WA, USA, 2001; Volume 2, pp. 973–978. [Google Scholar]
  16. Zhang, Y.; Zhou, Z.-H. Cost-Sensitive Face Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1758–1769. [Google Scholar] [CrossRef]
  17. Frumosu, F.D.; Khan, A.R.; Schiøler, H.; Kulahci, M.; Zaki, M.; Westermann-Rasmussen, P. Cost-sensitive learning classification strategy for predicting product failures. Expert Syst. Appl. 2020, 161, 113653. [Google Scholar] [CrossRef]
  18. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar] [CrossRef]
  19. Wu, X.; Sahoo, D.; Hoi, S.C.H. Recent advances in deep learning for object detection. Neurocomputing 2020, 396, 39–64. [Google Scholar] [CrossRef] [Green Version]
  20. Fernandez, A.; Garcia, S.; Herrera, F.; Chawla, N.V. SMOTE for Learning from Imbalanced Data: Progress and Challenges, Marking the 15-year Anniversary. J. Artif. Intell. Res. 2018, 61, 863–905. [Google Scholar] [CrossRef]
  21. Zhu, X.; Yang, J.; Zhang, C.; Zhang, S. Efficient Utilization of Missing Data in Cost-Sensitive Learning. IEEE Trans. Knowl. Data Eng. 2021, 33, 2425–2436. [Google Scholar] [CrossRef]
  22. Fernández, A.; García, S.; Galar, M.; Prati, R.C.; Krawczyk, B.; Herrera, F. Cost-sensitive learning. In Learning from Imbalanced Data Sets; Springer: Berlin/Heidelberg, Germany, 2018; pp. 63–78. [Google Scholar] [CrossRef]
  23. Kukar, M.; Kononenko, I. Cost-Sensitive Learning with Neural Networks. In European Conference on Artificial Intelligence; John Wiley & Sons, Ltd.: Hoboken, NJ, USA, 1998; pp. 445–449. [Google Scholar] [CrossRef]
  24. Petrides, G.; Verbeke, W. Cost-sensitive ensemble learning: A unifying framework. Data Min. Knowl. Discov. 2022, 36, 1–28. [Google Scholar] [CrossRef]
  25. Khan, S.H.; Hayat, M.; Bennamoun, M.; Sohel, F.A.; Togneri, R. Cost-Sensitive Learning of Deep Feature Representations from Imbalanced Data. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 3573–3587. [Google Scholar] [CrossRef] [Green Version]
  26. Zhou, Z.H.; Liu, X.Y. Training cost-sensitive neural networks with methods addressing the class imbalance problem. IEEE Trans. Knowl. Data Eng. 2006, 18, 63–77. [Google Scholar] [CrossRef]
  27. Li, T.; Cao, P.; Yuan, Y.; Fan, L.; Yang, Y.; Feris, R.; Indyk, P.; Katabi, D. Targeted Supervised Contrastive Learning for Long-Tailed Recognition. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 6908–6918. [Google Scholar] [CrossRef]
  28. Raj, V.; Magg, S.; Wermter, S. Towards Effective Classification of Imbalanced Data with Convolutional Neural Networks. In Artificial Neural Networks in Pattern Recognition; Springer: Cham, Switzerland, 2016; pp. 150–162. [Google Scholar] [CrossRef]
  29. Zhou, Z.-H. Cost-sensitive learning. In Modeling Decision for Artificial Intelligence: 8th International Conference; Springer: Berlin/Heidelberg, Germany, 2011; pp. 17–18. [Google Scholar] [CrossRef]
  30. Chung, Y.-A.; Lin, H.-T.; Yang, S.-W. Cost-aware pre-training for multiclass cost-sensitive deep learning. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 2016; AAAI Press: New York, NY, USA, 2016; pp. 1411–1417. [Google Scholar] [CrossRef]
  31. Collell, G.; Prelec, D.; Patil, K.R. A simple plug-in bagging ensemble based on threshold-moving for classifying binary and multiclass imbalanced data. Neurocomputing 2018, 275, 330–340. [Google Scholar] [CrossRef] [PubMed]
  32. Kuo, W.; Häne, C.; Yuh, E.; Mukherjee, P.; Malik, J. Cost-Sensitive Active Learning for Intracranial Hemorrhage Detection. In Medical Image Computing and Computer Assisted Intervention—MICCAI; Springer: Berlin/Heidelberg, Germany, 2018; pp. 715–723. [Google Scholar] [CrossRef] [Green Version]
  33. Zhang, H.; Jiang, L.; Li, C. CS-ResNet: Cost-sensitive residual convolutional neural network for PCB cosmetic defect detection. Expert Syst. Appl. 2021, 185, 115673. [Google Scholar] [CrossRef]
  34. Liu, M.; Miao, L.; Zhang, D. Two-Stage Cost-Sensitive Learning for Software Defect Prediction. IEEE Trans. Reliab. 2014, 63, 676–686. [Google Scholar] [CrossRef]
  35. Natarajan, N.; Dhillon, I.S.; Ravikumar, P.; Tewari, A.J.J.M.L.R. Cost-Sensitive Learning with Noisy Labels. J. Mach. Learn. Res 2017, 18, 5666–5698. [Google Scholar]
  36. Seliya, N.; Khoshgoftaar, T.M. The use of decision trees for cost-sensitive classification: An empirical study in software quality prediction. WIREs Data Min. Knowl. Discov. 2011, 1, 448–459. [Google Scholar] [CrossRef] [Green Version]
  37. Wan, J.; Yang, M. Survey on Cost-sensitive Learning Method. J. Softw. 2019, 113–136. [Google Scholar] [CrossRef]
  38. Lu, S.; Liu, L.; Lu, Y.; Wang, P. Cost-sensitive neural network classifiers for postcode recognition. Int. J. Pattern Recognit. Artif. Intell. 2012, 26, 1263001. [Google Scholar] [CrossRef]
  39. Jiang, X.; Mahadevan, S. Bayesian risk-based decision method for model validation under uncertainty. Reliab. Eng. Syst. Saf. 2007, 92, 707–718. [Google Scholar] [CrossRef]
  40. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  41. Song, K.; Yan, Y. A noise robust method based on completed local binary patterns for hot-rolled steel strip surface defects. Appl. Surf. Sci. 2013, 285, 858–864. [Google Scholar] [CrossRef]
  42. Espíndola, R.P.; Ebecken, N.F.F. On extending f-measure and g-mean metrics to multi-class problems. In Sixth International Conference on Data Mining, Text Mining and Their Business Applications; WIT Press: Ashurst Lodge, UK, 2005; Volume 35, pp. 25–34. [Google Scholar] [CrossRef]
  43. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot Multibox Detector; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar] [CrossRef] [Green Version]
  44. Chen, J.; Liu, Z.; Wang, H.; Nunez, A.; Han, Z. Automatic Defect Detection of Fasteners on the Catenary Support Device Using Deep Convolutional Neural Network. IEEE Trans. Instrum. Meas. 2018, 67, 257–269. [Google Scholar] [CrossRef] [Green Version]
  45. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
  46. He, Y.; Song, K.; Meng, Q.; Yan, Y. An End-to-End Steel Surface Defect Detection Approach via Fusing Multiple Hierarchical Features. IEEE Trans. Instrum. Meas. 2020, 69, 1493–1504. [Google Scholar] [CrossRef]
Figure 1. CS-YOLOv5 model and classification loss structure.
Figure 1. CS-YOLOv5 model and classification loss structure.
Sensors 23 02610 g001
Figure 2. (a) NEU dataset sample statistics (b) Paint dataset sample statistics.
Figure 2. (a) NEU dataset sample statistics (b) Paint dataset sample statistics.
Sensors 23 02610 g002
Figure 3. (a) NEU Data augmentation (b) Paint Data augmentation. Numbers are class indexes.
Figure 3. (a) NEU Data augmentation (b) Paint Data augmentation. Numbers are class indexes.
Sensors 23 02610 g003
Figure 4. (a) Example of NEU division (b) Example of Paint division.
Figure 4. (a) Example of NEU division (b) Example of Paint division.
Sensors 23 02610 g004
Figure 5. (a) Cost (×10−3) with varying risk coefficients on NEU (b) Cost (×10−3) with varying risk coefficients on Paint.
Figure 5. (a) Cost (×10−3) with varying risk coefficients on NEU (b) Cost (×10−3) with varying risk coefficients on Paint.
Sensors 23 02610 g005
Figure 6. (a) Cost (×10−3) with varying positive classes on NEU (b) Cost (×10−3) with varying positive classes on Paint.
Figure 6. (a) Cost (×10−3) with varying positive classes on NEU (b) Cost (×10−3) with varying positive classes on Paint.
Sensors 23 02610 g006
Figure 7. (a) Cost (×10−3) with varying weight ratios on NEU (b) Cost (×10−3) with varying weight ratios on Paint.
Figure 7. (a) Cost (×10−3) with varying weight ratios on NEU (b) Cost (×10−3) with varying weight ratios on Paint.
Sensors 23 02610 g007
Figure 8. Training curves of (a) NEU (b) Paint.
Figure 8. Training curves of (a) NEU (b) Paint.
Sensors 23 02610 g008aSensors 23 02610 g008b
Table 1. Risk coefficient grouping.
Table 1. Risk coefficient grouping.
(a): λ N P : λ P N : λ N N : λ P P = 200 : 20 : 2 : 1 (b):   λ N P : λ P N : λ N N : λ P P = 675 : 45 : 3 : 1
(c):   λ N P : λ P N : λ N N : λ P P = 1600 : 80 : 4 : 1 (d):   λ N P : λ P N : λ N N : λ P P = 3125 : 125 : 5 : 1
Table 2. mAP and F1-Score with Different Risk Coefficients.
Table 2. mAP and F1-Score with Different Risk Coefficients.
ModelGroupNEUPaint
mAPF1mAPF1
YOLOv5a0.760.740.610.63
b0.740.720.610.63
c0.760.740.630.63
d0.740.720.620.63
CS-YOLOv5a0.750.740.630.63
b0.760.740.620.63
c0.760.740.620.62
d0.740.720.650.64
Table 3. mAP and F1-score with different positive classes.
Table 3. mAP and F1-score with different positive classes.
ModelGroupNEUPaint
mAPF1mAPF1
YOLOv510.730.720.630.61
20.740.720.630.63
30.740.730.640.65
40.770.74
CS-YOLOv510.750.730.610.63
20.750.720.640.64
30.780.750.620.61
40.740.71
Table 4. mAP and F1-score with different weight ratios.
Table 4. mAP and F1-score with different weight ratios.
Weight RatioNEUPaint
mAPF1mAPF1
00.75 0.73 0.61 0.63
0.250.74 0.72 0.61 0.63
0.50.72 0.70 0.61 0.64
0.750.76 0.73 0.63 0.63
10.68 0.67 0.59 0.62
1.250.68 0.66 0.60 0.61
1.50.68 0.66 0.59 0.59
Table 5. The cost performance on NEU with different batch sizes.
Table 5. The cost performance on NEU with different batch sizes.
Batch size64128 230
cost (×10−3)0.480.470.47
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, B.; Gao, F.; Li, Y. Cost-Sensitive YOLOv5 for Detecting Surface Defects of Industrial Products. Sensors 2023, 23, 2610. https://doi.org/10.3390/s23052610

AMA Style

Liu B, Gao F, Li Y. Cost-Sensitive YOLOv5 for Detecting Surface Defects of Industrial Products. Sensors. 2023; 23(5):2610. https://doi.org/10.3390/s23052610

Chicago/Turabian Style

Liu, Ben, Feng Gao, and Yan Li. 2023. "Cost-Sensitive YOLOv5 for Detecting Surface Defects of Industrial Products" Sensors 23, no. 5: 2610. https://doi.org/10.3390/s23052610

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop