Next Article in Journal
The Kinematic and Electromyographic Analysis of Roller Skating at Different Speeds on a Treadmill: A Case Study
Previous Article in Journal
Emerging Trends and Challenges of IoT in Smart Healthcare Systems, Smart Cities and Education
Previous Article in Special Issue
Beehive Smart Detector Device for the Detection of Critical Conditions That Utilize Edge Device Computations and Deep Learning Inferences
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimizing Automated Optical Inspection: An Adaptive Fusion and Semi-Supervised Self-Learning Approach for Elevated Accuracy and Efficiency in Scenarios with Scarce Labeled Data

by
Yu-Shu Ni
1,
Wei-Lun Chen
1,
Yi Liu
2,
Ming-Hsuan Wu
2 and
Jiun-In Guo
1,*
1
Department of Electronics Engineering, Institute of Electronics, National Yang Ming Chiao Tung University, Hsinchu City 300, Taiwan
2
Software & Service Business Development Program Center, AUO Company, Taichung 407, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(17), 5737; https://doi.org/10.3390/s24175737
Submission received: 30 July 2024 / Revised: 26 August 2024 / Accepted: 2 September 2024 / Published: 4 September 2024
(This article belongs to the Special Issue Deep-Learning-Based Defect Detection for Smart Manufacturing)

Abstract

:
In the field of automatic optical inspection (AOI), this study presents innovative strategies to enhance object detection accuracy while minimizing dependence on large annotated datasets. We initially developed a defect detection model using a dataset of 3579 images across 32 categories, created in collaboration with a major Taiwanese panel manufacturer. This model was evaluated using 12,000 ambiguously labeled images, with improvements achieved through data augmentation and annotation refinement. To address the challenges of limited labeled data, we proposed the Adaptive Fused Semi-Supervised Self-Learning (AFSL) method. This approach, designed for anchor-based object detection models, leverages a small set of labeled data alongside a larger pool of unlabeled data to enable continuous model optimization. Key components of AFSL include the Bounding Box Assigner, Adaptive Training Scheduler, and Data Allocator, which together facilitate dynamic threshold adjustments and balanced training, significantly enhancing the model’s performance on AOI datasets. The AFSL method improved the mean average precision (mAP) from 43.5% to 57.1% on the COCO dataset and by 2.6% on the AOI dataset, demonstrating its effectiveness in achieving high levels of precision and efficiency in AOI with minimal labeled data.

1. Introduction

AOI has become an indispensable technology in the manufacturing sector, ensuring product quality and consistency by automatically detecting defects and anomalies across a wide range of products. As the manufacturing industry continues to advance, the demand for higher efficiency and precision in quality control processes has intensified. Consequently, AOI systems have become crucial for reducing reliance on labor-intensive manual inspections, thus achieving consistent product quality. Figure 1 illustrates typical defect detection results in AOI systems, including common defect types such as surface scratches, discolorations, and structural anomalies. These examples underscore the critical role of AOI in maintaining high production standards by effectively identifying and categorizing a diverse range of defects [1].
Despite these advancements, the development of high-accuracy object detection models in AOI systems still faces significant challenges, particularly the reliance on large, annotated datasets [2,3]. These datasets are essential for training models to accurately identify defects; however, the process of acquiring such extensive labeled data is often time-consuming and costly. This challenge is further compounded in scenarios in which labeling requires expert knowledge or in which defects are subtle and diverse, leading to additional complications in model training. This brings us to a crucial point: the need for innovative methods that can overcome these limitations and enhance the efficiency of AOI systems [4,5,6].
To address these significant hurdles, this study introduces two innovative strategies aimed at enhancing object detection model accuracy while minimizing dependence on extensive annotated datasets. The proposed approach leverages the concepts of semi-supervised learning and adaptive learning, enabling the more efficient use of available data and facilitating self-learning within models [7,8]. These methodologies not only address the limitations of traditional object detection models but also contribute to the broader field of machine learning by demonstrating how self-learning and adaptive strategies can be applied in industrial applications [9].
The AFSL method is a central innovation in this study, designed to optimize AOI systems by improving their ability to learn from both labeled and unlabeled data. This method is particularly relevant given the previously mentioned challenges of limited labeled data and the complexity of defect detection [10,11,12]. AFSL is structured around several key components, each with a specific purpose:
  • Bounding Box Assigner: This component is responsible for generating and refining the bounding boxes used to identify defects within images. It uses an intersection over union (IOU) filter to ensure that only the most reliable bounding boxes are passed on for further training [13]. This step is critical for reducing errors in the early stages of the model’s learning process [14];
  • Adaptive Training Scheduler: To further refine the training process, this scheduler dynamically adjusts the learning rate and other parameters based on the confidence levels of the pseudo-labels generated during training. By doing so, it helps to prevent the model from being misled by unreliable data, which is a common challenge in semi-supervised learning [15,16,17];
  • Data Allocator: This module applies various data augmentation techniques to both labeled and unlabeled datasets, increasing the diversity of the training data. The goal is to expose the model to a wide range of possible defect scenarios, thereby improving its ability to generalize across different types of products and manufacturing conditions [18,19].
While AFSL offers several innovative solutions, it is important to acknowledge the challenges associated with its implementation. One of the primary challenges involves managing the quality of pseudo-labels generated during the semi-supervised learning process. Inaccurate [20,21] or noisy pseudo-labels [22,23] can mislead the model, potentially degrading its performance. Additionally, balancing the training process to ensure that the model effectively utilizes both labeled and unlabeled data without overfitting to the more confident labels is another significant hurdle. These challenges highlight the importance of the careful implementation and ongoing refinement of the AFSL methodology to achieve optimal results.
In the initial phase of this research, a preliminary defect detection model was developed using a dataset of 3579 images across 32 categories, provided by a leading panel manufacturer in Taiwan. The model’s performance was subsequently evaluated on a set of 12,000 ambiguously labeled images, which presented a significant challenge due to the variability and complexity of the defects. Through iterative refinement processes, including data augmentation and annotation correction, the model’s accuracy and generalizability were substantially improved.
Building on these foundational improvements, the second phase introduced the AFSL method to further enhance the self-learning capabilities of the AOI system. While a detailed evaluation of the results, such as the improvements in the model’s mean average precision (mAP), will be discussed in the results section, it is noteworthy that AFSL played a critical role in achieving significant performance gains. Specifically, the mAP improved from 43.5% to 57.1% on the COCO dataset and by 2.6% on an in-house AOI dataset. Furthermore, recent studies have shown that incorporating reinforcement learning and transfer learning into the AFSL framework could further improve defect detection accuracy and adaptability across different manufacturing environments. These preliminary indicators suggest the potential of AFSL to revolutionize AOI systems, which will be explored in greater detail later in this thesis.
The primary objective of this research is to enhance the accuracy and efficiency of object detection models within AOI systems while reducing the reliance on large, annotated datasets. This goal is achieved through the development of the AFSL method, which integrates semi-supervised and adaptive learning techniques to optimize the use of both labeled and unlabeled data [24,25].
The main conclusions of this study highlight the significant improvements in model performance achieved through these methods. While detailed results will be provided in subsequent sections, it is evident that the AFSL approach not only enhances accuracy but also offers a scalable framework for future AOI systems. By addressing the critical challenge of limited labeled data, this research paves the way for more efficient, cost-effective, and scalable AOI solutions in the manufacturing industry.

2. Related Work

The field of AOI has experienced tremendous advancements over the past few decades, driven by the increasing demand for efficient and accurate quality control mechanisms in manufacturing. Various methodologies have been explored to enhance the performance of AOI systems, particularly in the detection and classification of defects. This section provides an in-depth review of the key contributions and methodologies that have shaped the current landscape of AOI, highlighting their strengths and limitations.

2.1. Traditional Machine Learning Approaches

Traditional machine learning techniques have laid the groundwork for defect detection in AOI systems. Techniques such as support vector machines (SVMs), k-nearest neighbors (k-NNs), and decision trees have been extensively utilized for defect classification based on extracted features from images. Kang et al. (2009) employed SVMs to detect defects in thin-film-transistor liquid-crystal display (TFT-LCD) panels, optimizing feature selection and classification parameters to achieve a notable accuracy [26]. Similarly, Iwahori (2017) utilized a k-NN algorithm for defect detection in electronic circuit boards, in which keypoint extraction and CNN features played a critical role in identifying defect patterns [27]. Additionally, Akbar et al. (2013) used decision trees for detecting surface defects, demonstrating the efficacy of rule-based classification methods [28]. Traditional machine learning approaches also explored various feature extraction techniques, such as wavelet transforms and Gabor filters, which were crucial in the early stages of AOI development [29]. However, these methods often required manual feature extraction, which was labor-intensive and limited in capturing the full complexity of defect patterns. This constraint posed challenges in the scalability and adaptability of these models in dynamic manufacturing environments, which this research addresses by automating feature extraction through deep learning techniques. Future trends in this area may involve the integration of traditional machine learning with newer technologies, such as edge computing, to enhance real-time processing capabilities in AOI systems.

2.2. Deep-Learning-Based Methods

The advent of deep learning has brought a paradigm shift in AOI systems, enabling more automated and accurate defect detection. CNNs, in particular, have become the cornerstone of modern AOI systems due to their ability to learn hierarchical feature representations directly from raw image data. Ren et al. (2018) proposed a deep-learning-based approach for automated surface inspection, significantly outperforming traditional methods in terms of accuracy and robustness [30]. Similarly, Yuan-Fu (2019) developed a deep learning model for identifying defect patterns in semiconductor wafers, achieving remarkable improvements over conventional machine learning models [31]. Beyond CNNs, other deep learning architectures like recurrent neural networks (RNNs) and generative adversarial networks (GANs) have been explored for AOI applications. For instance, Zheng et al. (2020) applied a semi-supervised deep learning approach for automated surface inspection, addressing the challenge of data scarcity in AOI systems [32]. GANs have also been used to generate synthetic defect data for training purposes, further tackling the issue of limited labeled datasets [33].
Despite their success, deep learning models require large amounts of annotated data for training, posing a significant challenge in scenarios in which acquiring labeled datasets is expensive and time-consuming. This study builds upon these foundations by introducing semi-supervised and adaptive learning methods to reduce dependency on large labeled datasets, making deep learning more accessible and efficient for AOI applications. Looking ahead, the combination of deep learning with transfer learning and meta-learning could further reduce the need for extensive labeled datasets, accelerating the deployment of AOI systems across various manufacturing environments.

2.3. Semi-Supervised Learning

SSL has emerged as a promising approach to address the limitations of traditional supervised learning by leveraging both labeled and unlabeled data [34]. SSL techniques aim to improve model performance using a limited amount of labeled data, augmented by a larger pool of unlabeled data. Richter et al. (2017) provided a comprehensive review of SSL methods and their applications, demonstrating the effectiveness of SSL in various domains, including AOI [35]. Their study highlighted how SSL can reduce the dependency on labeled data while maintaining a high accuracy. Hung et al. (2022) introduced a semi-supervised learning approach for defect detection in PCBs, showcasing how SSL can be effectively applied to scenarios in which labeled data are scarce and expensive to obtain [36]. Ebayyeh et al. (2020) further explored SSL in the context of PCB quality inspection, integrating a data-expanding strategy to improve defect detection accuracy [37]. The use of pseudo-labeling in SSL, in which the model generates labels for unlabeled data, has also been explored to boost the training process in AOI systems. This method has been shown to significantly enhance defect detection performance in manufacturing lines with limited labeled data [38]. However, SSL is not without its challenges; for instance, model degradation is possible if the pseudo-labels generated are inaccurate. This study addresses these limitations by integrating adaptive learning techniques that dynamically adjust the training process based on the confidence of the pseudo-labels, thereby improving the reliability and robustness of SSL in AOI. Future research could explore the integration of SSL with active learning, in which the model selectively queries the most informative data points for labeling, further enhancing the efficiency of AOI systems.

2.4. Adaptive Learning and Self-Learning Techniques

Adaptive learning and self-learning techniques represent significant advancements in AOI systems, enabling models to continuously improve and adapt to new data. Adaptive learning involves dynamically adjusting model parameters and learning rates based on the characteristics of the data, while self-learning allows models to iteratively refine their predictions using feedback from unlabeled data. Sun et al. (2023) proposed a continual learning framework that adapts defect classification and inspection processes dynamically, significantly improving the model’s performance by focusing on varying defect types [39]. This approach balances the trade-off between precision and recall, making it highly effective in environments in which defect types are diverse and challenging.
Self-learning techniques, such as the use of reinforcement learning for AOI, have also gained attention. Landgraf et al. (2021) applied reinforcement learning to optimize the inspection paths of AOI systems, resulting in faster and more accurate defect detection [40]. Similarly, Jing et al. (2018) developed a computational framework that combines coverage planning with reinforcement learning for automatic path generation in robotic inspection tasks, further enhancing the efficiency and accuracy of AOI systems [41]. These techniques have demonstrated their effectiveness in improving the robustness and generalizability of AOI systems, particularly in environments with limited labeled data. However, adaptive and self-learning methods can be complex to implement and may require significant computational resources.
This research contributes by optimizing these techniques for use in AOI, making them more practical and scalable in industrial applications. In the future, advancements in hardware, such as the use of specialized AI chips, could further enhance the efficiency of adaptive and self-learning methods in AOI, enabling real-time adaptation and decision making.

2.5. Hybrid Approaches

Recent research has also explored hybrid approaches that combine multiple methodologies to leverage their respective strengths. Schlosser et al. (2022) developed a hybrid model that integrates deep learning with traditional machine learning techniques for defect detection in semiconductor manufacturing. This approach combined the feature extraction capabilities of a CNN with the classification accuracy of traditional machine learning methods, resulting in a highly effective defect detection system [42]. Similarly, Chu et al. (2022) proposed a hybrid-learning framework that utilizes both deep learning and edge computing for operational visual quality inspection within Internet of Things (IoT) systems. Their method effectively combines the strengths of different approaches to enable real-time defect detection in manufacturing environments [43].
These hybrid models offer the potential to overcome the limitations of single-method approaches, providing more versatile and powerful solutions for AOI. Despite their benefits, hybrid models can be challenging to design and require the careful balancing of the different techniques involved. This study advances the field by proposing a hybrid approach that effectively integrates semi-supervised learning with adaptive learning, thereby overcoming the limitations of existing models and improving defect detection accuracy in AOI systems. Looking forward, the combination of hybrid models with cloud computing and the IoT could revolutionize AOI by enabling more flexible, scalable, and connected inspection systems [44].

2.6. Challenges and Future Directions

While significant progress has been made in AOI systems, several challenges remain. One of the primary challenges is the scarcity of labeled data, which limits the training of deep learning models. Additionally, the variability and complexity of defect patterns pose significant challenges for model generalization. This research addresses these challenges by leveraging semi-supervised and adaptive learning techniques that minimize dependency on labeled data while enhancing model generalization. Moreover, the integration of domain knowledge and expert insights into AI models can further enhance their performance and reliability. Leveraging advancements in explainable AI (XAI) could also provide valuable insights into the decision-making processes of AOI systems, enabling better interpretability and more trust in automated inspections. In summary, the advancements in machine learning, particularly deep learning, semi-supervised learning, and adaptive learning, have significantly enhanced the capabilities of AOI systems. However, ongoing research and innovation are essential to overcome the remaining challenges and unlock the full potential of AOI technology in industrial applications. By continuously refining these methodologies and exploring new hybrid approaches, we can pave the way for more accurate, efficient, and scalable AOI systems in the future.

3. Research Methodology

In this study, we focused on developing a robust AOI system starting from the initial stages of the manufacturing process. The proposed approach involved multiple phases, including initial dataset assembly, model development, data augmentation, semi-supervised learning implementation, and thorough evaluation. This comprehensive design aimed to enhance the accuracy of defect detection while minimizing reliance on extensive annotated datasets. The complete process of the proposed method is illustrated in Figure 2, showcasing everything from early data establishment to subsequent model development, which involves creating a foundational dataset and model, validating labeled images, and then integrating both labeled and unlabeled images to establish a self-learning model. This allows us to achieve better results using a minimal amount of labeled data.

3.1. Initial Dataset Assembly

The initial dataset was provided by our industrial collaborators and consisted of 3579 images encompassing 32 defect categories. These images, captured by high-resolution AOI systems, included various defect types such as scratches, discolorations, and structural anomalies. Each image was meticulously annotated by expert inspectors to ensure high-quality labels. This dataset served as the foundation for our initial defect detection model.
The collected data have been modified in order to avoid exposing company secrets. However, to ensure the repair status is clearly identified, the defect area of defective image has been left unmodified.
The first version of our defect detection model was constructed using the initial dataset. We employed a YOLOv7 [45] base model architecture tailored for object detection tasks. The key components of the model included the following:
  • Input Layer: High-resolution images (input size 512 × 512 pixels);
  • Feature Extraction Layers: Multiple convolutional layers with ReLU activation functions and max-pooling layers;
  • Region Proposal Network (RPN): Generates region proposals where defects are likely to be located;
  • Classification and Regression Heads: Predicts defect categories and refines bounding box coordinates.
The model was trained using the initial dataset with a stochastic gradient descent (SGD) optimizer at a learning rate of 0.001. The loss function combined classification loss (cross-entropy) and localization loss (smooth L1 loss). Training was conducted over 50 epochs with a batch size of 16.
Following the initial training, the model was evaluated using a set of 12,000 ambiguously labeled images. This evaluation aimed to assess the model’s baseline performance and verify the accuracy of the initial annotations. Key performance metrics such as precision, recall, F1-score, and mAP were computed to gauge the model’s effectiveness.

3.2. Dataset Adjustment

After refining the initial dataset and model, we identified some minor errors. Although in-plane and out-of-plane defects were categorized together, their characteristics were significantly different, as shown in Figure 3. We separated the original in-plane defects from the GOA region and the Fan-out region, and finally used post-processing to reclassify the categories. This method helps to improve the overall accuracy of the classification.

3.3. Data Augmentation and Annotation Refinement

To enhance the dataset’s diversity and improve the model’s generalizability, various data augmentation techniques were employed. These included the following:
  • Scaling: Random scaling within a range of 0.8 to 1.2 times the original size as shown in Figure 4;
  • Flipping: Horizontal and vertical flips;
  • Brightness and Contrast Adjustment: Random changes in brightness and contrast levels;
  • Color Adjustment: Applying different color mappings to simulate various imaging conditions.
In addition to data augmentation, the ambiguously labeled dataset underwent an iterative annotation refinement process. Expert inspectors reviewed and corrected the annotations based on the model’s predictions, ensuring high-quality labels. This refinement included re-annotating bounding boxes and adjusting annotation sizes to improve precision.

3.4. Implementation of Adaptive Fusion Semi-Supervised Self-Learning

To address the limitations of labeled data, we implemented the AFSL method. AFSL is designed to enhance the model’s ability to learn from both labeled and unlabeled data, facilitating continuous improvement and reducing human annotation efforts. The following sections will discuss the key components of AFSL.

3.5. Semi-Supervised Learning Object Detection

Firstly, AFSL will be introduced in detail. Figure 5 shows the architecture of AFSL. The labeled dataset also participates in AFSL training to guide model training. Our proposal integrates three blocks with several innovations into AFSL based on the Mean Teacher [46] teacher–student architecture. These are the Bounding Box Assigner, Adaptive Training Scheduler, and various training strategies. First, a robust pseudo-label needs to be generated, so the robustness of the complex model is crucial. AFSL fine-tunes the complex model to generate robust pseudo-labels. Once the complex model is sufficiently robust, self-learning training fine-tunes the target model. The ΔmAP is used to judge the AFSL complex model’s fine-tuning. If ΔmAP is larger than 1% in the first three epochs, it indicates the complex model needs improvement for the newly added dataset, and AFSL continues training. Conversely, if ΔmAP is less than 1% in the first three epochs, it means the complex model is robust enough, and AFSL will stop, starting the one-shot self-learning algorithm.
The following sub-sections will dive into the details of each AFSL method.

3.5.1. Complex Model Selection

Before diving into the architecture of AFSL, the backbone of AFSL will be introduced. To generate robust pseudo-labels, a powerful model is as crucial as an excellent semi-supervised learning algorithm. This paper focuses on the object detection anchor-based model due to its efficiency and robustness. YOLOv7 [45], one of the powerful and lightweight representative object detection anchor-based models, is known for its performance on open dataset validation and its integrated training methodology with previous augmentation and optimizer strategies.
Table 1 shows the performance and complexity comparison of YOLOv7 with other object detection models. YOLOv7 has higher AP50:95 and less model complexity. Additionally, YOLOv7 has an improved performance due to applying several training strategies, such as Mosaic, which improve AP@50 by 1.8%.

3.5.2. Bounding Box Assigner

Similar to traditional supervised learning, semi-supervised learning for the student model contains pseudo-labels. An adaptive unlabeled loss is proposed, replacing fixed coefficients with bounding box confidence. We proposed a dynamic and adaptive unlabeled loss function, improving the training process and balancing student model learning from labeled and unlabeled data. The proposed method prevents student models from learning unreliable pseudo-labels, ensuring effective training.
The Adaptive Training Scheduler includes both labeled loss and unlabeled loss. It optimizes the unlabeled loss function for semi-supervised object detection by using confidence weighting, allowing the model to assess the quality of pseudo-labels. This approach improves the model’s efficiency and accuracy during training. Additionally, a debiasing penalty is incorporated to reduce uncertainties during training, balancing the influence of labeled and unlabeled data.

3.5.3. Data Allocator and Adaptive Training Scheduler

In semi-supervised object detection, implementing an effective training strategy that works in tandem with the Bounding Box Assigner and Adaptive Training Scheduler is essential for maintaining training stability. This thesis introduces both strong and weak augmentation techniques for a consistency-based approach on the unlabeled dataset, aiding the inference processes of the teacher and student models.
Figure 6 depicts the block diagram of the Adaptive Fusion Semi-Supervised Learning system, highlighting the use of weak and strong augmentations for input data. Weak augmentation techniques, such as horizontal flips and Gaussian noise, are used to modify the unlabeled data without altering essential image features. For strong augmentation, methods like Mosaic with HSV adjustments, Mixup, and Cutout—techniques utilized in YOLOv7—are employed to enhance the student model’s training. Figure 7 provides an example of a training image after applying Mosaic augmentation.
Additionally, we proposed using the Efficient Teacher method to ensure smoother fitting of the student model during training [47]. Early in the training process, high learning rates and improperly fitted unlabeled data can result in empty labels for challenging frames. To address this, we isolate these empty-labeled frames from the training process of the current epoch and reintegrate them once they are properly labeled.

4. Performance Evaluation and Validation

In this section, we present a comprehensive evaluation of the performance and validation of our object detection models. We utilized various techniques to enhance the dataset’s diversity and improve the model’s generalizability. The proposed approach includes detailed data augmentation and annotation refinement, which are crucial steps to ensure the robustness and accuracy of the models. The effectiveness of these techniques is demonstrated through a performance comparison of different editions of our models, including the YOLOv7 model.

4.1. Data Augmentation and Annotation Refinement

To improve the model’s performance, various data augmentation techniques were employed, and the ambiguously labeled dataset underwent iterative annotation refinement. These steps were crucial in enhancing the model’s accuracy and generalizability as shown in Table 2 and Table A1 included in Appendix A.

4.2. AFSL Validation on COCO and AOI Datasets

The following results validate the proposed AFSL method, starting with the COCO dataset and followed by the AOI dataset.

4.2.1. COCO Dataset Validation

In this section, the COCO Standard dataset serves as a benchmark for validating the AFSL methodology. Before we dive into the experimental results, Equation (1) introduces an evaluation matrix designed to normalize the results from other research studies for comparison. This equation is particularly important in scenarios like AOI, in which both model performance and computational complexity play a crucial role, especially given the constraints of edge computing.
The relationship between model performance and complexity is critical in AOI systems because these systems often operate in environments in which real-time processing is required, and computational resources are limited. By normalizing the complexity of the model (measured in FLOPs) and the labeled data ratio, Equation (1) allows us to compare the efficiency and effectiveness of our proposed AFSL method against other approaches. This comparison is especially relevant for semi-supervised object detection tasks, for which reducing the amount of labeled data while maintaining or improving performance is a key objective.
Equation (1) is formulated as follows:
E M A F S L = m a p @ 0.5 0.95 I m p r o m e n t ( % ) × 100 L a b e l e d   R a t i o ( % ) × M o d e l   C o m p l e x i t y ( G   F L O P s )
This equation helps demonstrate that a model with lower FLOPs that consumes fewer computational resources and requires a shorter training period can achieve a superior performance in semi-supervised object detection. The normalization of the labeled ratio is optional when the labeled data ratio is the same across different models; in such cases, we only normalize the model’s complexity to focus on the computational efficiency.
The inclusion of model complexity, specifically measured in FLOPs, is justified by the practical demands of AOI systems, which often rely on edge computing. In these scenarios, it is essential that the models are not only accurate but also computationally efficient. Edge devices typically have limited processing power, and a model with a lower computational footprint (i.e., lower FLOPs) is more suitable for real-time applications. Thus, Equation (1) provides a comprehensive evaluation by considering both the accuracy and the computational demands of the model, ensuring that the proposed AFSL method is both effective and efficient in real-world AOI applications.
To further validate the scientific correctness of Equation (1), we will conduct additional experiments comparing the results of different models on the COCO dataset using this evaluation metric, as shown in Table 3. These experiments will help confirm that the equation appropriately reflects the trade-offs between accuracy and computational efficiency, particularly in the context of edge computing in AOI systems.
As shown in Table 3, we observe that the performance of our proposed AFSL not only excels in the one-stage anchor-based method but also achieves the highest EMAFSL value across all SSOD methodologies when complexity is considered. The method with the highest score other than AFSL in the 10% labeled data category is Dense Teacher [48], which achieved a score of 7.01. However, our proposed AFSL method had a score of 8.63, which is 1.62 points higher than that of Dense Teacher [48]. This indicates that the AFSL methodology we proposed is highly competitive in terms of performance and training time consumption.
Table 3. The performance for CNN-based semi-supervised learning on object detection under 10% labeled COCO Standard dataset (* means that the method is conducted on the backbone of Efficient Teacher [47]).
Table 3. The performance for CNN-based semi-supervised learning on object detection under 10% labeled COCO Standard dataset (* means that the method is conducted on the backbone of Efficient Teacher [47]).
MethodologyFLOPsAP@50:95 (%)EMAFSL
Two-stage anchor-basedBaseline202.31G23.6
STAC [24]28.642.49
Instant Teaching [49]30.403.36
Humble Teacher [50]31.613.96
Unbiased Teacher [51]31.503.90
Soft Teacher [52]34.045.16
LabelMatch [53]35.495.87
PseCo [54]36.066.15
One-stage anchor-freeBaseline200.59G23.07
Unbiased Teacher v2 [55]32.614.76
DSL [56]36.226.56
Dense Teacher [48]37.137.01
One-stage anchor-freeBaseline169.61G24.04-
Unbiased Teacher * [57]30.343.71
Efficient Teacher [47]34.095.92
Baseline104.7G25.16-
AFSL (ours)34.208.63
Fully Supervised49.7-
As detailed in Table 4, the AFSL training improved the [email protected] from 43.5% to 57.1% when using 10% labeled data, compared to the baseline model. Additionally, we incorporated the COCO Additional dataset to explore whether supplementing with more data could further enhance the model’s performance. The results indicated that adding the COCO Additional dataset improved the [email protected]:0.95 by 1.5%, compared to the fully supervised model trained solely on the COCO Standard dataset, demonstrating AFSL’s significant potential for continuous learning.

4.2.2. AOI Dataset Validation

Furthermore, we conducted experiments on the AOI dataset provided by AUO, as shown in Table 5. We experimented with 10% and 25% labeled datasets. The results demonstrate that even with just 10% labeled data, AFSL exhibits a high level of adaptability and continuous improvements. When the labeled dataset proportion was increased to 25%, the AFSL-trained model outperformed the fully supervised results, further proving AFSL’s powerful adaptive training capabilities.
Table 6 and Table A2 show the data for various categories in the AOI dataset based on 25% labeled data after AFSL training. The black bars represent the fully supervised results, and the white bars represent the AFSL training results. The table indicates that AFSL not only improved the [email protected]:0.95 by 2.6%, but also maintained stable training across various categories.
Most categories performed slightly better in comparison to the fully supervised results, confirming that AFSL effectively captures the characteristics of each category. Additionally, AFSL can handle data imbalance to some extent, mitigating the issues faced by traditional semi-supervised object detection methods.

5. Conclusions

The proposed method significantly enhanced the accuracy of AOI systems while reducing reliance on extensive annotated datasets. Beginning with an initial dataset of 3579 images spanning 32 defect categories, we developed and refined a defect detection model, which was subsequently validated on 12,000 ambiguously labeled images. Through iterative processes of data augmentation and annotation refinement, the model’s accuracy saw a substantial improvement.
Our experiments demonstrated that the AFSL method achieved remarkable enhancements. On the COCO dataset, the AFSL method increased [email protected]:0.95 from 43.5% to 51.7% using only 10% labeled data. Similarly, on the AOI dataset, the AFSL method with 25% labeled data outperformed the fully supervised results, showing a 2.6% improvement in [email protected]:0.95. Despite these advancements, challenges such as the scarcity of labeled data and the complexity of defect patterns persist. Future research should focus on developing more effective data augmentation techniques and integrating domain knowledge. In the AOI field, particularly in defect detection, relying solely on image data often proves insufficient due to the heavy reliance on expert experience. Thus, combining multi-modal models with self-learning techniques could potentially yield even more significant results.
In conclusion, this paper proposes a scalable and efficient approach for AOI systems, markedly improving defect detection accuracy and efficiency. The AFSL method provides a robust solution for reducing annotation efforts while achieving a high precision, laying a strong foundation for future advancements in manufacturing quality control and automation.

Author Contributions

Conceptualization, Y.-S.N., W.-L.C. and J.-I.G.; methodology, Y.-S.N., W.-L.C. and J.-I.G.; software, Y.-S.N. and W.-L.C.; validation, Y.-S.N. and J.-I.G.; formal analysis, Y.-S.N., W.-L.C. and J.-I.G.; investigation, Y.-S.N., W.-L.C. and J.-I.G.; resources, J.-I.G.; data curation, Y.-S.N., W.-L.C., Y.L., M.-H.W. and J.-I.G.; writing—original draft preparation, Y.-S.N.; writing—review and editing, Y.-S.N., Y.L., M.-H.W. and J.-I.G.; visualization, J.-I.G.; supervision, J.-I.G.; project administration, J.-I.G.; funding acquisition, J.-I.G. All authors have read and agreed to the published version of the manuscript.

Funding

We would like to extend our sincere thanks for the funding provided by the following: the National Science and Technology Council (NSTC), Taiwan R.O.C. projects with grants 111-2221-E-A49 -126 -MY3, 112-2218-E-A49-027 -, 113-2218-E-002-013 -, 113-2640-E-A49-009 -, and 113-2218-E-A49-028 -, and the Satellite Communications and AIoT Research Center/The Co-operation Platform of the Industry-Academia Innovation School, National Yang Ming Chiao Tung University (NYCU) through Taiwan R.O.C. projects with grants 112UC2N006.

Institutional Review Board Statement

Ethical review and approval were waived for this study because it did not involve any human or animal subjects, and therefore did not require ethical approval.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

We would like to express our sincere gratitude to all those who contributed to this work. The support provided by the National Science and Technology Council (NSTC), Taiwan R.O.C., and the Satellite Communications and AIoT Research Center/The Co-operation Platform of the Industry-Academia Innovation School, National Yang Ming Chiao Tung University (NYCU) was invaluable.

Conflicts of Interest

Author Yi Liu and Ming-Hsuan were employed by the AUO Corporation. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Appendix A

Table A1. Detailed performance comparison of YOLOv7 with different refinement stages.
Table A1. Detailed performance comparison of YOLOv7 with different refinement stages.
Defect NameOriginalFirst
Edition
Second
Edition
Final
Edition
Type179.70%76.60%84%87.1%
Type287.50%85.20%100%100%
Type357.20%82.10%88%97%
Type497.80%82.70%93%93%
Type595.30%82.90%86%100%
Type689.60%99.40%100%98.1%
Type7100.00%100.00%100%100%
Type897.70%60.90%74%91.3%
Type989.20%86.70%100%100%
Type1089.60%92.50%93%93%
Type1185.40%85.30%94%94%
Type1283.80%93.30%100%100%
Type1383.70%96.10%100%100%
Type1477.40%83.70%100%100%
Type1594.50%94.00%98%98%
Type1692.70%100.00%100%100%
Type1781.10%75.90%93%93%
Type1888.20%30.80%56%75%
Type1978.70%75.50%77%84%
Type2074.20%76.30%76%76%
Type2174.40%82.80%86%86%
Type2275.80%87.10%93%93%
Type2368.80%88.00%82%82%
Type2472.70%70.20%84%84%
Type2586.70%86.90%97%97%
Type2688.50%93.30%94.6%94.6%
Type2767.30%58.90%82%87%
Type2887.30%89.20%95%95%
Type2984.70%87.30%98%98%
Type3085.60%69.10%83%92%
Type31100.00%100.00%100%100%
Type3271.70%78.40%92%93.5%
Total84.00%83.50%91.30%93.88%

Appendix B

Table A2. Detailed performance metrics of various defect types between baseline model and AFSL model.
Table A2. Detailed performance metrics of various defect types between baseline model and AFSL model.
Defect NamePrecisionPrecisionAP (0.5)AP (0.5)AP (0.5-0.95)AP (0.5-0.95)
BaselineAFSLBaselineAFSLBaselineAFSL
Total0.9210.9270.9680.970.7040.73
Type10.8430.8690.9660.9660.5880.633
Type20.8790.910.9860.9930.6720.711
Type30.9670.950.9930.9920.9920.987
Type40.97210.9920.9950.7110.746
Type50.9780.9880.9920.9930.9880.988
Type60.9770.9780.9950.9950.6950.732
Type70.9920.9920.9950.9950.8020.842
Type80.9990.9980.9950.9950.9950.994
Type90.7670.8230.9660.970.7370.775
Type100.9290.9340.9760.9810.5840.607
Type110.9430.9470.9630.9710.6560.712
Type120.7880.8190.890.9120.5670.594
Type1310.9970.9950.950.8320.847
Type140.959670.9730.9820.5210.554
Type15O.9270.9360.990.9930.750.773
Type160.8050.7880.7510.7480.5230.521
Type170.7370.7690.8260.8590.4860.524
Type180.9370.9360.9780.9810.8030.822
Type190.9030.9150.9760.9740.580.616
Type200.9340.9640.9780.9820.8130.837
Type210.8590.8620.9730.9760.6410.66
Type22110.9950.9950.8370.859
Type230.9940.9930.9950.9950.7520.762
Type240.9020.9050.9780.9850.6820.728
Type250.8810.890.9790.9390.6120.63
Type260.9570.9420.9910.9920.8270.846
Type270.950.9590.9850.9940.6550.696
Type280.9970.39980.9950.9950.7540.804
Type290.9190.8770.9840.9730.590.622
Type300.7420.9330.9460.9490.5850.608
Type31110.9950.9950.5460.597
Type320.8490.8390.9860.9810.6940.733

References

  1. Noroozi, M.; Ghadermazi, J.; Shah, A.; Zayas-Castro, J.L. Toward Optimal Defect Detection in Assembled Printed Circuit Boards Under Adverse Conditions. IEEE Access 2023, 11, 127119–127131. [Google Scholar] [CrossRef]
  2. Rezaei, A.; Nau, J.; Streitferdt, D.; Schambach, J.; Vangelov, T. ReProInspect: Framework for Reproducible Defect Datasets for Improved AOI of PCBAs. In Engineering of Computer-Based Systems: 8th International Conference, ECBS 2023, Västerås, Sweden, 16–18 October 2023, Proceedings; Springer: Berlin/Heidelberg, Germany, 2023; pp. 205–214. [Google Scholar] [CrossRef]
  3. Lin, H.-I.; Wibowo, F.S. Image Data Assessment Approach for Deep Learning-Based Metal Surface Defect-Detection Systems. IEEE Access 2021, 9, 47621–47638. [Google Scholar] [CrossRef]
  4. Ye, R.; Chang, M.; Pan, C.-S.; Chiang, C.A.; Gabayno, J.L. High-resolution optical inspection system for fast detection and classification of surface defects. Int. J. Optomechatronics 2018, 12, 1–10. [Google Scholar] [CrossRef]
  5. Phoulady, A.; Suleiman, Y.; Choi, H.; Moore, T.; May, N.; Shahbazmohamadi, S.; Tavousi, P. Synthetic data augmentation to enhance manual and automated defect detection in microelectronics. Microelectron. Reliab. 2023, 150, 115220. [Google Scholar] [CrossRef]
  6. Dai, W.; Mujeeb, A.; Erdt, M.; Sourin, A. Towards Automatic Optical Inspection of Soldering Defects. In Proceedings of the 2018 International Conference on Cyberworlds (CW), Singapore, 3–5 October 2018; pp. 375–382. [Google Scholar] [CrossRef]
  7. Dupre, R.; Fajtl, J.; Argyriou, V.; Remagnino, P. Improving Dataset Volumes and Model Accuracy With Semi-Supervised Iterative Self-Learning. IEEE Trans. Image Process. 2020, 29, 4337–4348. [Google Scholar] [CrossRef]
  8. Dópido, I.; Li, J.; Marpu, P.R.; Plaza, A.; Dias, J.M.B.; Benediktsson, J.A. Semisupervised Self-Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 51, 4032–4044. [Google Scholar] [CrossRef]
  9. Ahmad, H.M.; Rahimi, A. Deep learning methods for object detection in smart manufacturing: A survey. J. Manuf. Syst. 2022, 64, 181–196. [Google Scholar] [CrossRef]
  10. Ghamisi, A.; Charter, T.; Ji, L.; Rivard, M.; Lund, G.; Najjaran, H. Anomaly detection in automated fibre placement: Learning with data limitations. Front. Manuf. Technol. 2024, 4, 1277152. [Google Scholar] [CrossRef]
  11. Lv, X.; Duan, F.; Jiang, J.-J.; Fu, X.; Gan, L. Deep Active Learning for Surface Defect Detection. Sensors 2020, 20, 6. [Google Scholar] [CrossRef] [PubMed]
  12. Pham, T.T.A.; Thoi, D.K.T.; Choi, H.; Park, S. Defect Detection in Printed Circuit Boards Using Semi-Supervised Learning. Sensors 2023, 23, 6. [Google Scholar] [CrossRef]
  13. Hurtik, P.; Vajgl, M. BBRefinement: An Universal Scheme to Improve Precision of Box Object Detectors. October 2020. Available online: https://openreview.net/forum?id=RB0iNPXIj60 (accessed on 22 August 2024).
  14. Cruciata, G.; Presti, L.L.; La Cascia, M. Iterative Multiple Bounding-Box Refinements for Visual Tracking. J. Imaging 2022, 8, 61. [Google Scholar] [CrossRef] [PubMed]
  15. Li, K.; Lian, Q.; Gao, C.; Zhang, F. AatMatch: Adaptive Adversarial Training in Semi-Supervised Learning Based on Data-Driven Decision-Making Models. Systems 2023, 11, 5. [Google Scholar] [CrossRef]
  16. Xie, Q.; Luong, M.-T.; Hovy, E.; Le, Q.V. Self-training with Noisy Student improves ImageNet classification. arXiv 2020, arXiv:1911.04252. [Google Scholar] [CrossRef]
  17. Oh, S.; Jeong, J.; Lee, C.-G.; Yoo, J.; Nam, G. Synergistic Training: Harnessing Active Learning and Pseudo-Labeling for Enhanced Model Performance in Deep Learning. WSEAS Trans. Comput. 2023, 22, 114–119. [Google Scholar] [CrossRef]
  18. Mertes, S.; Margraf, A.; Kommer, C.; Geinitz, S.; André, E. Data Augmentation for Semantic Segmentation in the Context of Carbon Fiber Defect Detection using Adversarial Learning. Delta 2020. Available online: https://publica.fraunhofer.de/handle/publica/410037 (accessed on 29 July 2024).
  19. Kim, E.K.; Lee, H.; Kim, J.Y.; Kim, S. Data Augmentation Method by Applying Color Perturbation of Inverse PSNR and Geometric Transformations for Object Recognition Based on Deep Learning. Appl. Sci. 2020, 10, 3755. [Google Scholar] [CrossRef]
  20. Rizve, M.N.; Duarte, K.; Rawat, Y.S.; Shah, M. In Defense of Pseudo-Labeling: An Uncertainty-Aware Pseudo-label Selection Framework for Semi-Supervised Learning. arXiv 2021, arXiv:2101.06329. [Google Scholar] [CrossRef]
  21. Chen, H.; Tao, R.; Fan, Y.; Wang, Y.; Wang, J.; Schiele, B.; Xie, X.; Raj, B.; Savvides, M. SoftMatch: Addressing the Quantity-Quality Trade-off in Semi-supervised Learning. arXiv 2023, arXiv:2301.10921. [Google Scholar] [CrossRef]
  22. Jin, Z.; Zhong, D.; Song, X.; Liu, Z.; Ye, N.; Zeng, Q. Filter and evolve: Progressive pseudo label refining for semi-supervised automatic speech recognition. arXiv 2022, arXiv:2210.16318. [Google Scholar] [CrossRef]
  23. Sun, Y.; Gao, C. NCMatch: Semi-supervised Learning with Noisy Labels via Noisy Sample Filter and Contrastive Learning. In Pattern Recognition and Computer Vision; Liu, Q., Wang, H., Ma, Z., Zheng, W., Zha, H., Chen, X., Wang, L., Ji, R., Eds.; Springer Nature: Singapore, 2024; pp. 15–27. [Google Scholar] [CrossRef]
  24. Sohn, K.; Zhang, Z.; Li, C.-L.; Zhang, H.; Lee, C.-Y.; Pfister, T. A Simple Semi-Supervised Learning Framework for Object Detection. arXiv 2020, arXiv:2005.04757. [Google Scholar] [CrossRef]
  25. Rosenberg, C.; Hebert, M.; Schneiderman, H. Semi-Supervised Self-Training of Object Detection Models. In Proceedings of the 2005 Seventh IEEE Workshops on Applications of Computer Vision (WACV/MOTION’05)–Volume 1, Breckenridge, CO, USA, 5–7 January 2005; pp. 29–36. [Google Scholar] [CrossRef]
  26. Kang, S.B.; Lee, J.H.; Song, K.Y.; Pahk, H.J. Automatic defect classification of TFT-LCD panels using machine learning. In Proceedings of the 2009 IEEE International Symposium on Industrial Electronics, Seoul, Republic of Korea, 5–8 July 2009; pp. 2175–2177. [Google Scholar] [CrossRef]
  27. Iwahori, Y. Defect Detection and Classification of Electronic Circuit Boards Using Keypoint Extraction and CNN Features. 2017. Available online: https://www.semanticscholar.org/paper/Defect-Detection-and-Classification-of-Electronic-Iwahori/c7771f41941cf3a146ff61022594f9e5348bcf70 (accessed on 22 August 2024).
  28. Akbar, H.; Suryana, N.; Akbar, F. Surface Defect Detection and Classification Based on Statistical Filter and Decision Tree. Int. J. Comput. Theory Eng. 2013, 5, 774–779. [Google Scholar] [CrossRef]
  29. Reshadat, V.; Kapteijns, R.A.J.W. Improving the Performance of Automated Optical Inspection (AOI) Using Machine Learning Classifiers. In Proceedings of the 2021 International Conference on Data and Software Engineering (ICoDSE), Bandung, Indonesia, 3–4 November 2021; pp. 1–5. [Google Scholar] [CrossRef]
  30. Ren, R.; Hung, T.; Tan, K.C. A Generic Deep-Learning-Based Approach for Automated Surface Inspection. IEEE Trans. Cybern. 2018, 48, 929–940. [Google Scholar] [CrossRef] [PubMed]
  31. Yuan-Fu, Y. A Deep Learning Model for Identification of Defect Patterns in Semiconductor Wafer Map. In Proceedings of the 2019 30th Annual SEMI Advanced Semiconductor Manufacturing Conference (ASMC), Saratoga Springs, NY, USA, 6–9 May 2019; pp. 1–6. [Google Scholar] [CrossRef]
  32. Zheng, X.; Wang, H.; Chen, J.; Kong, Y.; Zheng, S. A Generic Semi-Supervised Deep Learning-Based Approach for Automated Surface Inspection. IEEE Access 2020, 8, 114088–114099. [Google Scholar] [CrossRef]
  33. Bhatt, P.M.; Malhan, R.K.; Rajendran, P.; Shah, B.C.; Thakar, S.; Yoon, Y.J.; Gupta, S.K. Image-Based Surface Defect Detection Using Deep Learning: A Review. J. Comput. Inf. Sci. Eng. 2021, 21, 040801. [Google Scholar] [CrossRef]
  34. Paliulis, E.; Zemblys, R.; Daunys, G. IMAGE ANALYSIS PROBLEMS IN AOI SYSTEMS. Inf. Technol. Control. 2008, 37, 3. Available online: https://itc.ktu.lt/index.php/ITC/article/view/11947 (accessed on 22 August 2024).
  35. Richter, J.; Streitferdt, D.; Rozova, E. On the development of intelligent optical inspections. In Proceedings of the 2017 IEEE 7th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 9–11 January 2017; pp. 1–6. [Google Scholar] [CrossRef]
  36. Hung, Y.-H. Developing an Anomaly Detection System for Automatic Defective Products’ Inspection. Processes 2022, 10, 8. [Google Scholar] [CrossRef]
  37. Ebayyeh, A.A.R.M.A.; Mousavi, A. A Review and Analysis of Automatic Optical Inspection and Quality Monitoring Methods in Electronics Industry. IEEE Access 2020, 8, 183192–183271. [Google Scholar] [CrossRef]
  38. Yin, A.; Fung, A. Effective analysis of optical inspection machines (IMPACT 2011). In Proceedings of the 2011 6th International Microsystems, Packaging, Assembly and Circuits Technology Conference (IMPACT), Taipei, Taiwan, 19–21 October 2011; pp. 408–410. [Google Scholar] [CrossRef]
  39. Sun, W.; Kontar, A.R.; Jin, J.; Chang, T.-S. A continual learning framework for adaptive defect classification and inspection. J. Qual. Technol. 2023, 55, 598–614. [Google Scholar] [CrossRef]
  40. Landgraf, C.; Meese, B.; Pabst, M.; Martius, G.; Huber, M.F. A Reinforcement Learning Approach to View Planning for Automated Inspection Tasks. Sensors 2021, 21, 6. [Google Scholar] [CrossRef]
  41. Jing, W.; Goh, C.F.; Rajaraman, M.; Gao, F.; Park, S.; Liu, Y.; Shimada, K. A Computational Framework for Automatic Online Path Generation of Robotic Inspection Tasks via Coverage Planning and Reinforcement Learning. IEEE Access 2018, 6, 54854–54864. [Google Scholar] [CrossRef]
  42. Schlosser, T.; Friedrich, M.; Beuth, F.; Kowerko, D. Improving automated visual fault inspection for semiconductor manufacturing using a hybrid multistage system of deep neural networks. J. Intell. Manuf. 2022, 33, 1099–1123. [Google Scholar] [CrossRef]
  43. Chu, Y.; Feng, D.; Liu, Z.; Zhao, Z.; Wang, Z.; Xia, X.G.; Quek, T.Q. Hybrid-Learning-Based Operational Visual Quality Inspection for Edge-Computing-Enabled IoT System. IEEE Internet Things J. 2022, 9, 4958–4972. [Google Scholar] [CrossRef]
  44. Schlosser, T.; Beuth, F.; Friedrich, M.; Kowerko, D. A Novel Visual Fault Detection and Classification System for Semiconductor Manufacturing Using Stacked Hybrid Convolutional Neural Networks. In Proceedings of the 2019 24th IEEE International Conference on Emerging Technologies and Factory Automation (ETFA), Zaragoza, Spain, 10–13 September 2019; pp. 1511–1514. [Google Scholar] [CrossRef]
  45. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar] [CrossRef]
  46. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. Adv. Neural Inf. Process. Syst. 2017, 30, 1195–1240. [Google Scholar]
  47. Xu, B.; Chen, M.; Guan, W.; Hu, L. Efficient Teacher: Semi-Supervised Object Detection for YOLOv5. arXiv 2023, arXiv:2302.07577. [Google Scholar] [CrossRef]
  48. Zhou, H.; Ge, Z.; Liu, S.; Mao, W.; Li, Z.; Yu, H.; Sun, J. Dense Teacher: Dense Pseudo-Labels for Semi-supervised Object Detection. In Computer Vision—ECCV 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer Nature: Cham, Switzerland, 2022; pp. 35–50. [Google Scholar] [CrossRef]
  49. Zhou, Q.; Yu, C.; Wang, Z.; Qian, Q.; Li, H. Instant-Teaching: An End-to-End Semi-Supervised Object Detection Framework. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 4079–4088. [Google Scholar] [CrossRef]
  50. Tang, Y.; Chen, W.; Luo, Y.; Zhang, Y. Humble Teachers Teach Better Students for Semi-Supervised Object Detection. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE Computer Society, Nashville, TN, USA, 20–25 June 2021; pp. 3131–3140. [Google Scholar] [CrossRef]
  51. Liu, Y.-C.; Ma, C.Y.; He, Z.; Kuo, C.W.; Chen, K.; Zhang, P.; Wu, B.; Kira, Z.; Vajda, P. Unbiased Teacher for Semi-Supervised Object Detection. arXiv 2021, arXiv:2102.09480. [Google Scholar] [CrossRef]
  52. Xu, M.; Zhang, Z.; Hu, H.; Wang, J.; Wang, L.; Wei, F.; Bai, X.; Liu, Z. End-to-End Semi-Supervised Object Detection with Soft Teacher. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 3040–3049. [Google Scholar] [CrossRef]
  53. Chen, B.; Chen, W.; Yang, S.; Xuan, Y.; Song, J.; Xie, D.; Pu, S.; Song, M.; Zhuang, Y. Label Matching Semi-Supervised Object Detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 14361–14370. [Google Scholar] [CrossRef]
  54. Li, G.; Li, X.; Wang, Y.; Wu, Y.; Liang, D.; Zhang, S. PseCo: Pseudo Labeling and Consistency Training for Semi-Supervised Object Detection. In Computer Vision—ECCV 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer Nature: Cham, Switzerland, 2022; pp. 457–472. [Google Scholar] [CrossRef]
  55. Liu, Y.-C.; Ma, C.-Y.; Kira, Z. Unbiased Teacher v2: Semi-supervised Object Detection for Anchor-free and Anchor-based Detectors. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE Computer Society, New Orleans, LA, USA, 18–24 June 2022; pp. 9809–9818. [Google Scholar] [CrossRef]
  56. Chen, B.; Li, P.; Chen, X.; Wang, B.; Zhang, L.; Hua, X.-S. Dense Learning based Semi-Supervised Object Detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 4805–4814. [Google Scholar] [CrossRef]
  57. Schmutz, H.; Humbert, O.; Mattei, P.-A. Don’t fear the unlabelled: Safe semi-supervised learning via simple debiasing. arXiv 2022, arXiv:2203.07512. [Google Scholar] [CrossRef]
Figure 1. Detection result of different defects in AOI.
Figure 1. Detection result of different defects in AOI.
Sensors 24 05737 g001
Figure 2. The proposed system workflow.
Figure 2. The proposed system workflow.
Sensors 24 05737 g002
Figure 3. The differences between the (a) in-plane region, (b) the GOA region, and (c) the Fan-out region.
Figure 3. The differences between the (a) in-plane region, (b) the GOA region, and (c) the Fan-out region.
Sensors 24 05737 g003
Figure 4. Random scaling of the image.
Figure 4. Random scaling of the image.
Sensors 24 05737 g004
Figure 5. Proposed AFSL system architecture diagram.
Figure 5. Proposed AFSL system architecture diagram.
Sensors 24 05737 g005
Figure 6. The block diagram of mean bounding box loss.
Figure 6. The block diagram of mean bounding box loss.
Sensors 24 05737 g006
Figure 7. The training image for student after YOLOv7 Mosaic was implemented.
Figure 7. The training image for student after YOLOv7 Mosaic was implemented.
Sensors 24 05737 g007
Table 1. The performance and complexity comparison of YOLOv7 and other object detection models.
Table 1. The performance and complexity comparison of YOLOv7 and other object detection models.
MethodResolutionMosaicParam.FLOPsAP50:95 (%)MethodResolutionMosaicParam.
FRCNN[1333,800]X39.8M202.31G40.3FRCNN[1333,800]X39.8M
FCOS[1333,800]X32.02M200.59G38.5FCOS[1333,800]X32.02M
YOLOv5w/o[640,640]X42.56M109.59G41.2YOLOv5w/o[640,640]X42.56M
YOLOv5[640,640]42.56M109.59G49YOLOv5[640,640]42.56M
YOLOv7[640,640]37.62M106.59G51.5YOLOv7[640,640]37.62M
RetinaNet[1333,800]X37.74M239.32G39.5RetinaNet[1333,800]X37.74M
DenseNet[640,640]42.13M169.61G44.86DenseNet[640,640]42.13M
Table 2. Performance comparison of YOLOv7 with different refinement stages.
Table 2. Performance comparison of YOLOv7 with different refinement stages.
DatasetsModel Accuracy
No modifications84.00%
First Edition (Category Refinement)83.50%
Second Edition (Category Refinement and Data Augmentation)91.30%
Final Edition (Category Refinement, Data Augmentation, and Full Data Reorganization)93.88%
Table 4. The performance and complexity comparison of YOLOv7 and other object detection models [51] using COCO dataset for AFSL validation with 10% labeled data.
Table 4. The performance and complexity comparison of YOLOv7 and other object detection models [51] using COCO dataset for AFSL validation with 10% labeled data.
PrecisionRecallAP
@0.5
AP
@0.5:0.95
Yolov7_640 × 640
(Baseline 10% labeled)
35.40%31.80%43.50%25%
Yolov7_640 × 640
(AFSL 10% labeled)
51.70%45.30%57.10%38.10%
Yolov7_640 × 640 (AFSL 100% labeled, additional)62.50%58.70%70.4%50.5%
Yolov7_640 × 640
(Fully Supervised)
63.20%56.20%69.70%49%
Table 5. AOI dataset for AFSL validation with different ratios of labeled data.
Table 5. AOI dataset for AFSL validation with different ratios of labeled data.
Labeled RatioPrecisionRecallAP
@0.5
AP
@0.5-0.95
Yolov7_640 × 640(Baseline 10% labeled)70.70%76.50%78.80%46.70%
Yolov7_640 × 640(Semi-supervised 10% labeled)79.60%83.70%84.7%
(+5.9%)
61.4%
(+14.7%)
Yolov7_640 × 640(Semi-supervised 10% labeled)81.40%82.20%86.70%53.20%
Yolov7_640 × 640(Semi-supervised 25% labeled)92.70%95.40%97%
(+0.2%)
73%
(+2.6%)
Yolov7_640 × 640 (Fully Supervised)92.10%94.60%96.80%70.40%
Table 6. AFSL training under 25% labeled AOI dataset: a comparison between supervised learning (Baseline) and AFSL.
Table 6. AFSL training under 25% labeled AOI dataset: a comparison between supervised learning (Baseline) and AFSL.
ItemPrecisionPrecisionAP@
0.5
AP@
0.5
AP@
0.5:0.95
AP@
0.5:0.95)
MethodBaselineAFSLBaselineAFSLBaselineAFSL
Accuracy0.9210.9270.9680.970.7040.73
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ni, Y.-S.; Chen, W.-L.; Liu, Y.; Wu, M.-H.; Guo, J.-I. Optimizing Automated Optical Inspection: An Adaptive Fusion and Semi-Supervised Self-Learning Approach for Elevated Accuracy and Efficiency in Scenarios with Scarce Labeled Data. Sensors 2024, 24, 5737. https://doi.org/10.3390/s24175737

AMA Style

Ni Y-S, Chen W-L, Liu Y, Wu M-H, Guo J-I. Optimizing Automated Optical Inspection: An Adaptive Fusion and Semi-Supervised Self-Learning Approach for Elevated Accuracy and Efficiency in Scenarios with Scarce Labeled Data. Sensors. 2024; 24(17):5737. https://doi.org/10.3390/s24175737

Chicago/Turabian Style

Ni, Yu-Shu, Wei-Lun Chen, Yi Liu, Ming-Hsuan Wu, and Jiun-In Guo. 2024. "Optimizing Automated Optical Inspection: An Adaptive Fusion and Semi-Supervised Self-Learning Approach for Elevated Accuracy and Efficiency in Scenarios with Scarce Labeled Data" Sensors 24, no. 17: 5737. https://doi.org/10.3390/s24175737

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop