Next Article in Journal
Effect of Difference of Sensory Modality in Cognitive Task on Postural Control During Quiet Stance
Previous Article in Journal
A Multi-Task Causal Knowledge Fault Diagnosis Method for PMSM-ITSF Based on Meta-Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Contrastive Learning with Global and Local Representation for Mixed-Type Wafer Defect Recognition

by
Shantong Yin
,
Yangkun Zhang
and
Rui Wang
*
School of Mechanical Engineering and Automation, Harbin Institute of Technology, Shenzhen 518055, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(4), 1272; https://doi.org/10.3390/s25041272
Submission received: 8 January 2025 / Revised: 12 February 2025 / Accepted: 18 February 2025 / Published: 19 February 2025
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
Recognizing defect patterns in semiconductor wafer bin maps (WBMs) poses a critical challenge in the integrated circuit (IC) manufacturing industry. The accurate classification and segmentation of these defect patterns are of utmost significance as they are key to tracing the root causes of defects, thereby reducing costs and enhancing both product efficiency and quality. As the manufacturing process grows in complexity, the WBM becomes intricate when multiple defect patterns coexist on a single wafer, making the recognition task increasingly complicated. In addition, traditional supervised learning methods require a large number of labeled samples, which is labor-intensive. In this paper, we present a self-supervised contrastive learning framework for the classification and segmentation of mixed-type WBM defect patterns. Our model incorporates a global module for contrastive learning that captures image-level representations, alongside a local module that targets the comprehension of regional details, which is helpful for the segmentation of defective patterns. Experimental results demonstrate that our model performs effectively in scenarios where there is a limited number of labeled examples and a wealth of unlabeled ones.

1. Introduction

The production of semiconductor devices is a complex and highly detailed operation, involving hundreds of steps to fabricate integrated circuit (IC) chips on silicon wafers. After wafer fabrication, the wafer needs to be inspected. Common inspection methods include a probing test, scanning electron microscopy (SEM), transmission electron microscopy (TEM), X-ray inspection, and automatic optical inspection (AOI), each generating information at different levels. For probing tests, the electrical performance of each die is tested using wafer probes. Dies that meet quality requirements are cut off from the wafer and utilized as final qualified chips. As the worldwide need for these devices grows, manufacturers are concentrating their efforts on boosting manufacturing efficiency by employing techniques such as maximizing the number of chips per wafer. Nevertheless, the enhanced complexity of these advanced techniques also gives rise to various defects, which are inevitable during the manufacturing processes. Therefore, the results of wafer test are stored in wafer bin maps (WBMs) for defect analysis, contributing to reducing defect rates and ensuring market competitiveness [1].
WBM defect pattern recognition is crucial for detecting systemic manufacturing issues, as the accurate classification and segmentation results of WBM defect patterns can effectively infer the root causes of defects in the manufacturing process [2]. The defects on the wafer can generally be divided into two categories: random errors and systematic faults. When dealing with wafer defect patterns, the analysis of systematic faults is generally the main focus, as the defect patterns caused by random errors are irregular and difficult to eliminate. As shown in Figure 1a, the typical single defect patterns contain the following: “Center”, “Donut”, “EdgeLoc”, “EdgeRing”, “Loc”, “NearFull”, “Scratch”, and “Random” [3]. With the ongoing advancement in the IC industry, the complexity of manufacturing raises the likelihood of mixed-type WBM defect patterns appearing on the same wafer [4,5], as shown in Figure 1b. These defect patterns may interact with each other, resulting in more intricate issues. Consequently, accurate classification and segmentation of these mixed-type WBM defect patterns are essential for effective quality control.
Given the intricate nature of mixed-type WBM defect patterns, manual inspection becomes increasingly labor-intensive and time-consuming [6]. Alternative solutions are required in industrial production. Recent studies have explored the application of artificial intelligence and computer vision models for defect pattern analysis [2]. However, existing research mainly focuses on image classification and cannot accurately identify the size and location of mixed-type defect patterns. In addition, training a model to recognize defect patterns often requires a large number of labeled samples and undoubtedly wastes manpower and resources. Self-supervised learning can learn from unlabeled data, which is particularly attractive in situations where it is difficult to obtain large amounts of labeled data [7]. Although self-supervised learning is currently mainly used for image classification tasks, our work is based on the intuition that the WBM defect pattern segmentation can also benefit from representations learned through self-supervised learning from unlabeled data.
This study centers on contrastive learning, a branch of self-supervised learning [8,9]. According to our research, most existing contrastive learning approaches prioritize global representations, which are not optimal for segmentation tasks that require pixel-level distinctions [10]. If local regions of images are extracted for contrastive learning, then local representations can be adequately learned [11] for segmentation tasks. Consequently, we propose a self-supervised contrastive learning model that integrates both global and local contrastive learning modules, with queue mechanisms introduced to decouple the sample size from mini-batch constraints. Specifically, the global contrastive learning module focuses on the learning of global representations, while the local contrastive learning module is used to learn pixel-level discrimination. With the introduction of queues, each module receives a sufficient number of negative samples, improving the contrastive loss computation. To the best of our knowledge, this is the first application of contrastive learning to both the classification and segmentation of WBM defect patterns. The primary contributions of this paper are as follows:
(1)
We employ self-supervised contrastive learning for both the classification and segmentation of mixed-type defect patterns. In this way, we can obtain more precise information about the types and locations of different defect patterns.
(2)
Our proposed model integrates global and local contrastive learning modules, enabling the learning of both image-level and pixel-level features directly from unlabeled data. This approach provides a robust foundation for downstream classification and segmentation tasks using only a limited amount of labeled data.
(3)
Experimental evaluations on the mixed-type WBM defect pattern dataset demonstrate that our framework outperforms existing self-supervised methods.
The remainder of this paper is organized as follows: Section 2 presents a brief overview of recent studies on mixed-type WBM defect pattern classification and segmentation. Section 3 details the proposed methodology. Section 4 discusses the experimental results and analysis. Finally, Section 5 contains concluding remarks and directions for future research.

2. Related Works

2.1. WBM Defect Pattern Recognition

WBM defect pattern analysis is crucial for identifying the root cause of process failures in semiconductor manufacturing. The current research mainly focuses on the classification of single defect patterns and can be roughly divided into three categories. The first category mainly monitors defect patterns on WBMs by constructing statistical information [12,13], which can successfully separate normal and abnormal graphic patterns, but it is difficult to distinguish different defect patterns. The second category is model-based clustering methods. For example, Wang et al. [14] used Gaussian EM to detect elliptical and linear patterns and used the shell algorithm to estimate annular patterns. Yuan and Kuo [15] proposed a mixed model describing the distribution of defects and used Bayesian inference to estimate parameters. The last category is machine learning methods that have received more and more attention in recent years. Wang and Chen [16] trained a multilayer perceptron (MLP) on a set of hand-crafted spatial filters that reflect the rotational invariance property of WBM. Kulkarni [3] applied CNNs to classify defect patterns based on a synthesized WBM dataset. Kang et al. [17] presented a semi-supervised representation learning method, which fully utilizes the information in both unlabeled and labeled wafer maps. Most of the above research relies on large WBM databases and has problems with unsatisfactory performance and high computational costs.
The composition of mixed-type WBM defect patterns is more complex, as the combination of defect patterns can vary greatly. Thus, research on the classification of mixed-type WBM defect patterns is less extensive than that on single WBM defect patterns. In these studies, Byun and Baek [18] proposed a new approach using a convolutional autoencoder to initialize the weights of a CNN. Tello et al. [19] used feature rules to determine whether a wafer map has mixed-type patterns and then used a stochastic general regression neural network (GRNN) and a deep structured CNN for classification. Kyeong and Kim [20] applied a CNN to classify WBM defects with mixed-type defect patterns. Wang et al. [5] proposed DC-Net, which used a deformable convolutional network (DC-Net) to classify mixed-type defect patterns. These methods are unable to recognize and locate mixed-type defect patterns on WBMs simultaneously, and the research on the separation of multiple defect patterns is still insufficient.
Compared with classification, image segmentation can provide more detailed information and better understand the content of the image. For mixed-type WBM defect pattern segmentation, Nag et al. [21] recently proposed a novel network based on an encoder–decoder architecture for the simultaneous classification and segmentation of single and mixed defect patterns. Chiu et al. [2] proposed a method based on Mask R-CNN [22] to identify WBM defect patterns. Yan et al. [23] presented a new U-Net framework that used semantic segmentation methods to segment different defect patterns on WBMs. However, existing research tends to focus more on global image representation while overlooking the powerful auxiliary role that local information plays in image segmentation tasks.

2.2. Self-Supervised Contrastive Learning

Self-supervised learning mainly uses auxiliary tasks to mine its own supervision information from large-scale unsupervised data and trains the network through this constructed supervision information [7,24]. As shown in Figure 2, the core idea of the self-supervised learning paradigm is to first learn knowledge from unlabeled samples by designing self-supervised signals and then transfer it to downstream tasks. As a branch of self-supervised learning, contrastive learning [25] focuses on learning the common features between instances of the same class and distinguishing the differences between instances of different classes. Contrastive learning can effectively utilize unlabeled data for training [9,26] and has been widely applied in the field of natural image processing. In the field of WBM defect pattern classification, Kwak et al. [27] applied a method (SWaCo) for safe WBM classification with self-supervised contrastive learning to effectively exploit unlabeled data. Kahng et al. [28] presented a self-supervised learning framework that fully utilizes unlabeled data to pre-learn rich visual representations for data-efficient WBM defect pattern classification. Hu et al. [29] proposed a contrastive learning framework for semi-supervised learning and prediction of WBM defect patterns. It can be seen that self-supervised contrastive learning methods do not heavily rely on the number of labeled samples and are mostly used for the classification of WBM defect patterns, but there is still little research on segmentation tasks. The urgent challenge lies in establishing a self-supervised contrastive learning model that possesses both classification and segmentation capabilities for mixed-type WBM defect patterns.

3. Methodology

3.1. Overview

In this paper, we propose a classification and segmentation model for mixed-type WBM defect patterns based on self-supervised contrastive learning. In our problem setting, we consider a set of WBM data including N labeled data and M unlabeled data, where N M . For simplicity, the unlabeled set is defined as D u = { x i } i = 1 M , and the labeled set is defined as D l = { x i , y i } i = M + 1 M + N . As shown in Figure 3, the framework consists of a self-supervised pre-training stage and a supervised fine-tuning stage. In pre-training, global contrastive learning module is designed to learn image-level representation, while the local contrastive learning module is used to better understand the structure of local regions, which contributes to the image segmentation task. To distinguish extracted features with different connotations, we apply the concept of a query and a key to describe encoders and decoders, although they have the same structure respectively. A query can be understood as the target to be searched for, and a key is the template. The model is built by matching the query to a dictionary of keys, which is a queue of the encoded representations of data samples that is updated with each mini-batch. In fine-tuning, pre-trained weights are transferred to the downstream classification and segmentation tasks and fine-tuned with a limited number of labeled samples. This section describes the proposed framework in detail.

3.2. Data Augmentation

In contrastive learning, the model is trained by enforcing similarity between positive sample pairs and dissimilarity between negative sample pairs. Therefore, the core of contrastive learning is the construction of positive and negative samples. In practice, different augmented views of the same sample after data augmentation are seen as positive samples, while other samples are defined as negative samples. Supported by data augmentation, the model can learn better feature representations of the data, capture the fundamental structure and relationships between samples, and effectively prevent over-fitting. As shown in Figure 4, we perform data augmentation operators, i.e., random rotation, flipping, and random noise on samples for learning the rotational invariance of WBMs and improving the robustness and generalization ability of the model. Regarding the noise addition operations in data augmentation, given that noise in semiconductor manufacturing environments (such as dust, light variations, etc.) can affect the performance of detection systems, we introduce different types of artificial noise (Gaussian Noise, Salt-and-Pepper Noise, and Poisson Noise) to simulate real-world production noise conditions. Specifically, given unlabeled sample x i D u , two augmented WBMs x i I and x i II are generated by data augmentation t 1 and t 2 , i.e., x i I = t 1 ( x i ) and x i II = t 2 ( x i ) . These two augmented views are then used for the illustration of subsequent feature extraction.

3.3. Global Contrastive Learning Module

The global contrastive learning module mainly promotes the model to learn the ability of extracting global features. Since global features provide the context information of the image, the designing of this module contributes to the effective comprehension of the overall structure of mixed-type WBMs. Specifically, key and query encoders are used to extract high-dimensional key and query global features. Then, these high-dimensional features are mapped to lower-dimensional hidden spaces for a more compact representation through the global projection head. After that, the global queue that is a unique form of a linear list is employed to dynamically manage key features. The construction of the global contrastive learning module is similar to MoCo v2 [8]; we provide a detailed description as follows.

3.3.1. Global Feature Extraction

Global features refer to the overall attributes of an image, capable of capturing the global information and understanding the image from a holistic perspective. We apply the concept of a query and a key to describe encoders for distinguishing extracted features with different connotations. As given a series of key features and a query feature, contrastive learning is to find the unique key feature that matches the query feature the most and enforce them to be similar while distancing from others.
Firstly, a key encoder e k ( · ) and a query encoder e q ( · ) are used to obtain global key feature G i I and global query feature G i II from the augmented sample instances, respectively. The procedure can be formulated as
G i I = μ e k x i I , G i II = μ e q x i II ,
where μ ( · ) represents the average value of each channel in the feature map, i.e., the global average pooling. e k ( · ) denotes the key encoder and e q ( · ) denotes the query encoder, which have the same structure as the downstream encoder e ( · ) .
In addition, the global projection head h G ( · ) is an MLP with one hidden layer (with ReLU) in this module, whose presence has been proven to be beneficial to push the dissimilar samples away from each other and obtain more useful information for downstream tasks [9]. The global projection head can map the high-dimensional features G i I and G i II to lower-dimensional hidden spaces for a more compact representation, encouraging our model to learn more abstract and informative features. Therefore, we obtain the final global key feature g i I and global query feature g i II by passing G i I and G i II through the global projection head h G ( · ) to further calculate the global contrastive loss. The conversion can be formulated as follows:
g i I = h G G i I = W G ( 2 ) R G W G ( 1 ) G i I ,
g i II = h G G i II = W G ( 2 ) R G W G ( 1 ) G i II ,
where W G ( 1 ) and W G ( 2 ) are linear transformation matrices, and R G is a ReLU nonlinearity.
Furthermore, the global queue that is a unique form of a linear list for dynamically managing vectors is introduced, as shown in Figure 3. Since the core of contrastive learning is to construct sufficient negative sample pairs, we design the global queue to store abundant negative vectors (i.e., key features) generated by the key encoder. The size of key features in the global queue can far exceed the capacity of a single batch, which can increase the diversity of training and improve the generalization ability of the model. Therefore, the introduction of the queue can decouple the negative vector size from the mini-batch size. In the dynamic management of the global queue, the current mini-batch of encoded key features is enqueued into the queue and the oldest key features are dequeued from the queue. This process ensures that the model is always learning from the latest data by a dynamic and efficient management of key features.

3.3.2. Global Contrastive Loss

Global contrastive loss is a measure of the similarities between global features in a representation space and promotes the model to learn global representations. In contrastive learning, if a query feature (generated by query encoder) is given, only a unique key feature in the global queue can match it. So, the idea of contrastive learning is to compare the query feature with all the key features in the queue and hope that the query feature is as similar as possible to its matching key feature, and then it stays away from others. Consider an encoded query feature g i II and a set of encoded samples { g 0 I , g i I , g K I } that contains K + 1 key features in the global queue. Obviously, there is a single key feature (denoted as g i I ) that matches g i II . Then, the contrastive loss function, called InfoNCE [30], is used to expect matching pairs to be similar and mismatching pairs to be dissimilar, as follows:
L G = 1 N G i = 1 N G log exp g i II · g i I / τ g k = 0 K exp g i II · g k I / τ g ,
where τ g is a temperature hyper-parameter, and N G denotes the number of samples from a mini-batch. The sum in the denominator is calculated over one positive key feature g i I and other K negative key features. Intuitively, the InfoNCE loss tries to classify the query feature g i II as its corresponding positive key feature g i I rather than others.

3.4. Local Contrastive Learning Module

The local contrastive learning module is a local contrastive representation space for computing the contrastive loss associated with the local regions of a WBM. For a WBM with mixed-type defect patterns, the segmentation task is more complex due to the interconnected nature of defects and their variability in quantity, position, and type, etc. While global features can effectively capture the overall information of an image, relying solely on global features to measure WBM defect patterns will result in the neglect of crucial local details. This approach may not be optimal for segmentation tasks that require pixel-level discrimination. To address this issue, we design the local contrastive learning module to learn the representation of local regions. We separate the WBM into multiple regions, selectively identify relevant regions, and subsequently perform position matching and feature extraction for local contrastive learning. This method ensures that the segmentation process benefits from both the global context and the fine-grained local information, enhancing the overall accuracy and effectiveness of defect recognition. The details are as follows:

3.4.1. Local Region Selection and Matching

Randomly selected local regions will be biased towards the more dominant feature categories, as the “background” category accounts for a large proportion in mixed-type WBM defect patterns. Therefore, we need to make the selection of local regions reasonable to contain more valuable feature categories. For example, as shown in Figure 5, flipping and rotating operations are performed on x i , and two augmented views x i I and x i II are obtained respectively. The selected local regions are then matched according to their positions.
The local region selection and matching is performed by the selection of appropriate regions in x i I at first, and then by the determination of the local regions in x i II that match with the selected regions. The appropriateness of each local region to be selected is evaluated by its weight, which is calculated by a convolutional layer with the kernel size being the same as the local region size n s × n s . Specifically, x i I is input into the convolutional layer, and the weight of each local region is calculated and recorded by sliding the kernel. We select the local region with the maximum weight from x i I , and then the position of the matching local region in x i II is determined. To minimize overlap, after each local region selection, the region is excluded so that the centers of subsequently selected regions do not fall into the previously selected local regions. This weighted selection and matching process is repeated n r times, selecting distinct local regions in each iteration to obtain n r pairs of matched local regions from images x i I and x i II .

3.4.2. Local Feature Extraction

Local features l j I and l j II that capture the pixel-level discrimination of WBM defect patterns are extracted by the following steps. Firstly, feature maps d k ( e k ( x i I ) ) and d q ( e q ( x i II ) ) of a positive pair ( x i I , x i II ) are obtained from a key and query encoder–decoder network. Here, e k ( · ) and e q ( · ) are key and query encoders while d k ( · ) and d q ( · ) are key and query decoders, which have the same structure as the downstream encoder–decoder network, respectively. Then, the j-th ( j = 1 , , n r ) pair of matching local region feature maps L j I and L j II is extracted according to the idea of local region selection and matching. After that, average pooling μ ( · ) is performed on local region feature maps L j I and L j II , and the results then pass through the local projection head h L ( · ) . Therefore, the final local features can be defined as follows:
l j I = h L μ L j I = W L ( 2 ) R L W L ( 1 ) μ L j I ,
l j II = h L μ L j II = W L ( 2 ) R L W L ( 1 ) μ L j II ,
where feature maps d k ( e k ( x i I ) ) and d q ( e q ( x i II ) ) have the same size as the original image. The local projection head h L ( · ) can contribute to making the model better capture the differences between local regions. μ represents calculating the average value of each channel in feature maps, W L ( 1 ) and W L ( 2 ) are linear transformation matrices, and R L is a ReLU nonlinearity. Similar to the global contrastive learning, the queue is also applied to dynamically store and manage local region vectors in the training of the model, as shown in Figure 3.

3.4.3. Local Contrastive Loss

The local contrastive loss updates the network by forcing the feature representations of matching local regions to be similar and the feature representations of different local regions to be dissimilar. Specifically, considering a local query feature l j II and a set of vectors { l 0 I , l j I , l Γ I } that contain Γ + 1 key features in the local queue, there is a single key feature (denoted as l j I ) that matches l j II . Then, the local contrastive loss can also be calculated by InfoNCE as follows:
L L = 1 N L j = 1 N L log exp l j II · l j I / τ l γ = 0 Γ exp l j II · l γ I / τ l ,
where τ l is a temperature hyper-parameter, N L indicates the number of pairs of matched local regions in a mini-batch, and Γ represents the number of key features in the local queue.

3.5. Network Updating

3.5.1. Self-Supervised Pre-Training

In the pre-training period, the total contrastive loss function to update the query encoder parameters θ e q and query decoder parameters θ d q is defined as follows:
L P = λ 1 L G + ( 1 λ 1 ) L L ,
where L G represents the global contrastive loss in Equation (4), and L L represents the local contrastive loss in Equation (7).
Although using queues allows for the size of negative vectors to be larger, it also poses difficulties for updating the key encoder and key decoder via back-propagation, since the gradient should propagate to all samples in the queue. To address this issue, a momentum update approach is utilized for the key encoder and decoder, which accumulates exponential moving averages of the key network to ensure a stable and efficient update manner. To be specific, the query encoder parameters θ e q and the decoder parameters θ d q are updated via back-propagation by Equation (8), while the key encoder parameters θ e k and decoder parameters θ d k are updated by the momentum equations:
θ e k α 1 θ e k + ( 1 α 1 ) θ e q ,
and
θ d k α 2 θ d k + ( 1 α 2 ) θ d q ,
where α 1 and α 2 [ 0 , 1 ) are the momentum coefficients. The momentum updating makes parameters θ e k and θ d k evolve more smoothly than θ e q and θ d q . Therefore, although the keys in the queue are generated by different encoders and decoders, the differences between these encoders and decoders can be ignored.

3.5.2. Supervised Fine-Tuning

Supervised fine-tuning refers to training the model on a smaller and task-specific labeled dataset. The purpose of this stage is to make the model adapt to the specific requirements of the given task. In detail, the model retains the knowledge learned in the pre-training stage while also learning additional information relevant to the specific task. In our problem, given labeled data { x i , y i } D l , we fine-tune the whole network through a segmentation loss L S and classification loss L C .
For the computation of the segmentation loss L S , we employ the cross-entropy loss function [31], which computes the pixel-level difference between segmented defect patterns and the ground truth. However, merely calculating the loss value of pixels may result in over-segmentation. It is still necessary to compute the loss value associated with the type of defect. We use a C-bit one-hot value to represent the predicted types of WBM defect patterns and calculate the corresponding binary cross entropy loss L C . Since we need to simultaneously focus on the classification and segmentation of WBM defect patterns, the total loss function in the fine-tuning period is defined as follows:
L F = λ 2 L S + ( 1 λ 2 ) L C ,
where L S represents the segmentation loss and L C represents the classification loss.

4. Experimental Results

4.1. Data Description

A wafer bin map (WBM) is the result of the circuit probing process, recording the spatial distribution of defective chips on the wafer. The WBMs dataset collection system is shown in Figure 6. The primary method of wafer testing involves the coordination between the tester and the probe stage [32]. During the testing process, the tester cannot directly measure the wafer under test. Instead, it relies on the probes in the probe card to make electrical contact with the pads or bumps on the wafer. The test signals measured by the probes are then sent to the automatic test equipment (ATE) for analysis and judgment, thereby obtaining the electrical characteristics test results for each die on the wafer. All samples are collected in actual industrial scenarios by the wafer probe testing station. We standardized the image size to 52 × 52 .
As shown in Table 1, the defect types contain 8 kinds of single defect patterns and 29 kinds of mixed-type defect patterns (13 kinds of double mixed-type defect patterns, 12 kinds of triple mixed-type defect patterns, and 4 kinds of quadruple mixed-type defect patterns). We represent each type of defect using the uppercase initial letter of its name as shown in Figure 1. It is worth noting that Random and NearFull defects do not significantly point to specific manufacturing roots, and their impact in practice is minimal. Also, these two types of patterns are not suitable for generating mixed defects with other single defect patterns, so we do not consider the synthesis of these two types. Furthermore, we consider the WBMs of the normal category (C0) into the model training process, and there are 37 kinds of defect patterns and 1 kind of normal category in total. In the final dataset, we collected 19,000 WBMs with 500 instances for each defect pattern. In our problem setting, the original WBMs have only pixel values 0, 1, and 2, so x i 0 , 1 , 2 H × W mentioned earlier is the WBM; y i = m i , s i m i { 0 , 1 } C × H × W , s i { 0 , 1 } C is the corresponding ground truth category label, where C denotes the number of pattern types; m i is the pixel label map of the i-th sample; and s i represents the C-bit one-hot label.

4.2. Model Evaluation

In this paper, we focus on both the segmentation and classification performance of mixed-type WBM defect patterns. Specifically, we use P a c c (pixel predicted accuracy) and I o U (Intersection over Union) to evaluate the results of semantic segmentation, which can be defined for a specific class c as follows:
P a c c i c = A r e a ( m ^ i c m i c ) H × W ,
I o U i c = A r e a ( m ^ i c m i c ) A r e a ( m ^ i c m i c ) ,
where m i c and m ^ i c denote the actual and predicted pixel label maps of the c-th pattern in the i-th WBM. For mixed-type patterns, the mean I o U ( m I o U ) and the mean P a c c ( m P a c c ) are used to evaluate the model segmentation performance.
For the evaluation of classification results, we use A c c u r a c y to measure the classification accuracy of mixed-type WBM defect patterns. To evaluate the performance in the false classification of WBMs, P r e c i s i o n and R e c a l l are employed in our experiments. The definitions of the indices are shown as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N ,
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P T P + F N .
where T P denotes the true positives, T N is the true negatives, F P represents the false positives, and F N represents false negatives.

4.3. Performance and Analysis

4.3.1. Implementation Details

In this work, the encoders and decoders designed in our model have the same structure as those in DeepLab v3+ [33]. During the pre-training period, we used an abundant amount of unlabeled WBMs to pre-train the encoder and decoder for 350 epochs with a batch size of 64. We optimized the parameters with Adam optimizer, and the initial learning rate was set as 0.01 with the cosine decay schedule. In addition, we selected 6 local regions of size 12 × 12 from each sample in the local region selection and matching process (i.e., n s = 12 , n r = 6 ). The hyper-parameter λ 1 in Equation (8) was set to 0.5 to balance the global contrast loss and local contrast loss. Following [8], the momentum coefficients α 1 and α 2 were all set to 0.999, and the size of the queues was set to 2048. After pre-training, the model with the lowest loss was saved for the downstream segmentation task.
In the fine-tuning period, we trained our model using a small amount of labeled WBMs for a total of 150 epochs, with a batch size of 32. The initial learning rate was set to 0.001, which decreased to 98% per epoch. The hyper-parameter λ 2 in Equation (11) is used to balance pixel loss and one-hot prediction loss, so we also set it to a default 0.5. In addition, the experiment was implemented on an equipment with an Intel(R) Xeon(R) Gold 6226R @ 2.90 GHz CPU and an NVIDIA RTX A6000 GPU by Python 3.8.3.

4.3.2. Model Performance

In our experimental setting, we randomly split the dataset into three subsets: 80% for pre-training, 10% for fine-tuning, and 10% for testing. We obtained the P a c c , I o U , A c c u r a c y , P r e c i s i o n , and R e c a l l for 38 kinds of defect patterns to demonstrate the performance of our model. Table 2 primarily displays the segmentation and classification results for the normal pattern and the single-type defect patterns, Table 3 shows the segmentation results for the mixed-type defect patterns, and Table 4 presents the classification results for the mixed-type defect patterns. As shown in Table 2, the horizontal axis represents different categories (C0 to C8), while the vertical axis displays various evaluation metrics for segmentation and classification. The average P a c c , I o U , A c c u r a c y , P r e c i s i o n , and R e c a l l of our proposed model for C0–C8 are 95.78%, 93.87%, 96.80%, 94.61%, and 94.89%. In Table 3 and Table 4, to facilitate the presentation of the experimental results, the horizontal axis represents various evaluation metrics for segmentation and classification, respectively, while the vertical axis corresponds to different categories (C9 to C37). The average P a c c , I o U , A c c u r a c y , P r e c i s i o n , and R e c a l l for C9–C37 are 95.40%, 90.51%, 92.34%, 94.45%, and 94.66%. From an overall perspective, the more the defect classes that appear on the same WBM, the more difficult it is to recognize and segment. However, whether it is a mixture of two defects, three defects, or four defects, our model accurately accomplished the recognition and segmentation of WBMs in the testing set.
Figure 7 presents several mixed-type WBM segmentation results achieved with our proposed model. We randomly selected eight experimental samples, each displaying the original mixed-type WBM, label map, and corresponding segmentation output. We can see that the segmentation maps obtained using our model are smooth, complete, and robust to noise. Even in the cases where several mixed defect patterns are overlapped, our model can still distinguish them based on the different characteristics of each defect pattern. For all these defect patterns, the C and E R classes are best classified and segmented. This is because the location and shape of the C and E R defect patterns are relatively fixed and easier to distinguish. Even for the S class, which may easily interact with other pattern classes, our model still has an excellent segmentation performance. Therefore, our model has great potential and application value in the field of WBMs classification and segmentation.

4.3.3. Comparison with Other Methods

In this section, we compare the performance of our proposed model with other self-supervised learning models, including Jigsaw [34], SimCLR [9], and MoCo v2 [8]. These are all relatively successful contrastive learning models. To ensure fairness in comparison, all models used in the experiments have the same dataset and division ratio as in the previous section.
As shown in Table 5, the horizontal axis represents the overall evaluation metrics for segmentation and classification, including m P a c c , m I o U (used for segmentation), and A c c u r a c y , P r e c i s i o n , and R e c a l l (used for classification), while the vertical axis displays the numerical results for each method across these metrics. The overall m P a c c , m I o U , A c c u r a c y , P r e c i s i o n , and R e c a l l values of our model are as high as 95.59%, 92.19%, 93.57%, 94.43%, and 94.78%, respectively. These indicators are far superior to the other three self-supervised contrastive learning methods. SimCLR performs the worst, with the overall I o U and A c c u r a c y values only being 75.85% and 80.41%. This is related to the fact that SimCLR relies more on large batch data, and the model cannot learn more useful knowledge from limited data. Although self-supervised contrastive learning is widely used in the field of image classification, research on segmentation tasks is still somewhat lacking. The other three common contrastive learning methods used for comparison only focus on global representations; thus, they ignore the local information, which is important for segmentation tasks. So, even after the fine-tuning period, their performance is still not excellent. However, our approach is designed to focus on both the global and local information of the image. Therefore, the proposed model can perform well whether the WBM defect patterns are easy or difficult to distinguish.

4.4. Discussion

4.4.1. Effectiveness of the Amount of Self-Supervised Unlabeled Data

Self-supervised pre-training needs to use a large amount of unlabeled data, and an abundance of unlabeled WBMs is readily accessible in the industrial manufacturing process. Therefore, this section investigates the potential of more self-supervised pre-training unlabeled data improving the performance of the model. Building upon our WBM dataset, we randomly utilized 1%, 25%, 50%, 75%, and 100% pre-training unlabeled WBMs. As presented in Table 6, the horizontal axis represents different amounts of pre-training unlabeled data, while the vertical axis displays various evaluation metrics for segmentation and classification. The experimental results indicate an overall upward trend as the amount of unlabeled data increases. This is because more unlabeled pre-training data contain richer features, so the model can learn more distinctive feature representations, thereby improving the generalization ability. Hence, it can be anticipated that the proposed method may yield even greater benefits when applied to a larger dataset for self-supervised pre-training.

4.4.2. Effectiveness of the Amount of Fine-Tuning Labeled Data

This section examines the influence of utilizing additional labeled data for fine-tuning on model performance. Initially, we expanded the fine-tuning labeled dataset to the same size as the pre-training unlabeled dataset. Then, we fine-tuned the model using varying proportions of labeled data, specifically 1%, 5%, 10%, 25%, 50%, and 100% of the self-supervised unlabeled data. The results, presented in Table 7, indicate that as the amount of labeled fine-tuning data increased, the evaluation metrics exhibited an upward trend, although at a diminishing rate. This can be attributed to the model’s ability to learn more information with an increasing number of labels but with a diminishing rate for each additional label. When the amount of labeled data reaches 25% of the self-supervised unlabeled data, the model achieves outstanding performance, and almost all of the evaluation metrics reach a fairly satisfactory result. Therefore, our proposed self-supervised contrastive learning model is not heavily dependent on the size of the labeled data but still remains effective even when a large amount of labeled data are available.

4.4.3. Ablation Study

In this section, we describe ablation experiments we conducted to investigate the effectiveness of the modules in our proposed method. Our proposed model mainly consists of three parts: global contrastive learning module, local contrastive learning module, and fine-tuning module. We chose to remove one or two modules separately for experimental research. The experimental results are shown in Table 8. We present the results of the ablative experiments conducted to assess the effectiveness of different modules. The horizontal axis represents the evaluation metrics, while the vertical axis lists various combinations of the modules. When there is only the global contrastive learning module or only the local contrastive learning module, the performance is slightly inferior to the experimental results obtained from the complete network. This indicates that both the global information and local information are beneficial for our segmentation and recognition task from different perspectives. Therefore, when removing both the global and local modules, the model performance becomes extremely poor. The purpose of the fine-tuning module is to make the self supervised pre-trained model better adapt to the specific downstream task. Hence, when the fine-tuning module is removed, the model does not adapt well to the recognition and segmentation of mixed-type WBM defect patterns. To sum up, each module is proven to be crucial for the performance of the model through the ablation experiments.

5. Conclusions

In this paper, we introduce a self-supervised contrastive learning model designed for the classification and segmentation of mixed-type WBM defect patterns. The experimental results indicate that our model surpasses other self-supervised contrastive learning approaches. Furthermore, we demonstrate that the performance of our proposed model does not heavily depend on labeled data, but more self-supervised unlabeled pre-training data can improve the model’s performance. And the ablation study further assesses the contribution of each model component to the overall performance. In practical scenarios where large amounts of unlabeled WBM data can be easily obtained, our proposed model will have significant real-world applications. However, in the process of chip manufacturing, even minor defects can lead to chip failure or performance degradation. Therefore, the requirements for defect detection are extremely stringent. We will focus on improving the detection performance with a limited number of labeled samples. In future research, we aim to integrate network architecture optimizations with self-supervised learning to increase model evaluation accuracy, reduce processing time, and lower computational resource demands. We also plan to extend the application of our model to other fields, such as medical image segmentation and remote sensing image analysis.

Author Contributions

All authors contributed to the study conception and design. Material preparation, data collection, and analysis were performed by S.Y., Y.Z. and R.W. The first draft of the manuscript was written by S.Y., and all authors commented on previous versions of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 72101065; in part by the Guangdong Basic and Applied Basic Research Foundation under Grants 2021A1515110336, 2020A1515110246; and in part by the Shenzhen Science and Technology Program under Grant RCBS20221008093124063.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this study are available from the corresponding author upon reasonable request. These data were used under license for the current study and cannot be redistributed without permission from the data provider.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Shin, E.; Yoo, C.D. Efficient Convolutional Neural Networks for Semiconductor Wafer Bin Map Classification. Sensors 2023, 23, 1926. [Google Scholar] [CrossRef] [PubMed]
  2. Chiu, M.-C.; Chen, T.-M. Applying Data Augmentation and Mask R-CNN-Based Instance Segmentation Method for Mixed-Type Wafer Maps Defect Patterns Classification. IEEE Trans. Semicond. Manuf. 2021, 34, 455–463. [Google Scholar] [CrossRef]
  3. Nakazawa, T.; Kulkarni, D.V. Wafer Map Defect Pattern Classification and Image Retrieval Using Convolutional Neural Network. IEEE Trans. Semicond. Manuf. 2018, 31, 309–314. [Google Scholar] [CrossRef]
  4. Lee, H.; Kim, H. Semi-Supervised Multi-Label Learning for Classification of Wafer Bin Maps With Mixed-Type Defect Patterns. IEEE Trans. Semicond. Manuf. 2020, 33, 653–662. [Google Scholar] [CrossRef]
  5. Wang, J.; Xu, C.; Yang, Z.; Zhang, J.; Li, X. Deformable Convolutional Networks for Efficient Mixed-Type Wafer Defect Pattern Recognition. IEEE Trans. Semicond. Manuf. 2020, 33, 587–596. [Google Scholar] [CrossRef]
  6. Liu, C.W.; Chien, C.F. An intelligent system for wafer bin map defect diagnosis: An empirical study for semiconductor manufacturing. Eng. Appl. Artif. Intell. 2013, 26, 1479–1486. [Google Scholar] [CrossRef]
  7. Qian, Y.; Tang, S.-K. Multi-Scale Contrastive Learning with Hierarchical Knowledge Synergy for Visible-Infrared Person Re-Identification. Sensors 2025, 25, 192. [Google Scholar] [CrossRef]
  8. Chen, X.; Fan, H.; Girshick, R.; He, K. Improved baselines with momentum contrastive learning. arXiv 2020, arXiv:2003.04297. [Google Scholar]
  9. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Los Angeles, CA, USA, 12–18 July 2020; pp. 1597–1607. [Google Scholar]
  10. Zhang, Y.; Zhang, Q. Semantic Segmentation of Traffic Scene Based on DeepLabv3+ and Attention Mechanism. In Proceedings of the International Conference on Neural Networks, Information and Communication Engineering, Guangzhou, China, 10–12 December 2023; pp. 542–547. [Google Scholar]
  11. Li, H.; Li, Y.; Zhang, G.; Liu, R.; Huang, H.; Zhu, Q.; Tao, C. Global and Local Contrastive Self-Supervised Learning for Semantic Segmentation of HR Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5618014. [Google Scholar] [CrossRef]
  12. Tong, L.I.; Wang, C.H.; Huang, C.L. Monitoring defects in IC fabrication using a Hotelling T/sup 2/ control chart. IEEE Trans. Semicond. Manuf. 2005, 18, 140–147. [Google Scholar] [CrossRef]
  13. Kim, B.; Jeong, Y.S.; Tong, S.H.; Chang, I.K.; Jeong, M.K. Step-Down Spatial Randomness Test for Detecting Abnormalities in DRAM Wafers with Multiple Spatial Maps. IEEE Trans. Semicond. Manuf. 2016, 29, 57–65. [Google Scholar] [CrossRef]
  14. Wang, C.H.; Kuo, W.; Bensmail, H. Detection and classification of defect patterns on semiconductor wafers. IIE Trans. 2006, 38, 1059–1068. [Google Scholar] [CrossRef]
  15. Yuan, T.; Kuo, W. Spatial defect pattern recognition on semiconductor wafers using model-based clustering and Bayesian inference. Eur. J. Oper. 2008, 190, 228–240. [Google Scholar] [CrossRef]
  16. Wang, R.; Che, N. Wafer Map Defect Pattern Recognition Using Rotation-Invariant Features. IEEE Trans. Semicond. Manuf. 2019, 32, 596–604. [Google Scholar] [CrossRef]
  17. Kang, H.; Kang, S. Semi-supervised rotation-invariant representation learning for wafer map pattern analysis. Eng. Appl. Artif. Intell. 2023, 120, 105864. [Google Scholar] [CrossRef]
  18. Byun, Y.; Baek, J.G. Mixed Pattern Recognition Methodology on Wafer Maps with Pre-trained Convolutional Neural Networks. In Proceedings of the International Conference on Agents and Artificial Intelligence, Lisbon, Portugal, 19–21 February 2020; pp. 53–62. [Google Scholar]
  19. Tello, G.; Al-Jarrah, O.Y.; Yoo, P.D.; Al-Hammadi, Y.; Muhaidat, S.; Lee, U. Deep-Structured Machine Learning Model for the Recognition of Mixed-Defect Patterns in Semiconductor Fabrication Processes. IEEE Trans. Semicond. Manuf. 2018, 31, 315–322. [Google Scholar] [CrossRef]
  20. Kyeong, K.; Kim, H. Classification of Mixed-Type Defect Patterns in Wafer Bin Maps Using Convolutional Neural Networks. IEEE Trans. Semicond. Manuf. 2018, 31, 395–402. [Google Scholar] [CrossRef]
  21. Nag, S.; Makwana, D.; Mittal, S.; Mohan, C.K. Wafersegclassnet—A light-weight network for classification and segmentation of semiconductor wafer defects. Comput. Ind. 2022, 142, 103720. [Google Scholar] [CrossRef]
  22. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Washington, DC, USA, 22–29 October 2017; pp. 33–42. [Google Scholar]
  23. Yan, J.; Sheng, Y.; Piao, M. Semantic Segmentation-Based Wafer Map Mixed-Type Defect Pattern Recognition. IEEE Trans. Comput.-Aided Des. Integr. Circuits Syst. 2023, 42, 4065–4074. [Google Scholar] [CrossRef]
  24. Wang, Y.; Liu, G. Self-Supervised Dam Deformation Anomaly Detection Based on Temporal–Spatial Contrast Learning. Sensors 2024, 24, 5858. [Google Scholar] [CrossRef] [PubMed]
  25. Kumari, P.; Kern, J.; Raedle, M. Self-Supervised and Zero-Shot Learning in Multi-Modal Raman Light Sheet Microscopy. Sensors 2024, 24, 8143. [Google Scholar] [CrossRef]
  26. Wu, Z.; Xiong, Y.; Yu, S.X.; Lin, D. Unsupervised feature learning via non-parametric instance discrimination. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3733–3742. [Google Scholar]
  27. Kwak, M.G.; Lee, Y.J.; Kim, S.B. SWaCo: Safe Wafer Bin Map Classification With Self-Supervised Contrastive Learning. IEEE Trans. Semicond. Manuf. 2023, 6, 416–424. [Google Scholar] [CrossRef]
  28. Kahng, H.; Kim, S.B. Self-supervised representation learning for wafer bin map defect pattern classification. IEEE Trans. Semicond. Manuf. 2020, 34, 74–86. [Google Scholar] [CrossRef]
  29. Hu, H.; He, C.; Li, P. Semi-supervised Wafer Map Pattern Recognition using Domain-Specific Data Augmentation and Contrastive Learning. In Proceedings of the 2021 IEEE International Test Conference, Anaheim, CA, USA, 1–4 November 2021; pp. 113–122. [Google Scholar]
  30. Oord, A.V.D.; Li, Y.; Vinyals, O. Representation learning with contrastive predictive coding. arXiv 2018, arXiv:1807.03748. [Google Scholar]
  31. Jadon, S. A survey of loss functions for semantic segmentation. In Proceedings of the 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology, Via del Mar, Chile, 3–5 August 2020; pp. 1–7. [Google Scholar]
  32. Mann, W.R.; Taber, F.L.; Seitzer, P.W.; Broz, J.J. The leading edge of production wafer probe test technology. In Proceedings of the 2004 International Conferce on Test, Charlotte, NC, USA, 25–29 April 2004; pp. 1168–1195. [Google Scholar]
  33. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  34. Noroozi, M.; Favaro, P. Unsupervised learning of visual representations by solving jigsaw puzzles. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 69–84. [Google Scholar]
Figure 1. (a) Eight common single-type WBM defect patterns. (b) Examples of mixed-type WBM defect patterns.
Figure 1. (a) Eight common single-type WBM defect patterns. (b) Examples of mixed-type WBM defect patterns.
Sensors 25 01272 g001
Figure 2. The self-supervised learning paradigm.
Figure 2. The self-supervised learning paradigm.
Sensors 25 01272 g002
Figure 3. The overall framework of our proposed model. Note that the figure omits the display of negative samples. In the self-supervised pre-training stage, we pre-train the network using a large amount of unlabeled data x i D u . After that, we transfer the knowledge embedded in the query encoder and query decoder to the downstream task and fine-tune the network with few labeled samples { x i , y i } D l .
Figure 3. The overall framework of our proposed model. Note that the figure omits the display of negative samples. In the self-supervised pre-training stage, we pre-train the network using a large amount of unlabeled data x i D u . After that, we transfer the knowledge embedded in the query encoder and query decoder to the downstream task and fine-tune the network with few labeled samples { x i , y i } D l .
Sensors 25 01272 g003
Figure 4. Illustration of the data augmentation operators.
Figure 4. Illustration of the data augmentation operators.
Sensors 25 01272 g004
Figure 5. Process of local region selection and matching.
Figure 5. Process of local region selection and matching.
Sensors 25 01272 g005
Figure 6. Wafer probe testing station.
Figure 6. Wafer probe testing station.
Sensors 25 01272 g006
Figure 7. Examples of visualization results obtained using our proposed model.
Figure 7. Examples of visualization results obtained using our proposed model.
Sensors 25 01272 g007
Table 1. Description of mixed-type WBM defect patterns.
Table 1. Description of mixed-type WBM defect patterns.
Single-TypeC (C1)D (C2)EL (C3)ER (C4)
L (C5)S (C6)NF (C7)R (C8)
Two-Mixed-TypeC + EL (C9)C + ER (C10)C + L (C11)C + S (C12)
D + EL (C13)D + ER (C14)D + L (C15)D + S (C16)
EL + L (C17)EL + S (C18)ER + L (C19)ER + S (C20)
L + S (C21)---
Three-Mixed-TypeC + EL + L (C22)C + EL + S (C23)C + ER + L (C24)C + ER + S (C25)
D + EL + L (C26)D + EL + S (C27)D + ER + L (C28)D + ER + S (C29)
C + L + S (C30)D + L + S (C31)EL + L + S (C32)ER + L + S (C33)
Four-Mixed-TypeC + L + EL + S (C34)C + L + ER + S (C35)D + L + EL + S (C36)D + L + ER + S (C37)
Table 2. Segmentation and classification results for normal and single-type defect patterns (%).
Table 2. Segmentation and classification results for normal and single-type defect patterns (%).
CategoryC0C1C2C3C4C5C6C7C8
SegmentationPacc-98.6996.4095.9598.0296.9091.4998.7490.11
IoU-96.2194.1993.2395.8896.0589.3597.5588.46
ClassificationAccuracy98.7598.6795.8396.8998.4197.0494.5498.7292.32
Precision96.6697.2493.4691.8998.7497.3092.0795.4188.74
Recall93.1992.9595.2096.1197.6894.6996.3892.8895.01
Table 3. Segmentation results for mixed-type defect patterns (%).
Table 3. Segmentation results for mixed-type defect patterns (%).
CategoryCDELERLS
PaccIoUPaccIoUPaccIoUPaccIoUPaccIoUPaccIoU
C998.5796.12--95.9493.23------
C1098.6696.19----97.9495.48----
C1198.4996.10------96.8896.01--
C1298.4896.10--------91.4889.35
C13--96.2393.8995.6693.10------
C14--96.3194.08--97.9195.47----
C15--96.2093.88----96.7995.12--
C16--96.2293.85------91.4689.34
C17----95.2192.16--96.2095.03--
C18----95.2992.17----91.3489.19
C19------97.8295.4096.7795.13--
C20------97.7795.35--91.2289.15
C21--------96.1894.9990.7988.94
C2298.4396.02--95.3292.20--96.1494.90--
C2398.4296.02--95.3292.21----90.1988.03
C2498.4596.04----97.5295.0696.0294.78--
C2598.4596.03----97.5595.07--90.1287.94
C26--95.1891.9893.4091.03--94.9992.11--
C27--95.1591.9093.3991.03----90.1089.90
C28--95.2091.92--96.1894.8995.0392.12--
C29--95.1891.97--96.1594.87--90.1087.89
C3098.3995.95------94.9592.0390.0287.86
C31--95.1191.86----94.9292.0290.0187.85
C32----93.0890.02--94.8691.9589.9087.80
C33------96.0594.8094.8791.9589.9087.81
C3497.9895.07--92.1189.20--93.8790.2688.4085.98
C3598.0095.08----95.1193.5793.8990.2788.4185.98
C36--93.8590.0792.0388.79--93.0889.8888.1685.21
C37--93.9090.10--94.7993.1493.1189.9588.1885.25
Table 4. Classification results for mixed-type defect patterns (%).
Table 4. Classification results for mixed-type defect patterns (%).
CategoryAccuracyPrecisionRecall
C995.6298.6793.30
C1096.1997.9894.02
C1194.4597.5693.88
C1293.8792.4398.14
C1391.9293.0097.44
C1493.8594.8695.12
C1591.6896.3392.88
C1691.6191.4498.05
C1794.3197.8793.30
C1892.4191.5597.20
C1992.8492.3494.99
C2092.6990.5996.22
C2191.8896.8191.74
C2292.4589.8493.10
C2389.9795.4492.13
C2491.9398.0190.74
C2590.7693.7189.67
C2690.5591.3597.28
C2789.6891.5596.98
C2892.6190.2498.00
C2989.7799.0192.74
C3090.1898.1294.17
C3189.1194.3896.05
C3290.5697.3593.69
C3390.3492.4496.77
C3488.7688.2995.13
C3589.5291.3093.27
C3687.4196.6689.35
C3788.4987.8094.06
Table 5. Comparison with other methods using limited labeled data on WBM defect pattern recognition and segmentation task (%).
Table 5. Comparison with other methods using limited labeled data on WBM defect pattern recognition and segmentation task (%).
MethodsSegmentation (Overall)Recognition (Overall)
mPaccmIoUAccuracyPrecisionRecall
Jigsaw83.6676.1980.9478.4481.71
SimCLR83.4375.8580.4175.9678.13
MoCo v284.1178.5282.4680.0883.25
Ours95.5992.1993.5794.4394.78
Table 6. Experimental results with different amounts of pre-training unlabeled data (%).
Table 6. Experimental results with different amounts of pre-training unlabeled data (%).
Different Amounts of Pre-Training
Unlabeled Data
1%25%50%75%100%
(152)(3800)(7600)(11,400)(15,200)
Segmentation results (Overall)mPacc31.1240.5572.6489.3395.59
mIoU29.3340.1763.8182.3092.19
Classification results (Overall)Accuracy27.9438.5570.6686.4593.57
Precision28.6041.0770.2385.0494.43
Recall30.4145.0271.3888.1094.78
Table 7. Experimental results with different amounts of fine-tuning labeled data (%).
Table 7. Experimental results with different amounts of fine-tuning labeled data (%).
Different Amounts of Fine-Tuning
Labeled Data
1%5%10%25%50%100%
(152)(760)(1520)(3800)(7600)(15,200)
Segmentation results (Overall)mPacc56.3270.9288.2595.5495.7295.80
mIoU48.6763.1181.5592.1092.1992.23
Classification results (Overall)Accuracy57.0871.1087.5593.4993.6093.64
Precision58.6875.4389.0294.3894.9295.17
Recall52.8774.1685.3994.7795.0695.33
Table 8. Results of ablation experiments on exploring the effectiveness of each module (%).
Table 8. Results of ablation experiments on exploring the effectiveness of each module (%).
ModulesSegmentation (Overall)Classification (Overall)
GlobalLocalFine-TuningmPaccmIoUAccuracyPrecisionRecall
95.5992.1993.5794.4394.78
75.0971.5472.6475.1671.81
84.1178.5282.4680.0883.25
86.4480.2284.1080.5985.67
52.7645.2850.4949.7755.01
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yin, S.; Zhang, Y.; Wang, R. Contrastive Learning with Global and Local Representation for Mixed-Type Wafer Defect Recognition. Sensors 2025, 25, 1272. https://doi.org/10.3390/s25041272

AMA Style

Yin S, Zhang Y, Wang R. Contrastive Learning with Global and Local Representation for Mixed-Type Wafer Defect Recognition. Sensors. 2025; 25(4):1272. https://doi.org/10.3390/s25041272

Chicago/Turabian Style

Yin, Shantong, Yangkun Zhang, and Rui Wang. 2025. "Contrastive Learning with Global and Local Representation for Mixed-Type Wafer Defect Recognition" Sensors 25, no. 4: 1272. https://doi.org/10.3390/s25041272

APA Style

Yin, S., Zhang, Y., & Wang, R. (2025). Contrastive Learning with Global and Local Representation for Mixed-Type Wafer Defect Recognition. Sensors, 25(4), 1272. https://doi.org/10.3390/s25041272

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop