Next Article in Journal
High-Voltage Power Supply for Four-Quadrant Dielectric Elastomer Actuators
Previous Article in Journal
Adaptation of a Differential Scanning Calorimeter for Simultaneous Electromagnetic Measurements
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Eddy Current Array for Defect Detection in Finely Grooved Structure Using MSTSA Network

School of Automation Engineering, University of Electronic and Scientific Technology of China, 2006 Xiyuan Ave., Gaoxin West District, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(18), 6078; https://doi.org/10.3390/s24186078
Submission received: 29 July 2024 / Revised: 8 September 2024 / Accepted: 18 September 2024 / Published: 20 September 2024
(This article belongs to the Section Sensor Networks)

Abstract

:
In this paper, we focus on eddy current array (ECA) technology for defect detection in finely grooved structures of spinning cylinders, which are significantly affected by surface texture interference, lift-off distance, and mechanical dither. Unlike a single eddy current coil, an ECA, which arranges multiple eddy current coils in a specific configuration, offers not only higher accuracy and efficiency for defect detection but also the inherent properties of space and time for signal acquisition. To efficiently detect defects in finely grooved structures, we introduce a spatiotemporal self-attention mechanism to ECA testing, enabling the detection of defects of various sizes. We propose a Multi-scale SpatioTemporal Self-Attention Network for defect detection, called MSTSA-Net. In our framework, Temporal Attention (TA) and Spatial Attention (SA) blocks are incorporated to capture the spatiotemporal features of defects. Depth-wise and point-wise convolutions are utilized to compute channel weights and spatial weights for self-attention, respectively. Multi-scale features of space and time are extracted separately in a pyramid manner and then fused to regress the bounding boxes and confidence levels of defects. Experimental results show that the proposed method significantly outperforms not only traditional image processing methods but also state-of-the-art models, such as YOLOv3-SPP and Faster R-CNN, with fewer parameters and lower FLOPs in terms of Recall and F1 score.

1. Introduction

Eddy current testing (ECT) has been a widely studied technique in the field of non-destructive testing, drawing significant attention from researchers [1,2]. Based on the principle of electromagnetic induction, ECT is used to detect surface or subsurface defects in conductive materials [3,4]. The core mechanism involves measuring changes in the impedance of a detection coil (probe), which occur due to variations in the eddy currents induced within the conductor [5]. Recent advancements in ECT technology have focused on visualization, aiming to enhance the precision of defect localization, as well as the characterization of defect shape and size. Since the original signal obtained from ECT is in the form of impedance, defect detection can be achieved by converting the impedance data into pseudo-images, which can then be analyzed using digital image processing techniques. As probes become smaller, eddy current arrays (ECA) arrange multiple eddy current coils in a specific configuration to form a planar structure for defect detection [6,7,8]. ECA is excited at high frequency in different directions to capture signals for identical defects near the surface, possessing inherent spatial and temporal properties. Compared to a single eddy current coil, ECA offers a larger measurement range, higher resolution, and greater accuracy [9], significantly enhancing testing speed, measurement accuracy, and sensor system reliability. Additionally, compared to ordinary arrays, ECA has good flexibility and can detect parts with complex surface shapes. Theoretically, if the captured signal is clean and free from interference, detecting the extreme value of the signal can effectively indicate the presence of defects.
In this paper, we focus on defect detection in finely grooved structures of spinning cylinders. Fine grooves are common textures on metal surfaces. The spinning cylinder with fine grooves is fixed on a robotic arm, and signals are captured through ECA. Defect detection is challenged by factors such as the surface texture of the spinning cylinder, lift-off distance, environmental magnetic fields, mechanical dither, and more. Consequently, traditional digital image processing methods result in high false positive and false negative rates. Therefore, a more powerful detection algorithm is needed to accurately detect defects in finely grooved structures. Instead of exploring what ECA can detect from excitation frequency or lift-off distance, we aim to develop an appropriate detection algorithm to achieve higher accuracy and recall rates.
ECA coils are stimulated at high frequencies in different directions to capture signals, which naturally possess spatiotemporal properties. Additionally, the spatiotemporal self-attention mechanism can focus on critical parts of the information from temporal and spatial dimensions, extracting features to improve detection accuracy. Therefore, we introduce the spatiotemporal self-attention mechanism into ECA defect detection in finely grooved structures for the first time and propose the Multi-scale SpatioTemporal Self-Attention Network (MSTSA-Net). The main contributions of this paper are:
(1) The principle of ECA for defect detection is analyzed in detail first, clearly showing that the signal collected by ECA has the natural properties of space and time, which leads to the proposed MSTSA-Net for defect detection.
(2) The MSTSA-Net takes multi-channel data from ECA as input, integrating and fusing spatial and temporal features. Multi-scale features from different channels are extracted separately by the Mixed Residual Attention Module (MRAM) in a pyramid manner and fused in the Up Sampling Block (UpS Block). Depth-wise and point-wise convolution strategies, instead of pooling, are utilized to compute the spatial and temporal weights in the Temporal Attention (TA) and Spatial Attention (SA) blocks.
(3) The proposed approach outperforms the state-of-the-art methods on the testing dataset in terms of Recall and F1 in experimental results.
A preliminary version of this work was published in [10]. This paper significantly improves upon [10] in the following aspects:
(1) The implementation strategies are different. The former work treats defect recognition as a classification problem. After extracting the spatiotemporal features, this paper treats defect detection as a regression problem to directly obtain bounding boxes for objects.
(2) The acquisition platform is described in Section 4, and an improved neural architecture is proposed for defect detection in Section 5.
(3) Compared with state-of-the-art detection methods, the performance of the proposed network outperforms YOLOv3-SPP and Faster-RCNN with fewer network parameters.
This paper is divided into six parts. Section 3 describes the principle of ECA for defect detection. Section 4 introduces the acquisition platform and analyzes the interference factors in our case. The MSTSA network is introduced in Section 5.2. The proposed method is evaluated qualitatively and quantitatively in Section 6, and the conclusions are presented in Section 7.

2. Related Work

2.1. Defect Detection by ECT

Traditional ECT methods have primarily concentrated on optimizing sensor design and excitation parameters [11,12,13] to enhance detection sensitivity. These optimization efforts typically focus on aspects such as the geometry of excitation coils, the use of pulsed or encoded excitation sources, and differential probes, among others [14,15,16,17]. As ECT is visualized in the form of images, multiple algorithms have been proposed to address defect detection. Liu et al. developed an image fusion algorithm for defect recognition and extraction based on pulsed eddy current thermal image analysis [18]. Mook et al. successfully visualized defect detection by employing eddy current micro-array probes [19]. Mekhalfa and Nacereddine [20] utilized support vector machines (SVM) to classify weld defects, while Gao et al. identified defect locations through the analysis of eddy current thermal images [21]. Long et al. proposed a resolution-enhanced ECA probe with four rows of coils to obtain finer spatial resolution without sacrificing sensitivity in [22]. Lysenko et al. used ECA to detect and evaluate defects in Aviation Components in [23]. Caetano et al. detected hole-like defects in materials using ECT Technique by neural network in [24].

2.2. Object Recognition Based on DEEP Neural Network

Deep artificial neural networks have been successfully used in image processing [25,26] in recent years. AlexNet [25] marked the beginning of the convolutional neural network (CNN) era. Numerous backbones and network architectures have since been proposed, including VGG [27], ResNet [26], DenseNet [28], Encoder-Decoder [29], GAN [30,31], and Transformer [32]. Object detection is a particularly active area of research. Girshick et al. proposed a two-stage object detection method called R-CNN, which first extracted CNN features from region proposals and solved a classification problem for object detection [33]. This early region-based method evolved into versions like Fast R-CNN [34], Faster R-CNN [35], and Mask R-CNN [36]. Fast R-CNN improved the execution efficiency of R-CNN by framing it as a regression problem [34]. Ren et al. introduced a Region Proposal Network (RPN) to extract region proposals and incorporated attention mechanisms into RPN and Fast R-CNN, leading to Faster R-CNN [35]. He et al. simultaneously generated high-quality segmentation masks during detection [36]. The YOLO series treated object detection as a regression problem [37,38,39,40], aiming for real-time detection at the cost of reasonable accuracy. Researchers have continuously improved YOLO methods from v1 to v7, making them state-of-the-art in object detection. Fang et al. introduced transformers to object detection in video sequences [39]. Liu et al. proposed a Single Shot Multi-box Detector (SSD), which fused multi-scale features to enhance detection accuracy (mAP) [41]. He et al. addressed the problem of inconsistent object expression by fixing image size through cropping and warping, then using spatial pyramid pooling for feature extraction [42].

2.3. Attention Mechanisms

Attention mechanisms are a popular approach in deep learning, focusing on important parts of objects [32]. They have been successfully applied to various tasks [43,44,45]. Attention mechanisms highlight where learning should focus and enhance feature representation. Wang et al. proposed a self-attention model for saliency prediction based on spatiotemporal features of videos [43]. Woo et al. proposed a convolutional attention module to extract channel and spatial features [46]. Researchers have also explored using CNNs for defect detection with promising results [47,48]. Yang et al. used the SSD framework to detect magnetic flux leakage signals, identifying locations of circumferential and spiral welds, as well as small defects [48]. Tao et al. designed ECA sensors for differential multi-mode array weld inspection and used Mask R-CNN for defect detection [47].

3. Principle Analysis of ECA for Defect Detection

According to the principle of eddy current testing, when the excitation coil is fed with alternating current, an induced magnetic field is generated, producing an eddy current in the tested conductor. When a defect falls in the area between the excitation coil and the detection coil, it alters the flow direction of the eddy current, changing the induced magnetic field generated by the eddy current. This change in the magnetic field can be detected by the detection coil. Therefore, there exists a sensitive area for detection between the excitation coil and the detection coil [49], as shown in Figure 1. When a strip-shaped defect is in the sensitive area and perpendicular to the direction of the eddy current, the eddy current will form the maximum accumulation at the two ends of the defect.
To detect defects in different directions, a time-sharing multi-excitation transmitting and receiving eddy current array scanning method is used, as shown in Figure 2, which can be realized using a multiplexer [9]. In this method, only one coil unit of the eddy current array sensor is used for excitation at a time to reduce the influence of mutual inductance interference between coil units on detection. In Figure 2, T represents the excitation coils, and R 1 and R 2 are the detection coils. There are two excitation modes: in the first mode, the sensitive area is roughly parallel to the scanning direction of the sensor; that is, the coil T sends a signal, and the coil R 1 receives it, referred to as Channel A (CA) in this paper. In the second mode, the sensitive area is perpendicular to the scanning direction of the sensor; that is, the coil T sends a signal, and the coil R 2 receives it, referred to as Channel T (CT). Because the two rows of coils are excited sequentially, this causes two signal abnormal areas in CT. With these two modes, different defects can be successfully detected.
As shown in Figure 2, the schematic diagram illustrates an eddy current array (ECA) consisting of two rows of coils. The excitation signal generated by coil T is received by coils R 1 and R 2 at different moments, respectively. As the device moves, coils from different rows are sequentially activated to sense the same region (defect). Figure 3 presents a real data visualization from the ECA. Specifically, Figure 3a shows the CT channel, while Figure 3b shows the CA channel. It can be observed that the CA measurement mode is more sensitive to longitudinal defects or the upper and lower edges of defects, whereas the CT mode is more sensitive to lateral defects or the left and right edges of defects.
There exists a mathematical relationship between the schematic diagram and real data based on our observations, as follows:
d d 2 = K
In this equation, d is the distance between the centers of signal change in the image coordinate system corresponding to the two lines of coils in Figure 2, d 2 is the physical distance between the two lines of coils, and K is a constant that represents the ratio between the physical distance and the distance in the image coordinate system. This mathematical relationship can be useful for estimating the real size of defects.
Using ECA sensors to detect defects on ferromagnetic objects is an extremely complex physical process. It is not only affected by the eddy current-induced magnetic field but also affected by magnetic flux leakage. We use the sensor shown in Figure 2 to collect signals on the transverse and longitudinal artificial defect features for analysis in Figure 4. The artificial defects from left to right are three transverse defects and three longitudinal defects with the same shape but different depths, respectively. For the three longitudinal defects on the right, the eddy current generated reaches the maximum since the sensitive area is just perpendicular to the defect. For the three transverse defects on the left, the shallow defect makes the magnetic flux leakage concentrated, so it shows dark, bright and dark changes at the shallow defect boundary. Since the second and third defects on the left are deeper, the magnetic flux leakage will be strengthened, so bright, dark, and bright are formed at both ends of the defect.

4. Acquisition Platform and Interference Analysis

In this section, we first introduce the data acquisition platform and then analyze its impact on ECA defect detection, providing insights for the detection algorithms. Our probe is flexible and consists of 48 coils, each with a diameter of 2 mm, arranged in 3 rows with 16 coils in each row. The object to be detected is a spinning cylinder with fine grooves, as shown in Figure 5a. The cylinder is clamped onto a robotic arm while the ECA sensor remains fixed. The robotic arm rotates the spinning cylinder at a constant, fixed speed, allowing simultaneous scanning of the outer and inner surfaces of the cylinder. Figure 5b illustrates the schematic diagram of the acquisition platform.
Signal interference mainly comes from these aspects [50] in our case:
(1) The mechanical spindle’s rotation error causes the lift-off distance change. As we all know, ECT is greatly affected by lifting distance. During the rotation of the scanning spindle, the deflection accuracy error will cause a large rotation error at the end of the object, which causes the variations of lift-off distance, see Figure 6.
(2) Mechanical jitter interference. During the detection process of the object, there may be local protrusion errors and mechanical jitter interference on the surface of the object due to the limitation of machining accuracy.
These two main reasons cause defect detection for ECA to be hard in our case. You can see some examples from the dataset in Section 6. It is impossible to distinguish the defects from the strength and shape of the signal.

5. Defect Detection Using Multi-Scale Spatiotemporal Self-Attention Network

After the ECA data are collected, we first preprocess the ECA data into image coordinates for computing and visualization and introduce the framework of the MSTSA network. The architecture details and implementation details of the network are given then. Unlike the work in [10], the output of the proposed method is to bind the defects.

5.1. Pseudo-Color Image Generation

It is necessary to convert this impedance signal matrix into pseudo-image data by rescaling the values to the range [ 0 , 255 ] since the original signal collected by the ECA consists of impedance values, typically assumed to fall within the range [ M , M ] . This preprocessing step is primarily performed to facilitate visualization, which is crucial for observing and verifying intermediate results. We apply a piecewise linear function for rescaling based on statistical analysis. It has been observed that most abnormal signal values associated with defects fall within the range [ S , S ] , while only a few defect signals exhibit very high amplitudes, exceeding S or even higher. The following equation is used to rescale the original data:
f = S , x < S , x , x [ S , S ] , S , x > S . f = f + S 2 S × 256 ,
in which f is the raw data, f is the value in image coordinate which is in the range of [ 0 , 255 ] .

5.2. Multi-Scale Spatiotemporal Self-Attention Network for Defect Detection

The network we propose is an end-to-end detection network, which mainly consists of four MRAMs, two UpS blocks, and three Output blocks. The network architecture is shown in Figure 7. The collected data from different channels are preprocessed by the method mentioned above, and then sent to the network as input. CA+CT means the image addition of CA and CT. For more details, please see the experiments section. The data from the three channels first passes through a convolution layer with stride 2, Batch Normalization layer, and ReLU layer, and then is sent to MRAM, which is composed of three different DRAM and RAM to extract features of different scales. Our network stacks one DRAM and two RAM four times in a pyramid manner. DRAM and RAM consist of a sequentially temporal self-attention block (TA block) and a spatial self-attention block (SA block), respectively. In TA and SA, the features from the different channels and spaces are emphasized through kernel convolution by computing a weight vector. And then features at adjacent layers are fused through the Up Sampling block, which includes a deconvolution layer, then connected to the Output block, which regresses to multiple six-dimensional vectors [ S w , S h , δ ( x ) , δ ( y ) , S c , L ] . [ S w , S h ] is the parameters corresponding to the size of the object, [ δ ( x ) , δ ( y ) ] is the offset relative to the upper left corner of the grid in normalized coordinates, S c means the score of confidence, and L is the category label. For more details, refer to Figure 7. Conv2D(3 × 3, i) represents a 2D 3 × 3 convolution with a stride of i. When the stride is 2, it leads to a result of lower resolution.
From a different perspective, four MRAMs are responsible for extracting features of different scales, the two Up Sampling blocks are responsible for fusing features of two different scales, and the three-layer Output blocks are responsible for outputting a 6-dimensional detection vector. The 6-dimensional detection vector of each layer corresponds to the expression of the same target at different resolutions. Small targets will be detected in the high-resolution layer, while big targets will be detected in the low-resolution layer. The number of six-dimensional detection vectors is determined by the number of anchor frames in the feature extraction layer. Assume our input image is with the size of ( W , L ) , then the number of anchor frames in our model is computed as follows,
N a = N a ( 2 ) + N a ( 3 ) + N a ( 4 ) = 1 16 ( W 1 2 2 · L 1 2 2 + W 1 2 3 · L 1 2 3 + W 1 2 4 · L 1 2 4 ) = 21 16 · W L
N a ( k ) represents the number of anchor frames in the kth layer, which corresponds to the size of object candidates and is set in advance. We only use feature maps from the last three MRAMs to regress the output vector due to the consideration of time efficiency, so k = 2 , 3 , , 4 .
The bounding boxes in the original input can be computed from the three different Output blocks at different scales. The final bounding box B is to merge the detection box by the following equation,
B = [ m i n { B x ( k ) } , m i n { B y ( k ) } , m a x { B x ( k ) } , m a x { B y ( k ) } ] , k = 2 , 3 , 4
B x ( k ) and B y ( k ) are x and y coordinate of the bounding box detected from the kth MRAM, respectively.

5.3. Mixed Residual Attention Block

As shown in Figure 7, MRAM is a mixture of three different DRAMs and RAMs. Three channels from CA, CT, and CA+CT are concatenated before Conv2D, BN, and ReLU, resulting in C channel data. Then, the C channel data are split into three branches, corresponding to CA, CT, and CA+CT, and connected to DRAM and RAM, respectively. DRAB and RAB are feature extraction blocks built on the residual attention block architecture. DRAM and RAM are quite similar; the primary difference is that the output size of DRAM is half of the input size. Downsampling in DRAM is achieved by C o n v 2 D ( , 2 ) , which means the stride is set to 2. Both DRAM and RAM are composed of sequential TA and SA blocks. The main idea of the TA and SA blocks is to learn temporal and spatial weights, respectively. These weights are then used to enhance attention in the feature maps.
We are inspired by the idea of [46] to create TA and SA blocks, but we use different strategies instead of pooling. In the TA block, a large depth-wise convolution kernel is used to compress the input of size H 2 × W 2 into a 1 × 1 × c tensor. This is followed by a Squeeze-and-Excitation (SE) structure [51], which computes multi-channel weights. These weights are then passed through a sigmoid function to generate channel-specific weights. The feature maps are linearly weighted by these channel weights to produce the input for the spatial self-attention (SA) module. In the SA block, the output from the TA block is convolved with a 1 × 1 × C kernel (a point-wise convolution) to compute spatial weights. These spatial weights are multiplied with the input feature map, and after applying a sigmoid function, the result is added back to the original input to form the input for the next MRAM module. The output from the previous MRAM is used as the input for the next MRAM. The resolution of the feature map will be reduced to 1 / 4 of the input after each MRAM because DRAM uses a convolution operation with a stride of 2. The deeper the MRAM, the lower the resolution of the feature map and the higher the number of channels. Therefore, the entire network has a pyramid structure. The resolution of the output from the kth MRAM is represented as follows:
R s o l ( k ) = 1 4 ( W 2 k × L 2 k )
Stacking the DRAMs and RAMs means stacking the TA and SA blocks multiple times. The original three-channel data are also fused and split in the DRAMs and RAMs to extract features at different scales. The details are shown in Figure 8.

5.4. Feature Fusion and Regression Output

After four MRABs, the feature map of the ( k + 1 ) th layer of MRAB is fused with the feature map of the kth layer in the Up Sampling block due to the different resolutions between the ( k + 1 ) th and the kth layers. The Up Sampling block upscales the ( k + 1 ) th layer to the same size as the kth layer by deconvolution. The output of the Up Sampling block is concatenated with the kth layer feature map, which becomes the input for the output block. The output block, including a convolutional layer of (33, 1), a ReLU layer, and a convolutional layer of (11, 1), regresses the multi-scale feature map into a six-dimensional vector as described in the previous Section 5.2 for each layer. The final estimated bounding boxes are combined into the original image scale.
Our spatiotemporal self-attention network is trained in a supervised manner. We manually label defects in the dataset (see the experiments section). The loss function L used in our framework is optimized to train the proposed network. The loss function consists of three terms: position loss L p , confidence loss L c , and class loss L c l s , which are commonly used in YOLO [37].
L = λ 1 · L p + λ 2 · L c + λ 3 · L c l s
where λ 1 , λ 2 , λ 3 are three balance factors, which are hyper-parameters set in advance.
The position loss L p constrains the estimation of the bounding box to be as close as possible to the ground truth of the bounding box. Assume the bounding box estimated by the network is denoted by S e , and the ground truth is denoted by S g t ,
L p = i ( 1 I o U i ) N
where I o U = S e S g t S e S g t , so I o U [ 0 , 1 ] . A larger I o U means the estimation is closer to the ground truth, resulting in a lower loss and vice versa. N is the number of objects.
The confidence loss is defined as follows:
L c = i ( o i · log c ^ i + ( 1 o i ) · log ( 1 c ^ i ) ) N a c i ^ = 1 1 + e S i c
where o i is an indicator for the ith patch; if it is an object, then o i is 1, otherwise it is 0. It can be generated from the ground truth. c i ^ is the sigmoid value of the confidence S i c of the ith patch.
Assume we have multiple classes, o i j is the indicator of the ith object in the jth class, c ^ i j is the sigmoid value of the confidence c ^ i j of the ith object in the jth class, and C j is the set of the jth class. Considering the number of samples in each class, the class loss is defined as follows:
L c l s = i j C j ( o i j log c ^ i j + ( 1 o i j ) log ( 1 c ^ i j ) ) N C j N
Among these three loss terms, the confidence number is the most important in our case, indicating whether the patch is a defect.

5.5. Implementation Details

The input and output sizes of all modules are provided to help readers understand the network implementation. The implementation details are summarized in Figure 9. The three-channel data are first convolved with a 3 × 3 kernel, then connected to a Batch Normalization layer and a ReLU layer. To control the number of anchor frames for detection (the total number of anchor frames is computed in Equation (3)), we set the stride to 2 for convolution to reduce the resolution of the next feature map but increase the number of channels to C groups. Each group has three 3 × 3 convolutions for CA, CT, and CA+CT, forming an output of 32 × 256 × C . After each MRAM, the resolution of the feature map is reduced to a quarter of the original, and the number of channels is doubled. After each UpS Block, the resolution of the feature map is increased by 2 times, the number of channels is halved, and finally, the Out Blocks regress to a vector of 6. The parameters to be learned in the network are mainly all convolution kernel parameters and the number of deconvolution kernels. There are a total of 36.4 M parameters in our model, compared with 62.6 M in YOLOv3-SPP and 41.4 M in Faster-RCNN. N a is the total number of object candidates estimated in our framework. Since N a is less than 700 in our approach, our method is faster than YOLOv3-SPP and Faster-RCNN in detection.

6. Experimental Results

In this section, we first introduce the datasets used in our experiments and provide the details of our experimental setup and evaluation metrics. We then evaluate our approach qualitatively and quantitatively and perform some ablation experiments as well.

6.1. Datasets

The raw data collected in the experiment comes from two channels using the acquisition platform introduced in Figure 2, called Channel A (CA) and Channel T (CT). According to the aforementioned detection physical mechanism, these two types of data correspond to different features of the same detected location by different coils and carry different information about the same defects. It is observed that the CA measurement mode is more sensitive to longitudinal defects or the upper and lower edges of defects, while CT is more sensitive to lateral defects or the left and right edges of defects. We manually align CA and CT and fuse them to form a third channel by averaging, named CA+CT. For CA+CT, it is the result of adding the absolute values of the original data from CA and CT and then converting them to image coordinates as described in Section 5.1.
After the original signal is converted into pseudo-color images, there are a total of 92 original raw images with identical widths but different lengths. We split the images into a training dataset and a testing dataset. The training dataset consists of 74 original images. These original images are cropped and padded into patches of 64 × 512 pixels, with a 100-pixel overlap for adjacent patches to ensure that the defects are not incomplete. Some training sample images from the dataset are shown in Figure 10. The details of the training and testing datasets are summarized in Table 1. There are 776 patches for training, of which 386 are defective. The total number of defect objects labeled manually is 495, including 248 large defects and 247 small defects (less than 900 pixel2). The testing dataset consists of 18 original images. After cropping and padding, there are 114 patches, of which 77 are defective. The total number of defects is 138, including 66 large defects and 72 small defects from man-made and natural defects.
The actual defect images are shown in Figure 11, where Figure 11a shows the man-made defects, and Figure 11b shows the natural defects. The targets in the red boxes are manually labeled defects. The targets in the green boxes are signal extrema but not defects caused by mechanical jitter and texture. In our experiments, the signal abnormalities caused by mechanical jitter and texture are considered an independent class in our dataset. It can be seen that the lift-off distance and signal anomalies caused by mechanical jitter and texture greatly increase the difficulty of identification.

6.2. Experimental Setup and Evaluation Indexes

6.2.1. Experimental Setup

Our experiments are conducted using an NVIDIA 3090 GPU made by Nvidia company in Santa Clara, CA, US. We set the same balance factors ( λ 1 = 1 , λ 2 = 64 , λ 3 = 1 ) and the same IoU threshold (0.05) for the proposed method and YOLOv3-SPP since they have the same objective. In all experiments, the batch size is set to 32, and the number of epochs is set to 400. The training process of the proposed network takes about 150 min. The learning rate we use is a variable learning rate l r , which is initially set to 0.01 and adjusted dynamically by the following equation:
γ = 1 + cos ( N e 400 · π ) 2 × ( 1 r ) + r
where γ is a decay factor of learning rate, N e is the current epoch index, and r is a constant set to 0.01. The learning rate after each epoch of training is then l r ( N e ) = l r 0 × γ .
We statistically analyze the size of the defect objects in the training set and set the potential target frames of three scales in advance: the large anchor frame as [139, 61], the medium anchor frame as [74, 49], and the small anchor frame as [31, 20]. The object bounding boxes at different scales estimated in the Out Block of the network are the scaling factors relative to these three potential anchor frames.

6.2.2. Evaluation Index

Compared with traditional evaluation indices, such as IoU, which is commonly used in detection tasks, we focus more on whether the detected signal anomaly is a defect, how many detected objects are correct, and how many objects are missed in the detection. Therefore, we use the following indices to evaluate the efficiency of the proposed method for defect detection. Precision ( P r ) and Recall ( R e ) are computed from True Positive ( T P ), False Positive ( F P ), True Negative ( T N ), and False Negative ( F N ), and the F 1 score is a combination of precision and recall:
P r = T P T P + F P R e = T P T P + F N F 1 = 2 × P r × R e P r + R e
The detection result contains the bounding box. If the detected boxes have an Intersection over Union (IoU) with the ground truth, they will be considered to be correct detections. In this way, Pr, Re, and F1 are computed to evaluate all methods of defect detection.

6.3. Experimental Results

The proposed approach is evaluated both qualitatively and quantitatively and compared with Thresholding [50], YOLOv3-SPP, and Faster-RCNN [35]. The locations where the detected signal is abnormal are considered to have defects. By setting a threshold interval as [ I m e a n ± 15 ] , where I m e a n is the mean of the image patch, defect candidates can be detected via the connected regions. This is a common method in the field. YOLOv3-SPP is an extension of YOLO [37]. YOLOv3-SPP and Faster-RCNN are state-of-the-art methods that are quite popular in the object detection field. Additionally, some ablation experiments are conducted to evaluate the necessity of each key module.

6.3.1. Qualitative Results

Compared to YOLOv3-SPP, and Faster-RCNN, two qualitative comparisons are shown in Figure 12 from our experiments. The first row shows the ground truth (GT), the second row shows the result from Thresholding, the third row shows the result from YOLOv3-SPP, the fourth row shows the result from Faster-RCNN, and the fifth row shows our results. The targets in red boxes are defects that need to be detected in the first row or detected defects from Thresholding, YOLOv3-SPP, Faster-RCNN [35], and the proposed approach from the second to the fifth rows. The targets in green boxes are non-defects, which are caused by mechanical jitter and texture in the first row.
Result 1 in Figure 12a shows signals from seven man-made defects on the right. The three deep learning methods, YOLOv3-SPP, Faster-RCNN, and MSTSAnet, can correctly detect all defects, while Thresholding has fewer correct bounding boxes. In Result 2, it can be seen that Thresholding almost cannot distinguish the targets in green boxes and misclassifies them as defects. Additionally, YOLOv3-SPP and Faster-RCNN cannot detect the second defect from left to right, but the proposed method can.

6.3.2. Quantitative Results

We also report the quantitative results from the proposed approach, Thresholding [50], YOLOv3-SPP, and Faster-RCNN [35] in Table 2. In Thresholding, the threshold interval is also set as [ I m e a n ± 15 ] for 1C, 2Cs, and 3Cs. The original inputs of YOLOv3-SPP and Faster-RCNN are three-channel data. When only CA data are used, the data are copied three times, and when CA and CT data are used, the third channel is filled completely with zeros.
It can be seen that the F1 score increases for all methods as the number of channels increases. The proposed method outperforms YOLOv3-SPP, Faster-RCNN, and Thresholding in F1 score while using nearly half the parameters of YOLOv3-SPP and only 6.4 % of the FLOPs of Faster-RCNN. There is a significant improvement in our method after increasing the number of channels, with about a 3 % improvement in Precision, Recall, and F1 score. The Thresholding method shows a reduction in F1 score after increasing the number of channels, despite having a 2 % higher recall but 17.56 % lower precision than the proposed method for 3Cs.
To show that TA and SA are key modules in our network, we also conducted some ablation experiments by removing TA, SA, and both. We also investigate using point-wise and depth-wise kernels instead of pooling [46] in TA and SA. It is found that the model with pooling struggles to train steadily, so we had to adjust the balance factors ( λ 1 = 1 , λ 2 = 50 , λ 3 = 20 ) to evaluate the ablation experiments. Recall and F1 scores are decreased by about 12% and 5%, respectively, after removing SA and TA in Table 3. The proposed method has higher Recall and F1 scores than all other methods because it perfectly combines the characteristics of multi-scale spatiotemporal self-attention by fusing CA and CT data.

7. Conclusions

In this paper, we propose a defect detection method based on a multi-scale spatiotemporal self-attention mechanism called MSTSA-Net for ECA. This is a deep learning approach that considers and introduces the spatiotemporal characteristics of ECA data during network learning with spatiotemporal attention. The proposed method is an end-to-end approach that primarily includes the SA, TA, and pooling modules. It evaluates the network’s performance using three different data inputs: 1 Channel (CT), 2 Channels (CA, CT), and 3Cs (CT, CA, CA+CT). The F1 score on all three different data inputs surpasses those of the Thresholding, YOLOv3-SPP, and Faster-RCNN methods comprehensively. Compared with the traditional Thresholding, the proposed method not only improves recall but also increases accuracy by over 10%, resulting in an F1 score of 87.77%. The proposed method outperforms the state-of-the-art methods, YOLOv3-SPP and Faster-RCNN, in both precision and recall. The number of parameters is only 36.4M, which is 58.1% of YOLOv3-SPP (62.6M) parameters, and its FLOPs are close to 50% of YOLOv3-SPP and 6.4% of Faster-RCNN. The experiment demonstrates that using ECA to detect defects on the surface of ferromagnetic finely grooved structures is a reliable method in practical applications. The essence of deep learning methods is to learn a nonlinear mapping relationship from input to output under the guidance of supervision. Learning a deep network requires a large amount of data. For a task with limited data, finding a neural network that matches the scale of the problem is an interesting direction and will be the future work of this paper.

Author Contributions

Conceptualization, L.B.; methodology, S.G. and Y.Z.; software, S.G.; validation, S.G., S.L. and J.Z.; formal analysis, Y.Z.; investigation, S.L.; resources, L.B. and Y.D.; data curation, J.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, Y.Z. and Y.D.; supervision, Y.D.; project administration, L.B.; funding acquisition, Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by the National Natural Science Foundation of China (No. 62303093, 61971106, 61903065, U2030205) and Sichuan Science and Technology Planning Project (No. 2022JDJQ00400, 2023YFG0044).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MSTSANetMulti-scale SpatioTemporal Self-Attention Network
ECTEddy Current Testing
ECAEddy Current Arrays
RPNRegion Proposal Network
TATemporal Attention
SASpatial Attention
BNBatch Normalization
RAMResidual Attention Module
DRAMDownsampled Residual Attention Module
MRABMixed Residual Attention Block
UpS BlockUp Sampling Block
CAChannel A
CTChannel T
YoloYou Look Only Once
FLOPsFloating point operations per second

References

  1. Song, K.; Kang, Y.H.; Zhang, L.P.; Peng, X.Z.N. Research progress in magnetic properties of steel pipe on the eddy current testing signal. J. Mater. Eng. 2015, 43, 106–112. [Google Scholar]
  2. Zhang, N.; Ye, C.; Peng, L.; Tao, Y. Novel Array Eddy Current Sensor With Three-Phase Excitation. IEEE Sensors J. 2019, 19, 7896–7905. [Google Scholar] [CrossRef]
  3. Obeid, S.; Tranjan, F.M.; Dogaru, T. Eddy current testing for detecting small defects in thin films. Aip Conf. Proc. 2007, 894, 340–345. [Google Scholar]
  4. Khanz, M.U. Detection of defect on aircraft multi-layered structure by eddy current technique. In Proceedings of the 15th World Conference on Non-Destructive Testing, Rome, Italy, 15–21 October 2000. [Google Scholar]
  5. Soni, A.K.; Sasi, B.; Thirunavukkarasu, S.; Bhagi, P.C.R. Development of eddy current probe for detection of deep subsurface defects. IETE Tech. Rev. 2016, 33, 386–395. [Google Scholar] [CrossRef]
  6. He, Y.; Luo, F.; Pan, M.; Weng, F.; Hu, X.; Gao, J.; Liu, B. Pulsed eddy current array imaging detection technology used for defect detection. Nondestruct. Test. 2010, 43, 176–181. [Google Scholar]
  7. Kiwa, T.; Kawata, T.; Yamada, H.; Tsukada, K. Fourier-transformed eddy current technique to visualize cross-sections of conductive materials. NDT E Int. 2007, 40, 363–367. [Google Scholar] [CrossRef]
  8. Grimberg, R.; Udpa, L.; Savin, A.; Steigmann, R.; Palihovici, V.; Udpa, S.S. 2D eddy current sensor array. NDT E Int. 2006, 39, 264–271. [Google Scholar] [CrossRef]
  9. Hurley, D.C.; Hedengren, K.H.; Young, J.D. Eddy Curren Arrays for Defect Detection. Rev. Prog. Quant. Nondestruct. Eval. 1992, 11, 1137–1144. [Google Scholar]
  10. Gao, S.; Zheng, Y.; Zhang, J.; Bai, L. Eddy Current Array for Defects Detection based on Spatiotemporal Self-attention Network. In Proceedings of the IEEE Conference on Sensing, Measurement and Data Analytics in the Era of Artificial Intelligence (ICSMD 2022), Harbin, China, 30 November–2 December 2022. [Google Scholar]
  11. Rifai, D.; Abdalla, A.N.; Khamsah, N.; Aizat, M. Subsurface defects evaluation using eddy current testing. Indian J. Sci. Technol. 2016, 9, 1–7. [Google Scholar] [CrossRef]
  12. Yu, Y.; Du, P.; Liao, Y. Study on effect of coil shape and geometric parameters on performance of eddy current sensor. Chin. J. Sci. Instrum. 2007, 28, 1045. [Google Scholar]
  13. Chao, W.; Kai, W.; Zheng, C.; Cui, Z. Eddy current testing on weld defect based on the dual frequency independent component analysis. In Proceedings of the IEEE International Instrumentation and Measurement Technology Conference, Turin, Italy, 22–25 May 2017. [Google Scholar]
  14. Zhou, H.T.; Hou, K.; Pan, H.L.; Chen, J.J.; Wang, Q.M. Study on the optimization of eddy current testing coil and the defect detection sensitivity. Procedia Eng. 2015, 130, 1649–1657. [Google Scholar] [CrossRef]
  15. Cao, M.; Zhang, W.; Zeng, W.; Chen, C. Research on the device of differential excitation type eddy current testing for metal defect detection. In Proceedings of the Nondestructive Evaluation/Testing: New Technology & Application, Jinan, China, 17–20 June 2013. [Google Scholar]
  16. Zhou, H.T.; Hou, K.; Pan, H.L.; Chen, J.J.; Wang, Q.M. Investigation of rectangular differential probes for pulsed eddy current non-destructive testing. Insight Non-Destr. Test. Cond. Monit. 2016, 2, 87–100. [Google Scholar] [CrossRef]
  17. Shu, L.; Songling, H.; Wei, Z.; Peng, Y. Improved immunity to lift-off effect in pulsed eddy current testing with two-stage differential probes. Russ. J. Nondestruct. Test. 2008, 2, 138–144. [Google Scholar] [CrossRef]
  18. Liu, Z.; Ge, L.; Liu, X.; Jiang, X.; Lodewijks, G. Image processing algorithms for crack detection in welded structures via pulsed eddy current thermal imaging. IEEE Instrum. Meas. Mag. 2017, 20, 34–44. [Google Scholar] [CrossRef]
  19. Mook, G.; Simonin, Y. Eddy current imaging using array probes. In Proceedings of the European Conference on Non-Destructive Testing, Prague, Czech Republic, 6–10 October 2014. [Google Scholar]
  20. Mekhalfa, F.; Nacereddine, N. Multiclass classification of weld defects in radiographic images based on support vector machines. In Proceedings of the International Conference on Signal-Image Technology & Internet-Based Systems, Marrakech, Morocco, 23–27 November 2014. [Google Scholar]
  21. Gao, B.; Li, X.; Woo, W.L.; Tian, G. Quantitative validation of eddy current stimulated thermal features on surface crack. NDT E Int. 2017, 85, 1–12. [Google Scholar] [CrossRef]
  22. Long, C.; Zhang, N.; Tao, X.; Tao, Y.; Ye, C. Resolution Enhanced Array ECT Probe for Small Defects Inspection. Sensors 2023, 23, 2070. [Google Scholar] [CrossRef]
  23. Lysenko, I.; Mirchev, Y.; Levchenko, O.; Kuts, Y.; Uchanin, V. Advantages of Using Eddy Current Array for Detection and Evaluation of Defects in Aviation Components. Int. J. NDT Days 2023, 84, 84–88. [Google Scholar]
  24. Caetano, D.M.; Rosado, L.S.; Fernandes, J.R.; Cardoso, S. Neural Networks for Defect Detection on Eddy-Currents-Based Non-Destructive Testing. In Proceedings of the IEEE SENSORS, Vienna, Austria, 29 October–1 November 2023. [Google Scholar]
  25. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012. [Google Scholar]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar]
  27. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. 2014. Available online: https://arxiv.org/pdf/1409.1556.pdf (accessed on 1 January 2020).
  28. Huang, G.; Liu, Z.; Maaten, L. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  29. Sundermeyer, M.; Ney, H.; Schlüter, R. From Feedforward to Recurrent LSTM Neural Networks for Language Modeling. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 517–529. [Google Scholar] [CrossRef]
  30. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  31. Zheng, X.; Zheng, Y.; Yang, S. Generating Multiple Hypotheses for 3D Human Mesh and Pose using Conditional Generative Adversarial Nets. In Proceedings of the Asian Conference on Computer Vision (ACCV), Macau, China, 4–8 December 2022. [Google Scholar]
  32. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  33. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  34. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  35. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  36. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  37. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  38. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  39. Fang, Y.; Liao, B.; Wang, X.; Fang, J.; Qi, J.; Wu, R.; Niu, J.; Liu, W. You Only Look at One Sequence: Rethinking Transformer in Vision through Object Detection. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Virtual, 6–14 December 2021. [Google Scholar]
  40. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-Time End-to-End Object Detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  41. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  43. Wang, Z.; Liu, Z.; Li, G.; Wang, Y.; Zhang, T.; Xu, L.; Wang, J.; Maaten, L. Spatio-Temporal Self-Attention Network for Video Saliency Prediction. IEEE Trans. Multimed. 2021, 25, 1161–1174. [Google Scholar] [CrossRef]
  44. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021. [Google Scholar]
  45. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 4 May 2021. [Google Scholar]
  46. Woo, S.; Park, J.; Lee, J.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  47. Tao, X.; Peng, L.; Tao, Y.; Ye, C. Inspection of Defects in Weld Using Differential Array ECT Probe and Deep Learning Algorithm. IEEE Trans. Instrum. Meas. 2021, 70, 2512209. [Google Scholar] [CrossRef]
  48. Yang, L.; Wang, Z.; Gao, S. Pipeline Magnetic Flux Leakage Image Detection Algorithm Based on Multiscale SSD Network. IEEE Trans. Ind. Inform. 2019, 16, 501–509. [Google Scholar] [CrossRef]
  49. Liu, B. Research on Methods of Crack Parameter Estimation and Imaging in Eddy Current Array Nondestructive Testing. Ph.D. Thesis, National University of Defense Technology, Changsha, China, 2011. [Google Scholar]
  50. Li, S. Research and Implementation of Automatic Detection of Spinning Cylinder Defects Based on Array Eddy Current. Master’s Thesis, University of Electronic and Scientific Technology of China, Chengdu, China, 2021. [Google Scholar]
  51. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
Figure 1. The sensitive area for defects detection [49].
Figure 1. The sensitive area for defects detection [49].
Sensors 24 06078 g001
Figure 2. A schematic diagram for ECA sensor. The excitation signal generated by T could be received by R 1 and R 2 , respectively. d is the diameter of coil, d 2 is the physical distance between two lines of coils.
Figure 2. A schematic diagram for ECA sensor. The excitation signal generated by T could be received by R 1 and R 2 , respectively. d is the diameter of coil, d 2 is the physical distance between two lines of coils.
Sensors 24 06078 g002
Figure 3. A real data from ECA in the manner of a pseudo-color image. (a,b) are from two different channels. d in (a) is the distance between the centers of signal change at two time points in the image coordinate.
Figure 3. A real data from ECA in the manner of a pseudo-color image. (a,b) are from two different channels. d in (a) is the distance between the centers of signal change at two time points in the image coordinate.
Sensors 24 06078 g003
Figure 4. A real data from ECA represented by a pseudo-color image. The first row shows the artificial defect shapes with depths of 0.1 mm, 0.2 mm, and 0.5 mm, respectively. The second row shows the pseudo-color image from CA with the arrangement of the roughly horizontal coils. The third row shows the pseudo-color image from CT with the arrangement of the vertical coils.
Figure 4. A real data from ECA represented by a pseudo-color image. The first row shows the artificial defect shapes with depths of 0.1 mm, 0.2 mm, and 0.5 mm, respectively. The second row shows the pseudo-color image from CA with the arrangement of the roughly horizontal coils. The third row shows the pseudo-color image from CT with the arrangement of the vertical coils.
Sensors 24 06078 g004
Figure 5. (a) shows a real example of fine grooves surface in our experiments. (b) shows the schematic diagram for the acquisition platform.
Figure 5. (a) shows a real example of fine grooves surface in our experiments. (b) shows the schematic diagram for the acquisition platform.
Sensors 24 06078 g005
Figure 6. A schematic diagram of the lift-off distance changing as the rotation of the mechanical spindle.
Figure 6. A schematic diagram of the lift-off distance changing as the rotation of the mechanical spindle.
Sensors 24 06078 g006
Figure 7. The network architecture of MSTSA-Net. Our network consists of four MRAMs, two UpS blocks, and three Output blocks, including DRAM and RAM, which are stacked four times into a pyramid. Conv2D(3 × 3, i) represents a 2D 3 × 3 convolution with a stride of i.
Figure 7. The network architecture of MSTSA-Net. Our network consists of four MRAMs, two UpS blocks, and three Output blocks, including DRAM and RAM, which are stacked four times into a pyramid. Conv2D(3 × 3, i) represents a 2D 3 × 3 convolution with a stride of i.
Sensors 24 06078 g007
Figure 8. The details of SA and TA block for self-attention [10].
Figure 8. The details of SA and TA block for self-attention [10].
Sensors 24 06078 g008
Figure 9. Implementation details for each block.
Figure 9. Implementation details for each block.
Sensors 24 06078 g009
Figure 10. Samples from the dataset of training, objects in the box on CA channel are defects candidates labeled by manual. The targets in the red box are defects, and the targets in the green box are signal extremum but non-defects, which are caused by mechanical jitter and texture.
Figure 10. Samples from the dataset of training, objects in the box on CA channel are defects candidates labeled by manual. The targets in the red box are defects, and the targets in the green box are signal extremum but non-defects, which are caused by mechanical jitter and texture.
Sensors 24 06078 g010
Figure 11. The visual image examples of man-made defects and natural defects. (a) shows four man-made defects, in which the right three defects have the same length but different depths. (b) shows an example of a natural defect.
Figure 11. The visual image examples of man-made defects and natural defects. (a) shows four man-made defects, in which the right three defects have the same length but different depths. (b) shows an example of a natural defect.
Sensors 24 06078 g011
Figure 12. Comparison results of defect detection. The targets in red boxes are defects that need to be detected in the first row or detected by Thresholding [50], YOLOv3-SPP, and Faster-RCNN [35] from the second to the fifth rows. The targets in green boxes are non-defects caused by mechanical jitter and texture. The four defects on the right in Result 1 are the four man-made defects in Figure 11a.
Figure 12. Comparison results of defect detection. The targets in red boxes are defects that need to be detected in the first row or detected by Thresholding [50], YOLOv3-SPP, and Faster-RCNN [35] from the second to the fifth rows. The targets in green boxes are non-defects caused by mechanical jitter and texture. The four defects on the right in Result 1 are the four man-made defects in Figure 11a.
Sensors 24 06078 g012
Table 1. The details of training and testing dataset.
Table 1. The details of training and testing dataset.
DatasetTrainingTesting
No. Patches776114
No. Defective patches36877
No. Objects495138
No. Big objects24866
No. Small objects24772
Table 2. Quantitative comparison results.
Table 2. Quantitative comparison results.
MethodsFLOPs1C2Cs3Cs
PrReF1PrReF1PrReF1
Thresholding [50]69.49%78.85%73.87%69.49%78.85%73.87%67.13%88.31%76.18%
YOLOv3-SPP 6.24 × 10 9 86.60%80.77%83.58%82.14%84.01%83.75%81.58%89.42%85.32%
Faster-RCNN [35] 5.10 × 10 10 87.49%69.46%77.44%74.38%86.54%79.99%78.38%83.65%80.93%
Ours 3.28 × 10 9 81.25%87.50%84.59%82.68%88.46%85.19%84.69%90.38%87.77%
Table 3. Ablative experimental results on 3Cs with ( λ 1 = 1 , λ 2 = 50 , λ 3 = 20 ).
Table 3. Ablative experimental results on 3Cs with ( λ 1 = 1 , λ 2 = 50 , λ 3 = 20 ).
MethodsPrecisionRecallF1
Without TA87.49%74.04%80.20%
Without SA79.41%77.88%78.64%
Without SA and TA84.24%76.92%80.39%
With pooling89.29%72.12%79.79%
Ours82.14%88.46%85.18%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, S.; Zheng, Y.; Li, S.; Zhang, J.; Bai, L.; Ding, Y. Eddy Current Array for Defect Detection in Finely Grooved Structure Using MSTSA Network. Sensors 2024, 24, 6078. https://doi.org/10.3390/s24186078

AMA Style

Gao S, Zheng Y, Li S, Zhang J, Bai L, Ding Y. Eddy Current Array for Defect Detection in Finely Grooved Structure Using MSTSA Network. Sensors. 2024; 24(18):6078. https://doi.org/10.3390/s24186078

Chicago/Turabian Style

Gao, Shouwei, Yali Zheng, Shengping Li, Jie Zhang, Libing Bai, and Yaoyu Ding. 2024. "Eddy Current Array for Defect Detection in Finely Grooved Structure Using MSTSA Network" Sensors 24, no. 18: 6078. https://doi.org/10.3390/s24186078

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop