Next Article in Journal
An Enhanced Direct Position Determination of Mixed Circular and Non-Circular Sources Using Moving Virtual Interpolation Array
Previous Article in Journal
Remote Sensing Inversion of Water Quality Grades Using a Stacked Generalization Approach
Previous Article in Special Issue
Ultrasonic Weld Quality Inspection Involving Strength Prediction and Defect Detection in Data-Constrained Training Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-Modal Illumination System for Defect Detection of Aircraft Glass Canopies

School of Computer Science and Technology, Xidian University, Xi’an 710126, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(20), 6717; https://doi.org/10.3390/s24206717 (registering DOI)
Submission received: 22 September 2024 / Revised: 13 October 2024 / Accepted: 16 October 2024 / Published: 18 October 2024
(This article belongs to the Special Issue Deep-Learning-Based Defect Detection for Smart Manufacturing)

Abstract

:
Defect detection in transparent materials typically relies on specific lighting conditions. However, through our work on defect detection for aircraft glass canopies, we found that using a single lighting condition often led to missed or false detections. This limitation arises from the optical properties of transparent materials, where certain defects only become sufficiently visible under specific lighting angles. To address this issue, we developed a dual-modal illumination system that integrates both forward and backward lighting to capture defect images. Additionally, we introduced the first dual-modal dataset for defect detection in aircraft glass canopies. Furthermore, we proposed an attention-based dual-branch modal fusion network (ADMF-Net) to enhance the detection process. Experimental results show that our system and model significantly improve the detection performance, with the dual-modal approach increasing the mAP by 5.6% over the single-modal baseline, achieving a mAP of 98.4%. Our research also provides valuable insights for defect detection in other transparent materials.

1. Introduction

Transparent materials are widely used in canopies, windshields, and windows of aircraft. As external structural components of the aircraft and vision windows for pilots, these materials must possess excellent mechanical strength and optical properties to ensure operational safety. However, due to the complexities inherent in the manufacturing processes and the continuous exposure to environmental factors, such as high speed dust impact and abrasion during flight [1], surface defects, such as cracks, scratches and pits, inevitably develop over time. These defects can interfere with the pilot’s field of vision and, in more severe cases, present significant structural safety risks [2,3].
Currently, the inspection and maintenance of aircraft transparent materials rely heavily on manual visual inspections. This approach is not only labor-intensive and time-consuming but also prone to human error, resulting in inconsistent outcomes and the potential for missed defects [4]. Therefore, the development of an automated, accurate and efficient defect detection system is crucial for improving the quality control and maintenance processes for aircraft transparent materials.
In recent years, machine vision technology has been increasingly applied to defect detection and quality control in industrial manufacturing. This paper seeks to bridge the gap between research and industrial application by introducing a novel defect detection system for aircraft glass canopies, along with a defect dataset and detection model.
Our contributions can be summarized as follows:
  • An automated prototype system with a dual-modal illumination structure is proposed. For the same sampling point, the system sequentially uses forward lighting and backward lighting to capture defect images. By integrating the information from both types of images, more accurate defect detection is achieved.
  • A curated dataset is presented, which, to the best of our knowledge, is the first benchmark in the field of defect detection for aircraft transparent materials.
  • A dual-modal baseline is designed and proved to be competitive, in which we designed two fusion detection methods: a data-level fusion method named RGB Channel Fusion, and an attention-based dual-branch modal fusion network using feature-level fusion. Both methods achieved excellent results on our dataset.

2. Related Work

Product defect detection is a crucial aspect of quality control in the manufacturing industry. Automated Optical Inspection (AOI), as a non-contact testing technique, is widely used across various product production lines [5]. However, research on defect detection systems for aircraft transparent materials lags far behind. Defect detection techniques in other fields can serve as valuable references for this study.

2.1. Traditional Machine Vision Method for Defect Detection in Transparent Materials

Traditional defect detection for transparent materials often relies on specialized optical or acoustic instruments [6,7]. With the development of image processing technology and machine learning, machine vision methods have started to be applied. These methods use conventional image processing techniques, such as edge detection, texture analysis, or shape matching, to extract and identify defect objects.
For instance, Zhou X et al. [8] proposed a defect detection strategy for glass bottle bottoms based on wavelet transform and multi-scale filtering algorithms, which effectively reduces the impact of texture on defect detection. Jian C et al. [9] captured images of mobile phone cover glass under coaxial parallel lighting and used template subtraction and projection methods for defect detection.
Traditional machine learning-based defect detection methods often rely on the comparison between defect images and defect-free templates [10], making them very sensitive to noise, background interference, and uneven lighting. These methods are mainly suitable for defect detection in controlled scenarios.

2.2. Deep Learning-Based Visual Defect Detection

In recent years, deep learning-based machine vision methods have been widely applied in the field of defect detection. Compared to traditional machine learning methods, deep learning offers more complex model fitting capabilities, allowing it to learn high-level abstract features [11]. This results in better accuracy and generalization in defect detection tasks.
For instance, Park J et al. [12] proposed a defect detection system for curved glass of mobile phones based on weighted multi-neural networks, effectively utilizing multi-channel measurement data. Cheng L et al. [13] introduced an infrared thermography-based generative adversarial network (IRT-GAN) for non-destructive testing of carbon fiber and glass fiber polymer composites. Zhu Y et al. [14] introduced a ResNeSt network based on deformable convolution to detect fine features of surface defects on mobile phones, addressing various defect shapes. They also used RoI Align to reduce localization errors, particularly for scratch defects with high aspect ratios.
R-CNN algorithms, including Fast R-CNN and Faster R-CNN, are well-established two-stage object detection algorithms. Pan Z et al. [15] developed an automatic surface scratch inspection system using a rail-based microscopic camera scanning system for detecting defects in architectural glass curtain walls. They used a Mask R-CNN network model to achieve pixel-level instance segmentation.
The YOLO series algorithms, as the most typical representatives of one-stage object detection algorithms, transform the object detection process into a regression task, significantly speeding up detection and maintaining a leading position in real-time object detection tasks [16]. Zhang M et al. [17] adopted an improved YOLOv5 algorithm for defect detection in solar cells, incorporating deformable convolutions, ECA-Net attention mechanism, Mosaic and MixUp data augmentation, and K-means++ for anchor box clustering, improving the algorithm’s mAP by 7.85% and achieving a detection speed of 36.24 FPS, meeting real-time detection requirements. Xie Y et al. [18] proposed a feature-enhanced surface defect detection algorithm (FE-YOLO) for practical industrial applications, combining depthwise separable convolution with dense connections and introducing an improved feature pyramid network to enhance spatial position correlation in multi-scale detection layers, achieving high accuracy while reducing model weight.

2.3. Multi-Modal Fusion in Object Detection

Multi-modal information fusion has gained attention in the field of object detection. These methods leverage the complementary information from different modalities to construct more comprehensive and enriched features, thereby improving detection accuracy.
In multi-modal image object detection, fusion methods are generally categorized into three types [19]: data-level fusion (early fusion), feature-level fusion (intermediate fusion), and decision-level fusion (late fusion).
A typical example is the fusion of visible light and infrared images for nighttime pedestrian detection. Data-level fusion involves combining visible light and infrared images into a single image using specific algorithms or multiple data channels before feeding them into the detection model. Decision-level fusion uses two separate models to detect visible and infrared images independently, focusing on post-processing the outputs of the two models. Feature-level fusion uses parallel branches to extract independent features from images at different scales, embedding a feature fusion layer between the feature layers, and using the fused features for detection.
Compared to decision-level fusion, a well-designed feature-level fusion method enables the model to learn the correlations and differences between different modalities, and reduces redundant computations.
Multi-modal fusion object detection has been applied and validated in various fields. Examples include RGB and infrared thermal image fusion for object detection [20,21], RGB and radar information fusion for assisted driving [22], polarization and infrared imaging fusion for PCB defect detection [23], and multi-spectral image fusion for medical image segmentation [24].

3. Methodology

3.1. Dual-Modal Illumination Structure

In industrial machine vision defect detection, an appropriate illumination structure design is crucial for capturing defect images. Among various illumination structures, forward lighting is the most widely used. Depending on whether the light is reflected directly into the camera lens, which is usually controlled by the incident angle of the light source, forward lighting can be divided into bright field forward lighting and dark field forward lighting, as shown in Figure 1a. Additionally, for highly reflective surfaces, scattering forward lighting can effectively reduce the strong reflections [25]. This type of lighting is usually achieved using a dome structure or multi-angle light source to provide uniform illumination, as shown in Figure 1b.
In contrast to forward lighting, the backward lighting method places the light source behind the object, as depicted in Figure 1c. Backward lighting is commonly used for defect detection in transparent materials as it can highlight internal defects, but it is less sensitive to surface defects [26].
To design an appropriate illumination structure for this study, we captured images of typical defects under different lighting conditions, with some of the results shown in Figure 2.
Through the analysis of these images collected, we found that bright field forward lighting easily captures surface scratch-type defects but is susceptible to reflections and uneven lighting. Backward lighting effectively reveals the depth information of defects, aiding in the analysis of defect severity and classification. However, it tends to lose surface detail and some boundary contour information. The effect of dark field forward lighting is similar to backward lighting, resembling an inverted operation.
We realized that the information provided by multiple modalities is complementary, contributing to a more accurate assessment of defect conditions. Our subsequent experiments in this paper also verified that the multi-modal defect detection method has significantly improved performance compared to the single-modal method.

3.2. Image Acquisition Platform

Considering the practical feasibility in industrial inspection environments, we ultimately chose to use a dual-modal illumination structure, employing both forward lighting and backward lighting to capture and detect defects. For the implementation of forward lighting, we use a ring light source to achieve more uniform illumination and reduce reflections. For backward lighting, we use a planar light source.
We built a prototype dual-modal illumination image acquisition platform, as shown in Figure 3. This platform mainly consists of a digital microscope, a flexible arm holder, a backlight panel, and a ring light source.
The defects on glass canopies mostly range in diameter from 80 μ m to 500 μ m, making it impractical to use conventional CCD cameras for imaging. Therefore, we use a digital microscope as the image acquisition device. This device features an adjustable focal length ranging from 8 mm to 60 mm, with a magnification range of 50× to 1000×, and a photo resolution of 1280 × 720 . During image acquisition, the focal length was set to 20 mm, with an approximate magnification of 400×.
Due to the necessity of maintaining a fixed position between the camera lens and the sample when capturing images under two lighting conditions, we used a flexible arm holder as the support device. This device has a length of 60 cm and a maximum load capacity of 1.0 kg, providing sufficient stability for the image acquisition process.
The forward lighting method uses 8 ring-shaped LED lights positioned around the camera lens. The backward lighting method uses a 30 cm × 42 cm backlight panel with a brightness of 500 nits and a color temperature of 6500 K. The two light sources are enabled alternately by an electrically controlled switch during image acquisition. See Figure 4.

3.3. Dataset Construction

Due to the specialized nature of the research, publicly available defect sample data for aircraft glass canopies are not currently accessible. Therefore, we collected data from our own samples of real aircraft cockpit canopies, which served as the source for our subsequent experiments. See Figure 5.
Based on the real aircraft glass canopies samples, we collected a total of 438 raw images, covering 219 defect points using two different lighting methods. Through image analysis and relevant experience, we categorized the collected defects into four types: contusion, scratches, crack, and spot. These categories essentially cover all possible defect types that may appear on the aircraft glass canopy. Typical images are shown in Figure 6.
Considering that many of the detected target objects are slender, that is, they have high aspect ratios, large image spans, and may be interlaced, we choose to use oriented bounding boxes (OBB) for annotation and detection.
We use X-AnyLabeling [27] (version 2.3.5) for data annotation, which allows the use of existing segmentation models to improve annotation efficiency. We choose the Segment Anything (ViT-Large) model for pre-labeling, which provides preparatory polygon boundaries of the selected defects. Then, a script is written to calculate the minimum enclosing rectangle for each defect, resulting in a more precise rotated bounding box.
Manual adjustments are also necessary. This is because some defects are only clearly outlined under a specific lighting condition. For example, scratch defects are usually easier to observe under forward lighting. However, some small spot-type defects that are not apparent under forward lighting are very prominent under backward lighting. Additionally, there are also subtle differences in the contour boundaries of the defects under the two types of lighting, and in that case, the union outer contour will be selected. We check each annotation file to ensure the accuracy and usability of the dataset.
To address the limited number of defect images available for aviation transparent parts, we also adopted several data augmentation methods. In order to improve the defect detection performance under any rotation angle, we performed pre-rotation augmentation on the defect samples. In this process, rotation and mirroring operations are simultaneously applied to the images and their labels to avoid redundant annotation work. By adjusting the step size of the rotation angle, the augmentation rate can be controlled. For example, using 90-degree rotations and mirroring operations, an 8× augmentation can be achieved without information loss. For general angles, the background color RGB(114, 114, 0) is filled in the blank areas to reduce interference. Moreover, since the standard OBB format does not allow point coordinates to be located outside the image boundaries, labels that overflow during the rotation process need to be adjusted or deleted. The correction method for overflow labels is as follows:
  • If all four points of the bounding box are outside the field of view, delete the target (process ends), otherwise go to step 2.
  • If all four points of the bounding box are within the field of view, no adjustment is required (process ends), otherwise go to step 3.
  • Select a point that is outside the field of view. Calculate the distance between the point and its adjacent two points to find the long and short sides connected to the point on the rectangle. Translate the short side along the direction of the long side until the point falls on the image boundary. The translation amount is determined by solving a linear programming problem. After adjustment, return to step 2 to re-evaluate.
By following these steps, we ensure that the labels are correctly normalized. Additionally, during the training process, we employed the Mosaic-4 data augmentation method, which includes random image combinations, random scaling and translation, contrast adjustments, exposure adjustments, and noise addition.
The final dataset of aircraft glass canopy defects, named ag_dual_obb, includes four types of defects, totaling 1752 defect points and 4784 defect objects. Each defect point includes two images: a forward lighting image and a backward lighting image, which are strictly spatially aligned. The dataset is divided into a training set with 1576 image pairs and a test set with 176 image pairs, following a 9:1 split ratio. See Table 1.
We also created some variant datasets for extra experiments, which follow similar naming. For example, the dataset labeled with standard rectangular box is named ag_dual_rect, and the dataset obtained through RGB Channel Fusion is named ag_composite_obb.

3.4. RGB Channel Fusion Method

To initially validate the effectiveness of fusion detection for this task, we first designed a data-level fusion method named RGB Channel Fusion, as shown in Figure 7. In the RGB Channel Fusion method, the forward lighting image and the backward lighting image are each converted to grayscale. These grayscale images are then assigned to the red channel and green channel of the fused image, respectively, while the blue channel is obtained by averaging the two grayscale images.
The design rationale for the RGB Channel Fusion method is based on the fact that, for this specific task, grayscale conversion hardly diminishes the image information. Moreover, the averaging of images can serve as a simple yet effective image fusion technique. In addition, the resulting color of the fused image can provide a reference for our data annotation work. For example, a dark red color usually means deeper damage, which helps distinguish surface scratches from more serious contusions.

3.5. Attention-Based Dual-Branch Modal Fusion Network (ADMF-Net)

To fully utilize the multimodal information obtained under different lighting conditions, this paper proposes an attention-based dual-branch modal fusion network (ADMF-Net) for defect detection. The overall architecture of the network is shown in Figure 8. It consists of two independent feature extraction backbones, feature fusion modules (AMFF, Attention-based Multi-modal Feature Fusion Module), a multi-scale feature pyramid pooling module (FSPPF, Fusion-based Spatial Pyramid Pooling - Fast), and detection layers.
In this network, we use two CSPDarknet-like feature extraction backbones, as proposed in YOLOv8, to extract multi-scale features from the forward lighting and backward lighting modalities separately. Next, two AMFF modules are inserted between the two backbones, applied to the P3 and P4 layers, to obtain fused features with different receptive fields. Then, the P5 layers from both modalities are connected to a spatial pyramid pooling layer, FSPPF. Finally, the three scales of features are fed into the detection layers, using an OBB detection head to predict the defects with rotated bounding boxes.
The AMFF module consists of a Concat module and a convolutional block attention module (CBAM). The feature fusion process involves three steps.
First, the Concat operation is applied to the feature maps of the two modalities to create the initial fused feature.
F = [ F forward ; F backward ]
Second, a channel attention module (CAM) is used to adjust the channel weights between and within the two modalities. The input feature map F R ( C × H × W ) is transformed into C × 1 × 1 using maximum pooling and average pooling, respectively. These two feature vectors are then fed into a shared MLP for learning the attention weights of channels. The results are combined using the addition operation and normalized by the sigmoid function. Multiplying the final weight coefficient M c R ( C × 1 × 1 ) with the original feature map gives the new feature map F .
M c ( F ) = σ ( MLP ( AvgPool ( F ) ) + MLP ( MaxPool ( F ) ) ) = σ W 1 W 0 F avg c + W 1 W 0 F max c
F = M c ( F ) F
Third, the fused feature is fed into the spatial attention module (SAM) to enhance the attention on the important defect areas and to reduce the weight of background areas. In this process, maximum and average pooling are applied to each channel to obtain two feature maps of 1 × H × W , then concatenated by channel and sent into a 7 × 7 convolution layer. After use of the sigmoid activation function, the spatial attention map M s R ( 1 × H × W ) is obtained. Lastly, the M s and F are multiplied to obtain the final fused feature map F .
M s ( F ) = σ f 7 × 7 ( [ AvgPool ( F ) ; MaxPool ( F ) ] ) = σ f 7 × 7 [ F avg s ; F max s ]
F = M s ( F ) F
The FSPPF module first concatenates the outputs of the P5 layers from the two feature extraction backbones, and then passes it into the SPPF (Spatial Pyramid Pooling-Fast) module. Then, similarly, the attention mechanism is applied to refine and produce the final feature map. The output of FSPPF, along with the outputs from the P3 and P5 layers’ AMFF modules, serves as the input to the neck layer.

4. Experiments

4.1. Evaluation Metrics

We evaluate the model’s detection performance using the mean average precision (mAP). Precision is the ratio of true positives (TP) to the sum of true positives and false positives (FP), indicating the accuracy of positive predictions. Recall is the ratio of true positives to the sum of true positives and false negatives (FN), reflecting how well the model detects all relevant instances. Average precision (AP) measures the area under the precision-recall curve, denoted as p ( r ) , which describes the relationship between precision and recall as the confidence threshold changes. The final mAP is obtained by averaging the AP values across all classes. In this study, we use both the conventional mAP50, which evaluates at an intersection over union (IoU) threshold of 0.5, and the stricter mAP50-95, which averages over multiple IoU thresholds from 0.5 to 0.95 in steps of 0.05.
Precision = T P T P + F P Recall = T P T P + F N A P = 0 1 p ( r ) d r m A P = 1 N A P i

4.2. Implementation Details

The training environment for the experiments was configured with Python version 3.10.14 and PyTorch version 2.2.2, with CUDA version 11.8 for GPU acceleration. The Ultralytics version 8.2.2 was employed to support model implementation. The hardware setup included an AMD Ryzen™ 7 7800X3D CPU with 8 cores and 16 threads, 48GB of memory, and a NVIDIA® GeForce RTX™ 4080 GPU with 16GB of VRAM. During the training phase, we adopted the hyperparameter settings shown in Table 2.

4.3. Results and Discussion

4.3.1. Baseline and Dual-Modal Validity Discussion

First, we used YOLOv8m as the baseline model, training on the single-modal datasets to test the baseline performance. The results show that using only forward lighting images achieved a mAP50 of 92.8%, while using only backward lighting images achieved a mAP50 of 79% (see Table 3). However, as seen in the detailed results for each defect type, backward lighting significantly outperforms forward lighting in detecting contusions and spot defects. In contrast, its AP for scratches is only 23%, which considerably lowers its overall mAP (see Table 4).
Next, we trained the model on the ag_composite_obb dataset, which is created using the RGB Channel Fusion method. The results show that our simple channel fusion method achieved impressive improvements. The overall mAP reached 97.9%, which is close to the natural error margin of manual annotations. Furthermore, the AP for each defect type surpassed the highest levels achieved using a single lighting method, with only a slight decrease in AP50-95 for the crack defect category.
Although the overall performance of forward lighting is barely satisfactory, it cannot fully replace the detection capabilities of backward lighting. Combining both methods for defect detection is obviously more effective and holds potential for further improvement.

4.3.2. ADMF-Net Experiment

Our fusion model achieved excellent results on the ag_dual_obb dataset. The detection rates for the contusion, crack, and spot categories all exceeded 99%. The detection rate for scratches was 95.4%, which is consistent with the results of the RGB Channel Fusion method. Overall, compared to the RGB Channel Fusion method, the total mAP50 improved from 97.9% to 98.4%. As shown in Figure 9, our fusion model demonstrated a better precision-confidence curve and precision-recall curve, although there was a slight decrease in mAP50-95.

4.3.3. Interpretability Experiment

A confusion matrix is a table used to evaluate classification models by comparing predicted and actual values. As shown in Figure 10, the model trained with forward lighting tends to misclassify lighter scratch defects as contusions and struggles to detect spot defects. For backward lighting, a significant number of scratch defects are incorrectly classified as background, which is the primary factor limiting its overall performance. These issues were significantly improved with the adoption of the fusion detection method.

4.3.4. Comparative Analysis of Results

We conducted several comparative experiments to verify the effectiveness of our proposed method. See Table 5. For the selection of the third channel in the RGB Channel Fusion method, we experimented with using the difference image of the two modalities instead of the mean value. This resulted in a slight decrease in mAP.
To validate the effectiveness of our fusion model, we conducted ablation experiments. The results showed that after removing the attention mechanism, mAP50 remained relatively stable, but mAP50-95 significantly dropped by 5.2%.
Additionally, we conducted further experiments to evaluate the impact of different attention mechanisms on detection performance, such as EMA attention and a state-of-the-art model based on transformer cross-attention. The experimental results indicate that these more complex attention mechanisms could have negative effects. Our proposed model proves to be more effective for defect detection in aircraft glass canopies.

5. Conclusions

To address the issue of surface defect detection on aircraft glass canopies, where relying on a single-modal lighting condition often leads to missed or false detections, we have proposed a dual-modal illumination system for transparent material defect detection. This system utilizes both forward lighting and backward lighting to capture defect images, facilitating a more comprehensive and accurate defect detection process.
In addition, based on real aircraft glass canopy samples, we have collected the first dataset for defect detection in aircraft glass canopies. This dataset includes 1752 defect points captured under both forward and backward lighting conditions, totaling 4784 defect samples. These samples are categorized into four classes and annotated using the OBB format.
Furthermore, we proposed a data-level fusion method named RGB Channel Fusion, and a feature-level fusion model, the attention-based dual-branch modal fusion network (ADMF-Net). The model uses two feature extraction backbones to independently extract multi-scale features from the forward and backward lighting modalities. Next, through the attention-based multi-modal feature fusion block (AMFF) and improved SPPF module, the network adaptively integrates the multi-scale features from the two modalities. Finally, the features are fed into the detection layers, where an OBB detection head is used to provide the rotated bounding boxes of defects.
Experiments demonstrate the effectiveness of our proposed system and dual-modal detect methods. Compared to the baseline model YOLOv8m, which individually detects forward lighting and backward lighting images, our ADMF-Net model improved the detection accuracy by 5.6% and 19.4%, reaching a mAP of 98.4%. The proposed RGB Channel Fusion method also achieved a strong mAP of 97.9%. The comparative experiments also confirmed the superiority of our model.
Our research effectively addresses the issue of defect detection in aircraft glass canopies. Moreover, it provides a new methodological support for defect detection in other transparent materials. Exploring specific multi-modal detection networks remains a worthwhile issue.

Author Contributions

Conceptualization: Y.Y. and Z.L.; methodology: Z.L. and Y.Y.; data curation: R.W., Z.L. and Q.L.; original draft: Z.L.; review and editing: R.W. and Q.L.; software: Z.L. and Q.L.; validation: Q.L. and R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are publicly available for research purposes at the following GitHub repository: https://github.com/core128/AGDD (accessed on 14 October 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Towsyfyan, H.; Biguri, A.; Boardman, R.; Blumensath, T. Successes and challenges in non-destructive testing of aircraft composite structures. Chin. J. Aeronaut. 2020, 33, 771–791. [Google Scholar] [CrossRef]
  2. Keršienė, N.; Raslavičius, L.; Keršys, A.; Kažys, R.; Žukauskas, E. Energo-mechanical evaluation of damage growth and fracture initiation in aviation composite structures. Polym.-Plast. Technol. Eng. 2016, 55, 1137–1144. [Google Scholar] [CrossRef]
  3. Zimmermann, N.; Wang, P.H. A review of failure modes and fracture analysis of aircraft composite materials. Eng. Fail. Anal. 2020, 115, 104692. [Google Scholar] [CrossRef]
  4. Kopardekar, P.; Mital, A.; Anand, S. Manual, Hybrid and Automated Inspection Literature and Current Research. Integr. Manuf. Syst. 1993, 4, 18–29. [Google Scholar] [CrossRef]
  5. Ebayyeh, A.A.R.M.; Mousavi, A. A Review and Analysis of Automatic Optical Inspection and Quality Monitoring Methods in Electronics Industry. IEEE Access 2020, 8, 183192–183271. [Google Scholar] [CrossRef]
  6. Costa, R.D.; Maia, J.M.; Assef, A.A.; Pichorim, S.F.; Gewehr, P.M.; Costa, E.T. Defect Detection in Aluminum Bars Using Impedance and Ultrasonic Attenuation. IEEE Sens. J. 2020, 20, 7400–7413. [Google Scholar] [CrossRef]
  7. Yao, H.B.; Ping, J.; Ma, G.D.; Li, L.W.; Gu, J.N. The system research on automatic defect detection of glasses. Appl. Mech. Mater. 2013, 437, 362–365. [Google Scholar] [CrossRef]
  8. Zhou, X.; Wang, Y.; Zhu, Q.; Mao, J.; Xiao, C.; Lu, X.; Zhang, H. A Surface Defect Detection Framework for Glass Bottle Bottom Using Visual Attention Model and Wavelet Transform. IEEE Trans. Ind. Inform. 2020, 16, 2189–2201. [Google Scholar] [CrossRef]
  9. Jian, C.; Gao, J.; Ao, Y. Automatic surface defect detection for mobile phone screen glass based on machine vision. Appl. Soft Comput. J. 2017, 52, 348–358. [Google Scholar] [CrossRef]
  10. Nand, G.K.; Noopur; Neogi, N. Defect detection of steel surface using entropy segmentation. In Proceedings of the 11th IEEE India Conference: Emerging Trends and Innovation in Technology, INDICON 2014, Pune, India, 11–13 December 2014. [Google Scholar] [CrossRef]
  11. Czimmermann, T.; Ciuti, G.; Milazzo, M.; Chiurazzi, M.; Roccella, S.; Oddo, C.M.; Dario, P. Visual-based defect detection and classification approaches for industrial applications—A SURVEY. Sensors 2020, 20, 1459. [Google Scholar] [CrossRef]
  12. Park, J.; Riaz, H.; Kim, H.; Kim, J. Advanced cover glass defect detection and classification based on multi-DNN model. Manuf. Lett. 2020, 23, 53–61. [Google Scholar] [CrossRef]
  13. Cheng, L.; Tong, Z.; Xie, S.; Kersemans, M. IRT-GAN: A generative adversarial network with a multi-headed fusion strategy for automated defect detection in composites using infrared thermography. Compos. Struct. 2022, 290, 115543. [Google Scholar] [CrossRef]
  14. Zhu, Y.; Ding, R.; Huang, W.; Wei, P.; Yang, G.; Wang, Y. HMFCA-Net: Hierarchical multi-frequency based Channel attention net for mobile phone surface defect detection. Pattern Recognit. Lett. 2022, 153, 118–125. [Google Scholar] [CrossRef]
  15. Pan, Z.; Yang, J.; Wang, X.; Wang, F.; Azim, I.; Wang, C. Image-based surface scratch detection on architectural glass panels using deep learning approach. Constr. Build. Mater. 2021, 282, 122717. [Google Scholar] [CrossRef]
  16. Hussain, M. YOLO-v1 to YOLO-v8, the Rise of YOLO and Its Complementary Nature toward Digital Manufacturing and Industrial Defect Detection. Machines 2023, 11, 677. [Google Scholar] [CrossRef]
  17. Zhang, M.; Yin, L. Solar Cell Surface Defect Detection Based on Improved YOLO v5. IEEE Access 2022, 10, 80804–80815. [Google Scholar] [CrossRef]
  18. Xie, Y.; Hu, W.; Xie, S.; He, L. Surface Defect Detection Algorithm Based on Feature-Enhanced YOLO. Cogn. Comput. 2023, 15, 565–579. [Google Scholar] [CrossRef]
  19. Zhang, Y.; Sidibé, D.; Morel, O.; Mériaudeau, F. Deep multimodal fusion for semantic image segmentation: A survey. Image Vis. Comput. 2021, 105, 104042. [Google Scholar] [CrossRef]
  20. Chen, H.; Li, Y.; Su, D. Multi-modal fusion network with multi-scale multi-path and cross-modal interactions for RGB-D salient object detection. Pattern Recognit. 2019, 86, 376–385. [Google Scholar] [CrossRef]
  21. Wang, Q.; Chi, Y.; Shen, T.; Song, J.; Zhang, Z.; Zhu, Y. Improving RGB-Infrared Object Detection by Reducing Cross-Modality Redundancy. Remote Sens. 2022, 14, 2020. [Google Scholar] [CrossRef]
  22. Feng, D.; Haase-Schutz, C.; Rosenbaum, L.; Hertlein, H.; Glaser, C.; Timm, F.; Wiesbeck, W.; Dietmayer, K. Deep Multi-Modal Object Detection and Semantic Segmentation for Autonomous Driving: Datasets, Methods, and Challenges. IEEE Trans. Intell. Transp. Syst. 2021, 22, 1341–1360. [Google Scholar] [CrossRef]
  23. Li, M.; Yao, N.; Liu, S.; Li, S.; Zhao, Y.; Kong, S.G. Multisensor Image Fusion for Automated Detection of Defects in Printed Circuit Boards. IEEE Sens. J. 2021, 21, 23390–23399. [Google Scholar] [CrossRef]
  24. Zhou, T.; Ruan, S.; Canu, S. A review: Deep learning for medical image segmentation using multi-modality fusion. Array 2019, 3–4, 100004. [Google Scholar] [CrossRef]
  25. Ren, Z.; Fang, F.; Yan, N.; Wu, Y. State of the Art in Defect Detection Based on Machine Vision. Int. J. Precis. Eng. and Manuf.-Green Tech. 2022, 9, 661–691. [Google Scholar] [CrossRef]
  26. Ming, W.; Shen, F.; Li, X.; Zhang, Z.; Du, J.; Chen, Z.; Cao, Y. A comprehensive review of defect detection in 3C glass components. Measurement 2020, 158, 107722. [Google Scholar] [CrossRef]
  27. Wang, W. Advanced Auto Labeling Solution with Added Features. 2023. Available online: https://github.com/CVHub520/X-AnyLabeling (accessed on 14 October 2024).
  28. Shen, J.; Chen, Y.; Liu, Y.; Zuo, X.; Fan, H.; Yang, W. ICAFusion: Iterative cross-attention guided feature fusion for multispectral object detection. Pattern Recognit. 2024, 145, 109913. [Google Scholar] [CrossRef]
  29. Ouyang, D.; He, S.; Zhang, G.; Luo, M.; Guo, H.; Zhan, J.; Huang, Z. Efficient Multi-Scale Attention Module with Cross-Spatial Learning. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of the light path of common illumination structures. (a) Bright field and dark field forward lighting. (b) Scattering forward lighting. (c) Backward lighting.
Figure 1. Schematic diagram of the light path of common illumination structures. (a) Bright field and dark field forward lighting. (b) Scattering forward lighting. (c) Backward lighting.
Sensors 24 06717 g001
Figure 2. Typical defect samples under different illumination structures.
Figure 2. Typical defect samples under different illumination structures.
Sensors 24 06717 g002
Figure 3. Prototype dual-modal illumination image acquisition platform.
Figure 3. Prototype dual-modal illumination image acquisition platform.
Sensors 24 06717 g003
Figure 4. Forward lighting and backward lighting implementation.
Figure 4. Forward lighting and backward lighting implementation.
Sensors 24 06717 g004
Figure 5. Aircraft glass canopy samples.
Figure 5. Aircraft glass canopy samples.
Sensors 24 06717 g005
Figure 6. Typical images of four types of defects.
Figure 6. Typical images of four types of defects.
Sensors 24 06717 g006
Figure 7. RGB Channel Fusion method.
Figure 7. RGB Channel Fusion method.
Sensors 24 06717 g007
Figure 8. The architecture of the proposed ADMF-Net.
Figure 8. The architecture of the proposed ADMF-Net.
Sensors 24 06717 g008
Figure 9. A comparison of the two proposed methods. First row: RGB Channel Fusion. Second row: ADMF-Net.
Figure 9. A comparison of the two proposed methods. First row: RGB Channel Fusion. Second row: ADMF-Net.
Sensors 24 06717 g009
Figure 10. Normalized confusion matrix of single lighting methods and our fusion methods. (a) Forward lighting. (b) Backward lighting. (c) RGB Channel Fusion. (d) ADMF-Net.
Figure 10. Normalized confusion matrix of single lighting methods and our fusion methods. (a) Forward lighting. (b) Backward lighting. (c) RGB Channel Fusion. (d) ADMF-Net.
Sensors 24 06717 g010
Table 1. Defect quantity statistics.
Table 1. Defect quantity statistics.
IndexDefect Type NameCount
0contusion1104
1scratches1728
2crack232
3spot1720
Table 2. Hyperparametric configuration.
Table 2. Hyperparametric configuration.
Parameter NameParameter Value
Epoch300
Batch size16
Image size 640 × 640
Data augmentationMosaic-4
OptimizerSGD
Learning rate0.01
Momentum0.937
Weight decay0.0005
Table 3. Experimental results overview.
Table 3. Experimental results overview.
IndexDataset NameModelmAP50mAP50-95
1forward_onlyYOLOv8m0.9280.641
2backward_onlyYOLOv8m0.7900.538
3ag_composite_obbRGB Channel Fusion (ours) 10.9790.691
4ag_dual_obbADMF-Net (ours)0.9840.669
1 Data-level fusion, using YOLOv8m as the back end. Note: In all experiments, the model scale was set to m if optional. Bold values indicate the best performance for each metric.
Table 4. Detailed AP results for the four types of defect.
Table 4. Detailed AP results for the four types of defect.
IndexContusionScratchesCrackSpot
10.908/0.6490.900/0.5170.995/0.7430.908/0.652
20.951/0.6790.233/0.1180.995/0.6850.980/0.669
30.980/0.7160.957/0.5560.995/0.7370.984/0.754
40.995/0.6860.954/0.5200.995/0.7140.993/0.756
Note: For each data item, the values from left to right represent AP50 and AP50-95. Bold values indicate the best performance for each metric.
Table 5. Additional experimental results.
Table 5. Additional experimental results.
IndexDataset NameModelmAP50mAP50-95
5ag_composite_diff_obbRGB Channel Fusion 10.9770.684
6ag_dual_obbADMF-Net 20.9830.639
7ag_dual_obbADMF-Net 30.9760.668
8ag_dual_rectICAFusion [28] 40.9630.640
1 Using differential image as the third channel. 2 Without attention mechanism, using only Concat operation. 3 Using EMA attention [29]. 4 Using transformer cross-attention.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Z.; Yao, Y.; Wen, R.; Liu, Q. Dual-Modal Illumination System for Defect Detection of Aircraft Glass Canopies. Sensors 2024, 24, 6717. https://doi.org/10.3390/s24206717

AMA Style

Li Z, Yao Y, Wen R, Liu Q. Dual-Modal Illumination System for Defect Detection of Aircraft Glass Canopies. Sensors. 2024; 24(20):6717. https://doi.org/10.3390/s24206717

Chicago/Turabian Style

Li, Zijian, Yong Yao, Runyuan Wen, and Qiyang Liu. 2024. "Dual-Modal Illumination System for Defect Detection of Aircraft Glass Canopies" Sensors 24, no. 20: 6717. https://doi.org/10.3390/s24206717

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop