Next Article in Journal
Correlation between Muscular Activity and Vehicle Motion during Double Lane Change Driving
Next Article in Special Issue
SMS Scam Detection Application Based on Optical Character Recognition for Image Data Using Unsupervised and Deep Semi-Supervised Learning
Previous Article in Journal
The Impact of 1030 nm fs-Pulsed Laser on Enhanced Rayleigh Scattering in Optical Fibers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Underwater Crack Detection System Combining New Underwater Image-Processing Technology and an Improved YOLOv9 Network

1
School of Infrastructure Engineering, Dalian University of Technology, Dalian 116024, China
2
China Institute of Water Resources and Hydropower Research, Beijing 100048, China
3
China Yangtze Power Co., Ltd., Yichang 443000, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(18), 5981; https://doi.org/10.3390/s24185981
Submission received: 23 August 2024 / Revised: 6 September 2024 / Accepted: 11 September 2024 / Published: 15 September 2024

Abstract

:
Underwater cracks are difficult to detect and observe, posing a major challenge to crack detection. Currently, deep learning-based underwater crack detection methods rely heavily on a large number of crack images that are difficult to collect due to their complex and hazardous underwater environments. This study proposes a new underwater image-processing method that combines a novel white balance method and bilateral filtering denoising method to transform underwater crack images into high-quality above-water images with original crack features. Crack detection is then performed based on an improved YOLOv9-OREPA model. Through experiments, it is found that the new image-processing method proposed in this study significantly improves the evaluation indicators of new images, compared with other methods. The improved YOLOv9-OREPA also exhibits a significantly improved performance. The experimental results demonstrate that the method proposed in this study is a new approach suitable for detecting underwater cracks in dams and achieves the goal of transforming underwater images into above-water images.

1. Introduction

Dams, serving as water conservation and hydropower infrastructure, play an important role in flood control, water storage, irrigation, power generation, and ecological environment protection [1,2,3]. Currently, many dams experience structural aging and corrosion during long-term operation, which increases the risk of damage [4,5,6]. Regularly inspecting a dam, comprehensively evaluating and monitoring its safety status, and timely identifying and addressing potential problems are important for the smooth operation of a dam structure.
Damage such as cracks, spalling, erosion, cavities, and wear may occur during the long-term use of dams [7]. These defects significantly affect the safe and stable operation of dams. Among these disasters, cracks have received widespread attention, owing to their proneness, danger, and ability to trigger other disasters [8]. In recent years, deep learning (DL) and computer vision technologies have made significant progress and have led to significant breakthroughs in artificial intelligence and the development of automatic crack detection methods. Devices equipped with high-resolution visual sensors, such as industrial cameras and unmanned aerial vehicles (UAVs), are used to collect numerous images of dam surfaces. The collected images are processed, a deep learning model is trained to learn the damage features in the images, and a crack detection method based on deep learning is then employed [9,10,11,12,13,14,15,16,17,18,19]. However, because of the complex underwater environment, dam-surface crack detection technology cannot be directly applied to underwater crack detection. Underwater images suffer from color distortion, low contrast, a low signal-to-noise ratio, and a lack of detail. Hence, if deep learning technology is applied, it is difficult to obtain and collect datasets of underwater crack areas. An innovative underwater crack detection system is urgently needed to ensure the safe operation of dams.
Currently, research on underwater crack detection is limited. Huang et al. [8] proposed a method to generate underwater dam crack images using the CycleGAN model, a type of adversarial learning network, to convert above-water dam crack images through image-to-image translation. Then, they used the converted underwater dam crack images for dataset annotation and deep learning network training to solve the problem of insufficient datasets. Cao et al. [20] proposed a large-scale underwater crack detection method based on image stitching and segmentation that can adapt to complex underwater environments and perform well in different research fields. Li et al. [21] proposed a comprehensive, pixel-level, underwater structure, multi-defect instance segmentation, and quantization framework based on machine vision and deep learning for hydraulic tunnels. They also developed a video dataset of multiple defects in underwater tunnel lining structures for training and detection. Li et al. [22] proposed a real-time pixel-level automatic segmentation and quantification framework for underwater cracks in dams using a lightweight semantic segmentation network (i.e., LinkNet) and two-stage hybrid transfer learning (TL) for the detection and segmentation of underwater cracks.
Previous studies mostly trained and segmented deep-learning networks based on underwater crack datasets. However, obtaining underwater crack datasets is difficult and time-consuming. Converting underwater crack images into above-water crack images with the same crack characteristics is an effective crack detection method. This method can be used to detect above-water crack images generated after image processing using a more easily accessible water surface crack dataset, which can improve work efficiency. However, new image-processing technology is required to convert underwater images into water surface images.
There are three main differences between the underwater and above-water environments [23]:
(1)
The optical imaging principle of underwater images is different from that of underwater images. Figure 1a shows that compared with above-water images, the image received by the underwater imaging system is affected by the direct attenuation, forward scattering, and backscattering components [24].
(2)
Underwater images are likely to be affected by bubbles, sand grains, and other impurities, resulting in more noise than that observed in above-water images [25].
(3)
The attenuation characteristics of light in water are significantly different from those in air. Figure 1b shows that in an underwater environment, light of different wavelengths has different attenuation rates when propagating underwater [26].
Figure 1. Underwater imaging principle: (a) underwater light loss diagram; (b) underwater attenuation rate graph of different light rays.
Figure 1. Underwater imaging principle: (a) underwater light loss diagram; (b) underwater attenuation rate graph of different light rays.
Sensors 24 05981 g001
To convert underwater crack images into above-water images, targeted image processing is required to address the three aforementioned differences. Underwater images require image enhancement, color balance, and noise removal [27]. There have been several breakthroughs in the field of underwater image processing [28,29,30,31]. Xin et al. [28] proposed a method based on the affine shadow transform and adaptive histogram equalization to achieve uniform illumination in a processed image. Ma et al. [29] proposed a method based on an affine shadow transform that reduces the color deviation of the processed image and improves its clarity. Zhang et al. [30] proposed a method based on biological principles that adaptively uniformizes the global brightness of an image according to its overall brightness distribution without manual intervention. Wang et al. [31] proposed a dual information modulation network that effectively supplemented the network’s structural and global perception capabilities by utilizing the differences and complementarities between features of different scales, thereby further improving the clarity and contrast of the image.
Previous research has played a significant role in the processing and optimization of underwater images. However, the optimization effect on the crack feature areas of underwater images is insignificant. Perhaps because of the different fields used in the research, most studies have only focused on optimizing the overall quality of underwater images and have neglected the optimization and denoising of specific areas (crack regions). Therefore, a more suitable underwater image-processing method must be developed to optimize underwater images with crack features. This study proposes a new underwater image-processing technology combining a novel white balance algorithm and bilateral filtering method to improve the image quality of processed underwater crack images, compared with previous image-processing methods.
After obtaining water crack images with underwater crack features, a crack detection network is used to achieve crack region detection. With the popularity of target detection networks, such as the YOLO (You Only Look Once) and U-net networks, digital image detection technology has become a detection method with a lower cost and higher efficiency. Moreover, it is widely used in underwater crack detection in concrete dams. Currently, there are many studies on the improvement and innovation of crack detection networks. Li et al. [32] proposed a new IDP-YOLOv8 network based on the newly launched YOLOv9 network, which utilizes a parallel architecture, including an image denoising and enhancement module as well as an improved YOLOv9 object detection module that can significantly improve image quality. Wu et al. [33] proposed an integrated approach based on the convolutional neural network DeepLabv3+ for crack detection as well as a pixel-level crack quantification algorithm that achieved high accuracy in crack detection. Abbasi et al. [34] proposed a fog perception adaptive YOLO algorithm for target detection in foggy weather environments, and this algorithm achieved a reasonable improvement in terms of average accuracy. Luo et al. [35] used a deep convolutional neural network for drone detection. In the process of feature extraction in convolutional layers of deep learning neural networks, there are often issues with insufficient and imprecise feature extraction, which can lead to decreased model accuracy. The OREPA convolutional structure addresses this problem. By integrating residual connections, edge information, and pyramid attention mechanisms, the OREPA convolutional layer enhances the model’s ability to extract complex features, thereby resolving the shortcomings of traditional convolutional layers in feature extraction and significantly improving the model’s recognition accuracy. A detailed introduction to the OREPA convolutional layer will be provided in Section 3.
Previous studies based on deep-learning networks have made significant progress in object detection [36]. A new deep-learning network, YOLOv9 [37], was proposed and used for object detection and segmentation. There has been relatively little research on the improvement of the YOLOv9 network; therefore, further research is required on the YOLOv9 network.
Building on previous research, this study proposes a novel image-processing technology to convert underwater crack images to water surface crack images and then replace the original convolutional blocks in YOLOv9 with OREPA convolutional blocks to improve the ability of the YOLOv9 network to accurately detect underwater cracks. The contributions of this study are as follows:
(1)
This study proposes a novel white balance algorithm that improves the image quality of processed underwater crack images, compared with previous white balance algorithms.
(2)
This study used a bilateral filtering denoising method to further denoise underwater crack images, reduce the impact of noise while preserving underwater crack features, and convert underwater crack images to water surface crack images. The denoising effect of bilateral filtering is superior to that of other denoising methods and can better restore the crack features of underwater images.
(3)
The original convolutional blocks of YOLOv9 were replaced with OREPA convolutional blocks. Compared with all the original models of YOLOv9, the proposed model shows better crack detection results in terms of the loss rate and other indicators, which applies to the method proposed in this study.
The remainder of this paper is organized as follows. Section 2 introduces the proposed underwater image-processing technology and various evaluation indicators for underwater images. Section 3 introduces the combination of the improved YOLOv9 model and the OREPA convolutional blocks. Section 4 describes the relevant experimental plans and results. Section 5 provides the conclusions. Section 5 provides the conclusions. Section 6 introduces the future work to be done and some limitations of this method.

2. A Novel Underwater Image-Processing Technology

This section introduces a novel white balance algorithm and the principle of bilateral filtering and then introduces the meaning and expressions of various indicators used to evaluate image quality.

2.1. A Novel White Balance Algorithm

White balance, also known as color constancy, is the ability to maintain color consistency in a scene under different lighting conditions. In an underwater environment, owing to different light source conditions, the light reflected from the same object may differ, leading to the final camera assuming that the same object has different colors under different lighting conditions. To address this issue, a white balance adjustment function is introduced in the image to increase or decrease the color values of the red, green, and blue channels at each pixel point, according to the perceived light, color, and temperature information. In this manner, the white objects in the image can be presented as truly white under different lighting conditions, and the other colors can maintain relative accuracy.
Four types of underwater image white balance algorithms are commonly used [38]: the average white balance method, perfect reflection method, image analysis-based chromatic aberration detection and correction method, and gray world white balance method. We introduce a new white balance method based on a dynamic threshold algorithm [39] to achieve image color uniformity.
This method introduces a new color space, namely, the YUV color space, where Y represents luminance and U and V represent chrominance (hue and saturation). The Y-component is constructed from the RGB input signals by combining specific portions of RGB signals. Chrominance is defined by two aspects of color: hue and saturation, which are represented by Cr and Cb, respectively. Here, Cr reflects the difference between the red component of the RGB input signal and the luminance of the RGB signal, whereas Cb reflects the difference between the blue component of the RGB input signal and the luminance of the RGB signal.
This method determines a reference point by converting the RGB color space into the YCrCb color space for analysis. By analyzing the YCrCb coordinate space of the image, a region that is close to white and contains the reference point can be identified. A threshold is then set to designate certain points as reference points. Next, the input image is divided into 16 regions, and each region is transformed from the RGB color space to the YCrCb space. A threshold is then set to determine the reference point position in each region. In Figure 2, the thirteenth region is used as an example.
The specific algorithm process is as follows.
(1)
Transform the image from the RGB space into the YCrCb space.
Y = 0.257 R + 0.054 G + 0.098 B + 16 255
C b = 0.148 R 0.291 G + 0.439 B + 128 255
C r = 0.439 R 0.368 G 0.071 B + 128 255
(2)
Calculate the mean values of Cr and Cb for each region, where Mr, Mb, and N are the numbers of pixels in each region, and (i, j) denotes the reference point coordinates for each region.
M r = C r i , j   N   ,   M b = C b i , J N
D r = C r i ,   j M r   N   ,   D b = C b i ,   j M b   N  
(3)
Find a reference point that satisfies the conditions in each region. The judgment criteria are shown in Equation (6).
D r   = 0.01 ;   D b   0.01
(4)
After determining the reference point position, calculate the maximum brightness Ymax of the reference point.
(5)
Calculate the gain coefficients (Rgain, Ggain, and Bgain) of each RGB channel. In Equation (7), Ravgw, Gavgw, and Bavgw are the mean values of the reference points for the three channels in each region.
R g a i n = Y m a x R a v g w   ,   G g a i n   = Y m a x G a v g w   ,   B g a i n = Y m a x B a v g w
(6)
Calculate the gain coefficients (Rgain, Ggain, and Bgain) for each RGB channel. In Equation (7), Ravgw, Gavgw, and Bavgw are the mean values of the reference points for the three channels in each region.
(7)
The RGB channel value of the image is calculated using the gain coefficient. In Equation (8), R′, G, and B′ are the RGB channel values of the new image calculated by the gain coefficient in each region.
R   = R × R g a i n   ,   G = G × G g a i n   ,   B = B × B g a i n
(8)
Combine the newly obtained RGB channel (R′, G′, and B′) values to generate new underwater crack images. Figure 3 shows different underwater crack images generated by the four classic white balance algorithms and the white balance algorithm proposed in this study.

2.2. Bilateral Filtering Denoising

Image denoising has many common methods, such as median filtering, Gaussian filtering, and mean filtering. This paper uses the bilateral filtering denoising method to process crack feature images [40,41]. Bilateral filtering is a non-linear filtering method that combines the spatial proximity and pixel similarity of an image, taking into account both spatial information and grayscale similarity to achieve edge preservation and denoising. The advantage of the bilateral filter is that it can preserve edges. Usually, Gaussian filtering is used for denoising, which blurs the edges more and does not have a significant protection effect for high-frequency details. The reason why the bilateral filter can achieve smooth denoising while preserving the edges well is that the filter kernel is generated by two functions: the spatial domain kernel and the value domain kernel. In the spatial domain kernel, the bilateral filter follows Equation (9). The ωd term is a template weight determined by the Euclidean distance between pixel positions, q(i, j) is the coordinates of other coefficients in the template window, p(k, l) is the center coordinate point of the template window, and σd is the standard deviation of the Gaussian kernel function in the spatial domain and is used to control the weights of pixel positions.
ω d i ,   j ,   k ,   l = exp ( i k 2 j l 2 2 σ d 2 )  
Within the range kernel, bilateral filtering is as follows in Equation (10). Here, ωr is the template weight determined by the difference in pixel values, f(i, j) represents the pixel value of the image at point q(i, j), f(k, l) denotes the pixel value of the image at point p(k, l), and σr is the standard deviation of the Gaussian kernel function in the pixel value domain that is used to control the weights of pixel values.
ω r i ,   j ,   k ,   l = exp ( f i , j f k , l 2 σ r 2 )  
Here, ωd measures the distance between two points, with lower weights as the distance increases, and ωr measures the degree of similarity in pixel values between two points, with higher weights assigned to points that are more similar. Finally, multiplying the above two templates yields the template weights ω for the bilateral filter.
ω i ,   j ,   k ,   l = e xp ( i k 2 j l 2 2 σ d 2 ) ( f i , j f k , l 2 σ r 2 )  
The data equation for the bilateral filter is given in Equation (12):
g i ,   j = k l f k ,   l ω i ,   j , k , l k l ω i , j , k , l
This study uses median filtering, mean filtering, Gaussian filtering, and bilateral filtering to denoise the input image and then compares the values of various evaluation indicators of the processed image to find the most suitable denoising method. Figure 4 illustrates the specific process.

2.3. Image Evaluation Indicators

The main image metrics used in this study were the UIQM [23], SSIM, and PSNR [42]. UIQM includes three underwater image attribute metrics: the underwater image color metric (UICM), the underwater image clarity metric (UISM), and the underwater image contrast metric (UIConM). Below are the methods used to calculate various evaluation indicators. The solution equations for the UICM, UISM, and UIConM terms in Equation (13) can be found in the literature [23].
UIQM = 0.0282 × UICM + 0.2953 × UISM + 3.5753 × UIConM
Given a reference image f and a test image g, both of size M × N, the PSNR between f and g is defined as follows:
PSNR f , g = 10 log 10 255 2 MSE f , g
MSE f , g = 1 M N   i = 1 M j = 1 N ( f i j g i j ) 2
The PSNR value approaches infinity as the MSE [42] approaches zero, indicating that a higher PSNR value provides a higher image quality. On the other end of the scale, a small PSNR value implies large numerical differences between the images.
The SSIM is a quality metric used to measure the similarity between two images. This is considered to correlate with the quality perception of the human visual system. The SSIM is defined as follows:
SSIM f , g = l f , g c f , g s ( f , g )
l f , g = 2 μ f μ g + C 1 μ f 2 + μ g 2 + C 1 c f , g = 2 σ f σ g + C 2 σ f 2 + σ g 2 + C 2 S f , g = σ f g   + C 3 σ f σ g + C 3
Here, l(f, g) is used to measure the closeness of the average brightness (μf and μg) between two images, c(f, g) is used to measure the closeness of contrast (σf and σg) between two images, and s(f, g) is used to measure the correlation coefficient between the images f and g. After describing the UIQM, PSNR, and SSIM, this study compares various underwater image-processing methods using these three indicators and thus determines the most suitable underwater image-processing method.

2.4. A New Underwater Image-Processing System

This study proposes a novel underwater image-processing technology combined with bilateral filtering and a novel white balance algorithm for underwater image denoising to achieve image enhancement and compensation. This method is used to convert underwater crack images into above-water crack images. Figure 5 shows a detailed structural diagram. Box 1 in Figure 5 represents the five white balance algorithms from top to bottom: the average white balance method, perfect reflection method, proposed white balance method, chromatic aberration detection and correction method based on image analysis, and gray world method. Box 2 in Figure 5 shows the four filtering and denoising methods from top to bottom: mean, median, bilateral, and Gaussian filtering. The results of the experiment described in Section 4.1 indicate that the proposed white balance method and bilateral filtering denoising method have significant advantages over other methods in processing underwater images, and can convert underwater crack images into above-water crack images without damaging the crack characteristics.

3. Improved Crack Detection Network Based on YOLOv9

3.1. Infrastructure of YOLOv9

YOLOv9 is a recently launched object detection network [37]. Figure 6 shows a structural diagram of the YOLOv9 network. The basic architecture of YOLOv9 is as follows. Many existing deep learning networks overlook the fact that a large amount of information is lost when the input data undergo feature extraction and spatial transformation, layer by layer. YOLOv9 delves into important data loss issues during data transmission in deep networks, namely, information bottlenecks and reversible functions. Moreover, it proposes the concept of programmable gradient information (PGI) [43] to address the various changes required for deep networks to achieve multiple objectives. PGI [44] can provide complete input information for the target task to calculate the objective function and thereby obtain reliable gradient information to update the network weights. In addition, a new lightweight network structure based on gradient path planning, namely, the generalized efficient layer aggregation network (GELAN) [45], was designed. The architecture of GELAN confirmed that the PGI achieved outstanding results with lightweight models.
The concept of PGI in YOLOv9 and the introduction of GELAN can help solve the feedforward problem of input data caused by information bottlenecks and reversible functions as well as reduce training losses.
The PGI mainly consists of three parts: the main branch, the auxiliary reversible branch, and multilevel auxiliary information. During the inference process, only the main branch was used without additional inference costs, whereas the other two branches were mainly used to solve information bottleneck problems. Figure 7 shows the overall architecture of the PGI.
The GELAN network combines the design of CSPNeT [46] and ELAN [45] by adopting the concepts of segmentation and reassembly from CSPNeT and introducing the hierarchical convolution processing from ELAN in each section. The design of GELAN makes the overall training model lightweight with a higher inference efficiency and accuracy. Figure 8 shows the structural diagrams of the three network architectures.

3.2. Improved YOLOv9 Network Combined with an OREPA Module

OREPA (Online Convolutional Re-parameterization) reduces the cost and complexity of training deep learning models via online convolution reparameterization. This method includes two stages. First, a special linear scaling layer is used to optimize the performance of the online block. Second, the training overhead can be reduced by compressing complex training modules into a single convolution. The OREPA process includes three main steps: removal, scaling, and normalization. After a series of steps, the OREPA method effectively simplifies the complexity of the model and optimizes the training efficiency during the training phase. Figure 9 illustrates the specific steps of the OREPA process.
A key component of OREPA is the linear scaling layer, which aims to maintain complexity management during training while improving the flexibility and adaptability of the model in learning features. The appropriate scaling of the weights helps stabilize the entire training process and enables the model to converge as quickly as possible. The linear scaling layer consists of the following four parts: a frequency prior filter, linear depth-separable convolution, 1 × 1 convolution with heavy parameters, and linear deep stem cells [47]. Figure 10 shows the four components of the OREPA.
During the model training process, the OREPA framework uses module compression to reduce memory usage and computational costs. The design of OREPA allows multiple convolutional layers and batch normalization layers to be merged into one convolutional layer via reparameterization, thereby reducing the training complexity and memory requirements. Figure 11 shows a comparison of the different convolutional layers during training.
The improved YOLOv9-OREPA model proposed in this study uses OREPA convolution layers instead of the original YOLOv9 network’s convolutional layers to reduce memory usage, but computational costs are incurred during model training to improve training accuracy. Figure 12 presents a flowchart of the YOLOv9-OREPA model. Figure 6 and Figure 12 together show that the difference from the original YOLOv9 structure lies in the replacement of the convolutional structure.

3.3. Loss Function and Evaluation Metrics of YOLOv9-OREPA

The loss function of YOLOv9-OREPA consists of two parts: classification loss (Losscla) and regression loss (Lossbox), as follows.
L o s s = L o s s c l a + L o s s b o x
The classification loss adopts binary cross-entropy (BCE), which is used to determine whether the detection target belongs to a particular category, and accordingly outputs the following network confidence result:
L o s s c l a B C E = 1 N i y i . log p i + 1 y i . log ( 1 p i )
where yi represents the target category and pi represents the probability that the predicted result is yi. In this article, there is only one classification category (cracks), so the classification loss can be ignored. Regression loss is caused by the difference in coordinates and dimensions between the predicted and target boxes. In the regression tasks, the degree of regression can be measured using the ratio of the predicted box to the target box. The IoU ratio (IoU) is a descriptive method expressed via the following formula, where A is the prediction and B is the target, as follows:
I o U = A + B A B A B
This study uses CIoU as the regression loss function, adds the aspect ratio to the basic IoU, and considers three factors, namely, the overlap area, center point distance, and aspect ratio of the box, to improve the regression accuracy [48]. Figure 13 illustrates the calculation principle of the loss function.
L C I o U = 1 I o U + ρ 2 b , b g t c 2 + α v
c 2 = ( w c ) 2 + ( h 2 ) 2
α = v 1 I o U + v 2
v = 4 π 2 ( a r c t a n w g t h g t a r c t a n w h ) 2
The final loss function is as follows:
L o s s = L o s s c l a BCE + L o s s b o x C I o U
The performance of the improved YOLOV9-OREPA model was evaluated using object detection evaluation metrics, including the accuracy P, recall R, F1-value F1, and average category accuracy mAP. The specific calculation method is as follows:
P = T P T P + F P
R = T P T P + F N
F 1 = 2 × P × R P + R
A P = 0 1 P R d R
m A P = q = 1 Q A P ( q ) Q
where TP is the number of correctly detected samples, FP is the number of incorrectly detected samples, FN is the number of undetected defect samples, F1 is the harmonic average of the accuracy P and recall rate R, the single-category AP is the area of the closed graph formed by the curve of the accuracy P and recall rate R, the coordinate axis Q represents the category, and the average category average precision mAP is the average single-category AP of each of the defect categories.

4. Case Study

4.1. Underwater Image Optimization Experiment

To obtain underwater images with the crack features required for the experiment, an underwater robot was used to capture images in a laboratory swimming pool. Underwater images were captured under low-light conditions to verify the superiority and reliability of the proposed method. Figure 14 shows the experimental setup used. Figure 15 shows the experimental process.
After capturing multiple images, five images in Figure 16 with obvious crack features were selected for use in the experiment. Subsequently, a white balance operation was performed on the five images. The initial image in Figure 16a was considered as an example. Five methods were used: the average white balance method (first method), perfect reflection method (second method), our method (third method), chromatic aberration detection and color correction method based on image analysis (fourth method), and gray world method (fifth method). In Figure 17, five processed images were obtained, and the UIQM, PSNR, and SSIM of each image were compared to determine the best white balance method.
The image metrics obtained from the first image using the different white-balance methods were different, as shown in Table 1.
Table 1 shows that the proposed method exhibits good processing performance in terms of the UIQM, PSNR, and SSIM evaluation indicators. Our proposed method increases the UICM index of the initial underwater image from 14.897 to 18.395, increases the UISM index from 0.136 to 0.140, decreases the UIConM index from −0.119 to −0.118, and increases the UIQM index from 0.032 to 0.136. The PSNR index is the highest for the proposed method, reaching 39.783, and the SSIM index is also the highest for the proposed method, reaching 0.988. Hence, this method not only retains the crack features of the original image but also enhances and compensates the original image. Figure 18 shows the processing and data tables for the remaining images.
Table 2 shows that in the processing of the five underwater images, our method performs better than the other four methods in terms of three image evaluation indicators (i.e., UIQM, SSIM, and PSNR). Especially in terms of the UISM evaluation index, our proposed white balance method has the highest gain in image quality, improving Figure 16a–e from the original (0.032, −0.063, 0.082, 0.094, 0.069) to (0.136, 0.051, 0.084, 0.112, 0.112). Therefore, it can be concluded that among the many white-balance processing methods, our method has the best processing effect on underwater images. The five images processed using our white balance method were the output results of the first step of the proposed method. Next, the five processed images were denoised, and these images were used as the input images for the second step.
The underwater images processed by our white balance method are the output images of the first step and the input images of the second-step denoising operation. Considering Figure 19a as an example, Figure 20 presents a comparison of the denoising processes. First, Figure 19a was used as the input image, and four different denoising methods (i.e., bilateral, median, mean, and Gaussian filtering) were used to denoise the image. The UIQM, PSNR, and SSIM of the denoised images were compared to determine the most suitable denoising method.
The other input images in Figure 19 also underwent various denoising operations. In the denoising experiments of mean filtering and Gaussian filtering, different convolution kernel sizes were added to denoise the image, and the various evaluation indicators of the processed images were compared to determine the denoising method that best suited the underwater image. Table 3 shows the results.
Table 3 shows that the UIQM and PSNR image metrics of the five images used in the experiment were the best among all denoising methods after bilateral filtering denoising. According to the definition of the SSIM, image denoising affects image similarity. The bilateral filtering method impacts the fourth decimal place, satisfying the accuracy requirements. In addition, the results of the Gaussian and mean filtering experiments indicate that the larger the convolution kernel used, the larger the UIQM of the image, whereas the PSNR and SSIM metrics are smaller. This is because the larger the convolution kernel used for denoising, the less noise in the image, the smoother the image itself, and the larger the UIQM metric. However, the denoising operation may cause a loss of crack features. Therefore, the PSNR and SSIM metrics are slightly smaller than those of the input image.
The new image-processing method proposed in this study combines a novel white balance method and a bilateral filtering denoising method and optimizes three of the image metrics used in the experimentation. Hence, it is suitable for underwater image processing and makes it possible to convert underwater crack images into above-water crack images that preserve crack features. Table 4 summarizes the overall improvements in the image metrics of the five experimental images processed using the proposed method. The UIQM index of the image processed by this method is significantly improved, whereas the PSNR and SSIM indexes also reach an acceptable level, indicating that this method not only retains the characteristics of the original image but also confers a significant improvement in image quality, thereby exceeding the performance of than other methods. In summary, the method proposed in this study is a new image-processing method suitable for underwater images.

4.2. Improved YOLOv9-OREPA Network Performance Evaluation

This study used 938 images with crack information as the dataset and divided the image dataset into training, validation, and test sets during the training process. The test set was not used for training. The training set consisted of 750 images, whereas the validation and test sets consisted of 96 images each, at a ratio of 8:1:1. Table 5 lists the basic operating parameters of this model.
To verify that the proposed YOLOv9-OREPA has improved performance compared with the original YOLOv9, this study compares all the original YOLOv9 models (i.e., YOlOv9, YOlOv9-e, YOlOv9-c, YOlOv9-m, YOlOv9-s, and YOlOv9-t) with the YOLOv9-OREPA model. Table 6 shows the results.
Table 6 shows that the performance of YOLOv9-OREPA proposed in this study exceeds those of various existing YOLOv9 models. Compared with the initial YOLOv9, R increased by 5.5%, F1 increased by 2%, mAP0.5 increased by 0.9%, and mAP0.5:0.95 increased by 2.1%. This proves that OREPA, a new convolution mode, is highly suitable for improving YOLOv9.
Figure 21a,b show that after 200 rounds of training, YOLOv9-OREPA had the smallest training and validation process losses, with a training process loss of 1.39% and a validation process loss of 1.10%. In addition, Figure 21c,d show that YOLOv9-OREPA has a better convergence trend than do the other YOLOv9 models, with relatively stable fluctuations in the P and R curves and a continuous upward trend. Furthermore, Figure 21e,f show that YOLOv9-OREPA is also superior to other ordinary models in terms of the average category accuracy mAP. YOLOv9-OREPA had the highest mAP0.5 (74.8%) and mAP0.5:0.95 (44.1%).
The precision–recall curves for the accuracy and summoning rate visually demonstrate the performance of the model. Plotting the precision–recall curves of the different models in Figure 22a reveals that the closed graph area enclosed by the YOLOv9-OREPA curve and coordinate axis is the largest, which can intuitively demonstrate the higher performance of the improved YOLOv9-OREPA. Moreover, Figure 22b shows that the F1 curve of YOLOv9-OREPA has the largest area with the coordinate axis; when the confidence interval is 0.357, the F1 value is the highest, reaching 0.73.
Figure 23 shows the crack detection results of the different original YOLOv9 models and the YOLOv9-OREPA model proposed in this study after training. The results show that the base YOLOv9 model exhibits reinspection and missed detection in the crack detection process, while the improved YOLOv9-OREPA model does not have reinspection or missed detection issues indicating that the improved YOLOv9-OREPA model enhances the accuracy of crack detection.
The YOLOv9-OREPA model can achieve the goal of crack detection, and its performance is superior to those of other YOLOv9 models.
Figure 24 shows the crack detection results of the original underwater image after combining the new underwater image-processing method proposed in this study with the improved YOLOv9-OREPA algorithm. After image processing, the overall indicators of the image are improved while ensuring the crack features. The proposed YOLOv9-OREPA network could fully perform crack detection. The images processed using the underwater image-processing technology proposed in this study can be accurately identified by the improved YOLOv9 OREPA network trained on underwater crack images to thereby complete the conversion from underwater to above-water crack images.
Figure 25 shows that the proposed method is suitable for crack detection in not only underwater images under low-light conditions but also underwater and above-water images with large color differences, which proves the superiority and universality of this method in crack detection.

5. Conclusions

This study proposes a novel underwater image-processing method that combines a new white balance technique with bilateral filtering denoising. The metrics of the processed underwater crack images were evaluated and compared with those of other methods, demonstrating superior performance. These results further demonstrate that the proposed underwater image-processing method can convert underwater crack images into high-quality images that retain the original crack features to thereby transform underwater crack images into above-water crack images. Subsequently, crack detection was conducted based on the improved YOLOv9-OREPA model, and comparisons were made between the improved YOLOv9-OREPA model and other basic YOLOv9 models. The experimental results indicate that YOLOv9-OREPA significantly improves the detection metrics of YOLOv9. In summary, the following conclusions were drawn:
(1)
In a comparative experiment of white balance algorithms, the proposed novel method was superior to the other four tested white balance methods in terms of the image UIQM, PSNR, and SSIM indices, which proves the superiority of our proposed method.
(2)
In the comparative experiment of the denoising effect, the bilateral filtering method was superior to the other denoising methods in the evaluation of various image indicators, thus proving the superiority of bilateral filtering in denoising underwater crack images. The innovative image-processing technique proposed in this study has significant advantages and can greatly improve various underwater image metrics to transform underwater images into above-water images.
(3)
The results of a comparative experiment on crack detection between YOLOv9-OREPA and the YOLOv9 series show that YOLOv9-OREPA improves various training indicators of crack detection and is a deep-learning network that is more suitable for crack detection.
In summary, this study innovates from the two aspects of underwater image processing and crack detection network optimization. The system proposed in this study is better aligned with the requirements of underwater crack detection, improves the image quality of underwater crack images and the accuracy of crack detection, and converts underwater crack images into above-water crack images. This study offers a new perspective on underwater crack detection by transforming underwater images into above-water images, thereby addressing the issue of insufficient datasets that can lead to imprecise underwater crack detection.

6. Future Works

The input images used in this study were all collected from a laboratory water tank, which differs from the crack areas in real reservoir dam environments. The water in the laboratory tank is relatively clear, which also differs significantly from the water quality conditions in actual environments. In the future, we plan to collect crack feature images in real environments using underwater robots at reservoirs and apply the image enhancement and crack detection methods proposed in this paper to validate their accuracy and superiority. Additionally, we will create a murky water environment in the laboratory to extract crack images and verify the general applicability of the method proposed in this paper.

Author Contributions

Conceptualization, X.H. and F.K.; methodology, X.H.; software, X.H. and C.L.; validation, X.H., F.K., C.L. and X.L.; formal analysis, X.H.; investigation, F.K.; resources, writing—original draft preparation, X.H. and F.K.; visualization, X.H.; supervision, F.K.; project administration, F.K.; funding acquisition, F.K. contributed to the work equally and should be regarded as co-first authors. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key R & D Program of China (2022YFB4703404), the National Natural Science Foundation of China (52079022, 51979027), Guangxi Key R & D Program (GuiKe AB 24010003), the Fundamental Research Funds for the Central Universities (DUT21TD106).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

Author Xinyu Li was employed by the company China Yangtze Power Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Bai, R.; Liu, X.; Liu, X.; Liu, L.; Wang, J.; Liao, S.; Zhu, A.; Li, Z. The development of biodiversity conservation measures in China’s hydro projects: A review. Environ. Int. 2017, 108, 285–298. [Google Scholar] [CrossRef] [PubMed]
  2. Ma, H.; Chi, F. Technical Progress on Researches for the Safety of High Concrete-Faced Rockfill Dams. Engineering 2016, 2, 332–339. [Google Scholar] [CrossRef]
  3. Li, M.; Yuan, J.; Ren, Q.; Luo, Q.; Fu, J.; Li, Z. CNN-Transformer hybrid network for concrete dam crack patrol inspection. Autom. Constr. 2024, 163, 105440. [Google Scholar] [CrossRef]
  4. Huang, B.; Kang, F.; Li, J.; Wang, F. Displacement prediction model for high arch dams using long short-term memory based encoder-decoder with dual-stage attention considering measured dam temperature. Eng. Struct. 2023, 280, 115686. [Google Scholar] [CrossRef]
  5. Kang, F.; Shi, Y.; Li, J.; Zhang, H. Impacts of leader anger expressions on newcomer proactive career behavior and work alienation in the construction industry. Eng. Constr. Archit. Manag. 2024. ahead of printing. [Google Scholar] [CrossRef]
  6. Guo, W.; Lu, W.; Kang, F.; Zhang, L. How to Foster Relational Behavior in Construction Projects: Direct and Mediating Effects of Contractual Complexity and Regulatory Focus. J. Constr. Eng. Manag. 2024, 150, 4. [Google Scholar] [CrossRef]
  7. Wu, Z.; Li, J.; Gu, C.; Su, H. Review on hidden trouble detection and health diagnosis of hydraulic concrete structures. Sci. China Ser. E Technol. Sci. 2007, 50 (Suppl. S1), 34–50. [Google Scholar] [CrossRef]
  8. Huang, B.; Kang, F.; Li, X.; Zhu, S. Underwater dam crack image generation based on unsupervised image-to-image translation. Autom. Constr. 2024, 163, 105430. [Google Scholar] [CrossRef]
  9. Huang, B.; Kang, F.; Tang, Y. A real-time detection method for concrete dam cracks based on an object detection algorithm. Qinghua Daxue Xuebao/J. Tsinghua Univ. 2023, 63, 1078–1086. [Google Scholar] [CrossRef]
  10. Ren, Q.; Li, M.; Shen, Y.; Zhang, Y.; Bai, S. Pixel-level shape segmentation and feature quantification of hydraulic concrete cracks based on digital images. Shuili Fadian Xuebao/J. Hydroelectr. Eng. 2021, 40, 234–246. [Google Scholar] [CrossRef]
  11. Yu, Z.; Shen, Y.; Shen, C. A real-time detection approach for bridge cracks based on YOLOv4-FPM. Autom. Constr. 2021, 122, 103514. [Google Scholar] [CrossRef]
  12. Beckman, G.H.; Polyzois, D.; Cha, Y.J. Deep learning-based automatic volumetric damage quantification using depth camera. Autom. Constr. 2019, 99, 114–124. [Google Scholar] [CrossRef]
  13. Li, J.; Liu, T.; Wang, X.; Yu, J. Automated asphalt pavement damage rate detection based on optimized GA-CNN. Autom. Constr. 2022, 136, 104180. [Google Scholar] [CrossRef]
  14. Courtial, A.; Touya, G.; Zhang, X. Constraint-Based Evaluation of Map Images Generalized by Deep Learning. J. Geovisualization Spat. Anal. 2022, 6, 13. [Google Scholar] [CrossRef]
  15. Tan, L.; Hu, X.; Tang, T.; Yuan, D. A lightweight metro tunnel water leakage identification algorithm via machine vision. Eng. Fail. Anal. 2023, 150, 107327. [Google Scholar] [CrossRef]
  16. Tan, L.; Chen, X.; Hu, X.; Tang, T. In DMDSNet: A Computer Vision-based Dual Multi-Task Model for Tunnel Bolt Detection and Corrosion Segmentation. In Proceedings of the IEEE Conference on Intelligent Transportation Systems (ITSC), Bizkaia, Spain, 24–28 September 2023; pp. 4827–4833. [Google Scholar]
  17. Hu, X.; Zhang, X.; Wang, Z.; Chen, Y.; Xia, J.; Du, Y.; Li, Y. In Railway Switch Machine Fault Diagnosis Considering Sensor Abnormality Scenarios. In Proceedings of the IEEE Conference on Intelligent Transportation Systems, Proceedings (ITSC), Bizkaia, Spain, 24–28 September 2023; pp. 4834–4839. [Google Scholar]
  18. Ren, Q.Y.; Wang, Y.H.; Xu, J. A DL method to detect multi-type hidden objects in tunnel linings using a comprehensive GPR dataset. Measurement 2024, 238, 115379. [Google Scholar] [CrossRef]
  19. Chen, D.; Kang, F.; Li, J.; Zhu, S.; Liang, X. Enhancement of underwater dam crack images using multi-feature fusion. Autom. Constr. 2024, 167, 105727. [Google Scholar] [CrossRef]
  20. Cao, W.; Li, J. Detecting large-scale underwater cracks based on remote operated vehicle and graph convolutional neural network. Front. Struct. Civ. Eng. 2022, 16, 1378–1396. [Google Scholar] [CrossRef]
  21. Li, Y.; Bao, T.; Huang, X.; Wang, R.; Shu, X.; Xu, B.; Tu, J.; Zhou, Y.; Zhang, K. An integrated underwater structural multi-defects automatic identification and quantification framework for hydraulic tunnel via machine vision and deep learning. Struct. Health Monit. 2023, 22, 2360–2383. [Google Scholar] [CrossRef]
  22. Li, Y.; Bao, T.; Huang, X.; Chen, H.; Xu, B.; Shu, X.; Zhou, Y.; Cao, Q.; Tu, J.; Wang, R.; et al. Underwater crack pixel-wise identification and quantification for dams via lightweight semantic segmentation and transfer learning. Autom. Constr. 2022, 144, 104600. [Google Scholar] [CrossRef]
  23. Panetta, K.; Gao, C.; Agaian, S. Human-visual-system-inspired underwater image quality measures. IEEE J. Ocean Eng. 2016, 41, 541–551. [Google Scholar] [CrossRef]
  24. Corchs, S.; Schettini, R. Underwater image processing: State of the art of restoration and image enhancement methods. Eurasip J. Adv. Signal Process. 2010, 2010, 746054. [Google Scholar] [CrossRef]
  25. Li, G.; Bu, W.; Yang, H. Research on noise reduction method for ship radiate noise based on secondary decomposition. Ocean Eng. 2023, 268, 113412. [Google Scholar] [CrossRef]
  26. Li, Y.; Huang, Y.; Zou, Z.; Yu, Q.; Zhang, Z.; Sun, Q. Multi-AUV underwater static target search method based on consensus-based bundle algorithm and improved Glasius bio-inspired neural network. Inf. Sci. 2024, 673, 120684. [Google Scholar] [CrossRef]
  27. Chen, D.; Huang, B.; Kang, F. A Review of Detection Technologies for Underwater Cracks on Concrete Dam Surfaces. Appl. Sci. 2023, 13, 3564. [Google Scholar] [CrossRef]
  28. Xin, G.; Fan, X.; Shi, P.; Luo, C.; Ni, J.; Cao, Y. A fine extraction algorithm for image-based surface cracks in underwater dams. Meas. Sci. Technol. 2023, 34, 035402. [Google Scholar] [CrossRef]
  29. Ma, Y.; Wu, Y.; Li, Q.; Zhou, Y.; Yu, D. ROV-based binocular vision system for underwater structure crack detection and width measurement. Multimed. Tools Appl. 2023, 82, 20899–20923. [Google Scholar] [CrossRef]
  30. Zhang, H.; Lee, S. Advancing the Robotic Vision Revolution: Development and Evaluation of a Bionic Binocular System for Enhanced Robotic Vision. Biomimetics 2024, 9, 371. [Google Scholar] [CrossRef]
  31. Wang, L.; Li, X.; Li, K.; Mu, Y.; Zhang, M.; Yue, Z. Underwater image restoration based on dual information modulation network. Sci. Rep. 2024, 14, 5416. [Google Scholar] [CrossRef]
  32. Li, J.; Feng, Y.; Shao, Y.; Liu, F. IDP-YOLOV9: Improvement of Object Detection Model in Severe Weather Scenarios from Drone Perspective. Appl. Sci. 2024, 14, 5277. [Google Scholar] [CrossRef]
  33. Ji, A.; Xue, X.; Wang, Y.; Luo, X.; Xue, W. An integrated approach to automatic pixel-level crack detection and quantification of asphalt pavement. Autom. Constr. 2020, 114, 103176. [Google Scholar] [CrossRef]
  34. Abbasi, H.; Amini, M.; Yu, F.R. Fog-Aware Adaptive YOLO for Object Detection in Adverse Weather, 2023. In Proceedings of the IEEE Sensors Applications Symposium, SAS 2023, Ottawa, ON, Canada, 18–20 July 2023. [Google Scholar] [CrossRef]
  35. Luo, K.; Luo, R.; Zhou, Y. UAV detection based on rainy environment. In Proceedings of the IMCEC 2021—IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control, Chongqing, China, 18–20 June 2021; pp. 1207–1210. [Google Scholar] [CrossRef]
  36. Chen, D.; Kang, F.; Chen, J.; Zhu, S.; Li, H. Effect of light source wavelength on surface defect imaging in deep-water concrete dams. NDT. E Int. 2024, 147, 103198. [Google Scholar] [CrossRef]
  37. Wang, C.; Yeh, I.H.; Liao, H.Y. Yolov9: Learning what you want to learn using programmable gradient information. arXiv 2024, arXiv:2402.13616. [Google Scholar]
  38. Muniraj, M.; Dhandapani, V. Underwater image enhancement by combining color constancy and dehazing based on depth estimation. Neurocomputing 2021, 460, 211–230. [Google Scholar] [CrossRef]
  39. Weng, C.C.; Chen, H.; Fuh, C.S. A novel automatic white balance method for digital still cameras. In Proceedings of the-IEEE International Symposium on Circuits and Systems, Kobe, Japan, 23–25 May 2005; pp. 3801–3804. [Google Scholar] [CrossRef]
  40. Tomasi, C.; Manduchi, R. Bilateral filtering for gray and color images. In Proceedings of the IEEE International Conference on Computer Vision, Bombay, India, 4–7 January 1998; pp. 839–846. [Google Scholar]
  41. Zhao, Y.; Wang, G.; Yang, J.; Li, T.; Li, Z. AU3-GAN: A Method for Extracting Roads from Historical Maps Based on an Attention Generative Adversarial Network. J. Geovisualization Spat. Anal. 2024, 8, 26. [Google Scholar] [CrossRef]
  42. Horé, A.; Ziou, D. In Image quality metrics: PSNR vs. SSIM. In Proceedings of the International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar] [CrossRef]
  43. Chien, C.T.; Ju, R.Y.; Chou, K.Y.; Chiang, J.S. YOLOv9 for fracture detection in pediatric wrist trauma X-ray images. Electron. Lett. 2024, 60, e13248. [Google Scholar] [CrossRef]
  44. Urrea, C.; Garcia-Garcia, Y.; Kern, J. Improving Surgical Scene Semantic Segmentation through a Deep Learning Architecture with Attention to Class Imbalance. Biomedicines 2024, 12, 1309. [Google Scholar] [CrossRef]
  45. Wang, C.Y.; Liao, H.Y.M.; Yeh, I.H. Designing Network Design Strategies Through Gradient Path Analysis. J. Inf. Sci. Eng. 2023, 39, 975–995. [Google Scholar] [CrossRef]
  46. Wang, C.Y.; Mark Liao, H.Y.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H. CSPNet: A new backbone that can enhance learning capability of CNN, IEEE Computer Society. In Proceedings of the Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 1571–1580. [Google Scholar] [CrossRef]
  47. Xie, T.; Yin, M.; Zhu, X.; Sun, J.; Meng, C.; Bei, S. A Fast and Robust Lane Detection via Online Re-Parameterization and Hybrid Attention. Sensors 2023, 23, 8285. [Google Scholar] [CrossRef]
  48. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI 2020–34th AAAI on Artificial Intelligence, New York City, NY, USA, 7–12 February 2020; pp. 12993–13000. [Google Scholar]
Figure 2. Principle of the new white balance method.
Figure 2. Principle of the new white balance method.
Sensors 24 05981 g002
Figure 3. Five white balance algorithms: (a) mean white balance; (b) perfect reflection; (c) ours; (d) color cast; (e) gray world.
Figure 3. Five white balance algorithms: (a) mean white balance; (b) perfect reflection; (c) ours; (d) color cast; (e) gray world.
Sensors 24 05981 g003
Figure 4. Four denoising methods: (a) bilateral filtering; (b) median filtering; (c) mean filtering; (d) Gaussian filtering.
Figure 4. Four denoising methods: (a) bilateral filtering; (b) median filtering; (c) mean filtering; (d) Gaussian filtering.
Sensors 24 05981 g004
Figure 5. New underwater image-processing workflow.
Figure 5. New underwater image-processing workflow.
Sensors 24 05981 g005
Figure 6. Infrastructure diagram of YOLOv9.
Figure 6. Infrastructure diagram of YOLOv9.
Sensors 24 05981 g006
Figure 7. Overall architecture of PGI.
Figure 7. Overall architecture of PGI.
Sensors 24 05981 g007
Figure 8. Structure diagrams of three network architectures: (a) CSPNeT; (b) GELAN; (c) ELAN.
Figure 8. Structure diagrams of three network architectures: (a) CSPNeT; (b) GELAN; (c) ELAN.
Sensors 24 05981 g008
Figure 9. Specific steps of the OREPA process.
Figure 9. Specific steps of the OREPA process.
Sensors 24 05981 g009
Figure 10. Four components proposed in OREPA: (a) frequency prior filter; (b) linear depth separable convolution; (c) 1 × 1 convolution with heavy parameters; (d) linear deep stem cells.
Figure 10. Four components proposed in OREPA: (a) frequency prior filter; (b) linear depth separable convolution; (c) 1 × 1 convolution with heavy parameters; (d) linear deep stem cells.
Sensors 24 05981 g010
Figure 11. Comparison of different convolutional layers during training: (a) standard convolutional layer (without intermediate parameterization); (b) typical reparameterization; (c) OREPA.
Figure 11. Comparison of different convolutional layers during training: (a) standard convolutional layer (without intermediate parameterization); (b) typical reparameterization; (c) OREPA.
Sensors 24 05981 g011
Figure 12. Infrastructure diagram of the YOLOv9-OREPA model.
Figure 12. Infrastructure diagram of the YOLOv9-OREPA model.
Sensors 24 05981 g012
Figure 13. Calculation principle of the loss function.
Figure 13. Calculation principle of the loss function.
Sensors 24 05981 g013
Figure 14. Experimental facilities: (a) laboratory pool; (b) crack defect wall; (c) underwater robots used.
Figure 14. Experimental facilities: (a) laboratory pool; (b) crack defect wall; (c) underwater robots used.
Sensors 24 05981 g014
Figure 15. Experimental process image.
Figure 15. Experimental process image.
Sensors 24 05981 g015
Figure 16. Five underwater images with crack features under low light conditions.
Figure 16. Five underwater images with crack features under low light conditions.
Sensors 24 05981 g016
Figure 17. Image white balance operations: (a) original image; (b) mean white balance; (c) perfect reflection; (d) ours; (e) color cast; (f) gray world.
Figure 17. Image white balance operations: (a) original image; (b) mean white balance; (c) perfect reflection; (d) ours; (e) color cast; (f) gray world.
Sensors 24 05981 g017
Figure 18. The white balance processing results of the remaining four original images: (a) processing results of Figure 16b; (b) processing results of Figure 16c; (c) processing results of Figure 16d; (d) processing results of Figure 16e.
Figure 18. The white balance processing results of the remaining four original images: (a) processing results of Figure 16b; (b) processing results of Figure 16c; (c) processing results of Figure 16d; (d) processing results of Figure 16e.
Sensors 24 05981 g018
Figure 19. Images processed via our proposed method: (a) original picture from Figure 16a; (b) original picture from Figure 16b; (c) original picture from Figure 16c; (d) original picture from Figure 16d; (e) original picture from Figure 16e.
Figure 19. Images processed via our proposed method: (a) original picture from Figure 16a; (b) original picture from Figure 16b; (c) original picture from Figure 16c; (d) original picture from Figure 16d; (e) original picture from Figure 16e.
Sensors 24 05981 g019
Figure 20. Image denoising process diagram: (a) input image; (b) bilateral filtering; (c) median filtering; (d) mean filtering (1 × 1 convolutional kernel); (e) mean filtering (3 × 3 convolutional kernel); (f) mean filtering (5 × 5 convolutional kernel); (g) mean filtering (7 × 7 convolutional kernel); (h) Gaussian filtering (1 × 1 convolutional kernel); (i) Gaussian filtering (3 × 3 convolutional kernel); (j) Gaussian filtering (5 × 5 convolutional kernel); (k) Gaussian filtering (7 × 7 convolutional kernel).
Figure 20. Image denoising process diagram: (a) input image; (b) bilateral filtering; (c) median filtering; (d) mean filtering (1 × 1 convolutional kernel); (e) mean filtering (3 × 3 convolutional kernel); (f) mean filtering (5 × 5 convolutional kernel); (g) mean filtering (7 × 7 convolutional kernel); (h) Gaussian filtering (1 × 1 convolutional kernel); (i) Gaussian filtering (3 × 3 convolutional kernel); (j) Gaussian filtering (5 × 5 convolutional kernel); (k) Gaussian filtering (7 × 7 convolutional kernel).
Sensors 24 05981 g020
Figure 21. Various performance indicators of different models: (a) Train_loss; (b) Val_loss; (c) precision; (d) recall; (e) mAP0.5; (f) mAP0.5:0.95.
Figure 21. Various performance indicators of different models: (a) Train_loss; (b) Val_loss; (c) precision; (d) recall; (e) mAP0.5; (f) mAP0.5:0.95.
Sensors 24 05981 g021aSensors 24 05981 g021b
Figure 22. Other performance indicators of different models: (a) the precision–recall curve; (b) F1 curve.
Figure 22. Other performance indicators of different models: (a) the precision–recall curve; (b) F1 curve.
Sensors 24 05981 g022
Figure 23. Crack detection results of different YOLOv9 models.
Figure 23. Crack detection results of different YOLOv9 models.
Sensors 24 05981 g023
Figure 24. Crack detection results of two experimental images.
Figure 24. Crack detection results of two experimental images.
Sensors 24 05981 g024
Figure 25. The detection process of other underwater and surface crack images: (a,b) underwater crack images under low light conditions; (c,d) underwater crack images with significant color differences; (e,f) above-water images.
Figure 25. The detection process of other underwater and surface crack images: (a,b) underwater crack images under low light conditions; (c,d) underwater crack images with significant color differences; (e,f) above-water images.
Sensors 24 05981 g025
Table 1. Image metrics corresponding to different white balance methods (Figure 16a).
Table 1. Image metrics corresponding to different white balance methods (Figure 16a).
Processing MethodUICMUISMUIConMUIQMPSNRSSIM
Original14.8970.136−0.119 0.032--
Mean white balance17.4410.145−0.1180.11130.5910.987
Perfect reflection18.1190.140−0.1260.10230.1020.985
Ours18.3950.140−0.1180.13639.7830.988
Color cast17.2600.133−0.1120.12337.4060.987
Gray world17.5510.141−0.1220.09922.4250.977
Table 2. Image metrics corresponding to different white balance methods.
Table 2. Image metrics corresponding to different white balance methods.
Original ImageProcessing MethodUICMUISMUIConMUIQMPSNRSSIM
Figure 16aOriginal14.8970.136−0.119 0.032--
Mean white balance17.4410.145−0.1180.11130.5910.987
Perfect reflection18.1190.140−0.1260.10230.1020.985
Ours18.3950.140−0.1180.13639.7830.988
Color cast17.2600.133−0.1120.12337.4060.987
Gray world17.5510.141−0.1220.09922.4250.977
Figure 16bOriginal15.2060.199−0.154 −0.063--
Mean white balance17.3140.213−0.1440.03521.7710.980
Perfect reflection16.1670.201−0.150−0.01716.9030.948
Ours17.9950.200−0.1440.05127.9600.982
Color cast17.6560.198−0.1440.04127.2370.982
Gray world15.710.156−0.146−0.02111.2780.890
Figure 16cOriginal18.2270.260−0.142 0.082--
Mean white balance17.1310.277−0.1430.05425.7380.983
Perfect reflection17.7060.266−0.1490.04524.0520.975
Ours18.4120.252−0.1430.08438.1760.984
Color cast17.6620.256−0.1390.07537.2440.984
Gray world17.6420.263−0.1440.06019.5320.964
Figure 16dOriginal17.5570.126−0.1220.094--
Mean white balance17.3410.138−0.1230.08928.4690.990
Perfect reflection17.4570.130−0.1340.05129.0820.988
Ours18.2300.111−0.1230.11239.0230.991
Color cast17.9820.123−0.1200.10536.3710.990
Gray world18.0340.129−0.1250.10631.4820.989
Figure 16eOriginal16.8500.158−0.1270.069--
Mean white balance17.4250.163−0.1250.09328.7310.984
Perfect reflection17.2620.161−0.1380.09923.1590.974
Ours18.0370.159−0.1240.11238.9290.984
Color cast16.7300.156−0.1210.08737.4490.984
Gray world16.1870.157−0.1250.05416.1300.949
Table 3. 1. Image metrics of different image denoising methods in Figure 19a–c. 2. Image metrics of different image denoising methods in Figure 19d,e.
Table 3. 1. Image metrics of different image denoising methods in Figure 19a–c. 2. Image metrics of different image denoising methods in Figure 19d,e.
1
Input ImageProcessing MethodUICMUISMUIConMUIQMPSNRSSIM
Figure 19aInput18.3950.140−0.1180.13639.7830.988
Bilateral filtering18.3330.124−0.1140.14540.3260.988
Median filtering18.3990.134−0.1180.13839.7680.975
Mean filtering (1 × 1)18.4040.140−0.1190.13639.7780.988
Mean filtering (3 × 3)18.3880.141−0.1180.13739.7230.987
Mean filtering (5 × 5)18.3470.135−0.1160.14139.2650.987
Mean filtering (7 × 7)18.3290.122−0.1140.14438.5390.972
Gaussian filtering (1 × 1)18.3980.140−0.1190.13639.7780.988
Gaussian filtering (3 × 3)18.3950.139−0.1180.13739.7990.988
Gaussian filtering (5 × 5)18.3860.135−0.1170.13839.6740.986
Gaussian filtering (7 × 7)18.3650.130−0.1160.14039.4130.983
Figure 19bInput17.9950.200−0.1440.05127.9600.982
Bilateral filtering17.9910.129−0.1370.05527.9900.982
Median filtering17.9880.174−0.1420.05127.9210.942
Mean filtering (1 × 1)17.9930.200−0.1440.05227.9600.974
Mean filtering (3 × 3)17.9900.174−0.1420.05127.9140.971
Mean filtering (5 × 5)17.9900.148−0.1390.05227.8320.971
Mean filtering (7 × 7)17.9900.128−0.1370.05527.7500.935
Gaussian filtering (1 × 1)17.9930.200−0.1440.05227.9600.982
Gaussian filtering (3 × 3)17.9850.177−0.1420.05127.9410.977
Gaussian filtering (5 × 5)17.9890.165−0.1410.05127.9110.970
Gaussian filtering (7 × 7)17.9860.149−0.1400.05227.8680.960
Input18.4120.252−0.1430.08438.1760.984
Figure 19cBilateral filtering18.3870.185−0.1350.09138.2030.984
Median filtering18.4050.230−0.1410.08438.0700.946
Mean filtering (1 × 1)18.4190.252−0.1430.08438.1670.982
Mean filtering (3 × 3)18.3860.234−0.1410.08437.9840.981
Mean filtering (5 × 5)18.3910.209−0.1380.08837.0180.981
Mean filtering (7 × 7)18.3810.181−0.1350.09035.8070.936
Gaussian filtering (1 × 1)18.4190.252−0.1430.08438.1670.984
Gaussian filtering (3 × 3)18.4010.235−0.1410.08438.1230.983
Gaussian filtering (5 × 5)18.3940.223−0.1400.08537.8400.978
Gaussian filtering (7 × 7)18.3930.205−0.1380.08637.2850.968
2
Input ImageProcessing MethodUICMUISMUIConMUIQMPSNRSSIM
Figure 19dInput18.230.111−0.1230.11239.0230.991
Bilateral filtering18.0180.131−0.1220.11439.0540.991
Median filtering18.0230.113−0.1230.10239.0480.983
Mean filtering (1 × 1)18.040.111−0.1230.10239.0130.991
Mean filtering (3 × 3)18.010.129−0.1240.10339.0460.99
Mean filtering (5 × 5)18.0190.139−0.1230.10938.7960.99
Mean filtering (7 × 7)18.0210.127−0.1220.1138.3350.982
Gaussian filtering (1 × 1)18.040.111−0.1230.10239.0130.991
Gaussian filtering (3 × 3)18.0280.12−0.1240.10239.0770.991
Gaussian filtering (5 × 5)18.0250.124−0.1230.10439.0250.99
Gaussian filtering (7 × 7)18.0270.123−0.1230.10638.8830.988
Figure 19eInput18.0370.159−0.1240.11238.9290.984
Bilateral filtering18.0030.113−0.1170.12438.9430.984
Median filtering18.0330.143−0.1220.11538.6970.952
Mean filtering (1 × 1)18.0430.16−0.1240.11238.9280.984
Mean filtering (3 × 3)18.0150.147−0.1230.11338.6450.979
Mean filtering (5 × 5)17.9950.128−0.1190.11837.8170.979
Mean filtering (7 × 7)17.9940.113−0.1170.12437.0620.946
Gaussian filtering (1 × 1)18.0430.16−0.1240.11238.9280.984
Gaussian filtering (3 × 3)18.0260.147−0.1230.11338.8270.982
Gaussian filtering (5 × 5)18.020.139−0.1210.11538.5390.977
Gaussian filtering (7 × 7)17.9960.128−0.1190.11838.0980.969
Table 4. Image metrics corresponding to different image denoising methods.
Table 4. Image metrics corresponding to different image denoising methods.
Original ImageProcessing MethodUICMUISMUIConMUIQMPSNRSSIM
Figure 16aOriginal14.8970.136−0.119 0.032--
Processed images18.3330.124−0.1140.14540.3260.988
Figure 16bOriginal15.2060.199−0.154 −0.063--
Processed images17.9910.129−0.1370.05527.9900.982
Figure 16cOriginal18.2270.260−0.142 0.082--
Processed images18.3870.185−0.1350.09138.2030.984
Figure 16dOriginal17.5570.126−0.1220.094--
Processed images18.0180.131−0.1220.11439.0540.991
Figure 16eOriginal16.8500.158−0.1270.069--
Processed images18.0030.113−0.1170.12438.9430.984
Table 5. Training parameters.
Table 5. Training parameters.
Parameter NameMeaningThe Set Value
Hardware configurationComputer parameters12th Gen Intel(R) Core(TM) i7-12700F 2.10 GHz (Intel, Santa Clara, CA, USA)
Environment configurationVirtual environmentPython 3.8, PyTorch==1.13, CUDA 11.7
Batch sizeInput size4
WorkersNumber of data-loading threads4
EpochsIterations200
lr0Initial learning rate0.01
lrfCosine annealing hyperparameter0.01
MomentumLearning rate momentum0.937
weight_decayWeight attenuation coefficient0.0005
Table 6. Comprehensive test results of different training models.
Table 6. Comprehensive test results of different training models.
Model NamePRF1mAP0.5mAP0.5:0.95
YOLOv973.2%68.1%0.71074.1%42.2%
YOLOv9-e70.2%68.0%0.69073.0%41.3%
YOLOv9-c66.0%68.9%0.67069.4%38.4%
YOLOv9-m74.1%65.9%0.70072.2%41.7%
YOLOv9-s68.6%69.5%0.68072.7%41.5%
YOLOv9-t64.1%66.7%0.65069.7%38.3%
YOLOv9-OREPA72.4%73.6%0.73074.8%44.1%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, X.; Liang, C.; Li, X.; Kang, F. An Underwater Crack Detection System Combining New Underwater Image-Processing Technology and an Improved YOLOv9 Network. Sensors 2024, 24, 5981. https://doi.org/10.3390/s24185981

AMA Style

Huang X, Liang C, Li X, Kang F. An Underwater Crack Detection System Combining New Underwater Image-Processing Technology and an Improved YOLOv9 Network. Sensors. 2024; 24(18):5981. https://doi.org/10.3390/s24185981

Chicago/Turabian Style

Huang, Xinbo, Chenxi Liang, Xinyu Li, and Fei Kang. 2024. "An Underwater Crack Detection System Combining New Underwater Image-Processing Technology and an Improved YOLOv9 Network" Sensors 24, no. 18: 5981. https://doi.org/10.3390/s24185981

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop