Next Article in Journal
Design of an Electronic Interface for Single-Photon Avalanche Diodes
Previous Article in Journal
Detection of Debonding Defects in Carbon Fiber-Reinforced Polymer (CFRP)–Rubber Bonded Structures Based on Active Lamb Wave Energy Analysis
Previous Article in Special Issue
A Discussion of Building a Smart SHM Platform for Long-Span Bridge Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Leakage Identification of Underground Structures Using Classification Deep Neural Networks and Transfer Learning

1
Shandong Zhiyuan Electric Power Design Consulting Co., Ltd., Jinan 250021, China
2
Economic & Technology Research Institute of State Grid Shandong Electric Power Company, Jinan 250021, China
3
Hunan Province Key Laboratory for Disaster Prevention and Mitigation of Rail Transit Engineering Structure, Central South University, Changsha 410075, China
4
School of Civil Engineering, Central South University, Changsha 410075, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(17), 5569; https://doi.org/10.3390/s24175569
Submission received: 18 July 2024 / Revised: 19 August 2024 / Accepted: 24 August 2024 / Published: 28 August 2024

Abstract

:
Water leakage defects often occur in underground structures, leading to accelerated structural aging and threatening structural safety. Leakage identification can detect early diseases of underground structures and provide important guidance for reinforcement and maintenance. Deep learning-based computer vision methods have been rapidly developed and widely used in many fields. However, establishing a deep learning model for underground structure leakage identification usually requires a lot of training data on leakage defects, which is very expensive. To overcome the data shortage, a deep neural network method for leakage identification is developed based on transfer learning in this paper. For comparison, four famous classification models, including VGG16, AlexNet, SqueezeNet, and ResNet18, are constructed. To train the classification models, a transfer learning strategy is developed, and a dataset of underground structure leakage is created. Finally, the classification performance on the leakage dataset of different deep learning models is comparatively studied under different sizes of training data. The results showed that the VGG16, AlexNet, and SqueezeNet models with transfer learning can overall provide higher and more stable classification performance on the leakage dataset than those without transfer learning. The ResNet18 model with transfer learning can overall provide a similar value of classification performance on the leakage dataset than that without transfer learning, but its classification performance is more stable than that without transfer learning. In addition, the SqueezeNet model obtains an overall higher and more stable performance than the comparative models on the leakage dataset for all classification metrics.

1. Introduction

With the rapid development of urban areas all over the world, many engineering structures are constructed underground, such as metro tunnels [1], underground garages [2], underground substations [3], etc. Underground structures are usually affected by many adverse internal and external effects, including corrosion, groundwater, and operation loads. These effects may lead to the performance deterioration of the underground structures. Water leakage is a very common disease of underground structures, which can accelerate structural aging and cause serious structural safety accidents. Therefore, it is very important to conduct water leakage detection for underground structures during the operation process.
Conventional manual detection methods require a lot of human and material resources, and efficiency is low. In recent years, structural health monitoring has been studied and applied widely [4,5,6,7,8,9]. Using advanced sensors and data mining algorithms, structural health monitoring systems can provide timely condition information on structures. Compared with the conventional manual detection method, the detection results of structural health monitoring are more objective, labor saving, and cost-effective [10]. There are many kinds of sensors used in structural health monitoring, which can obtain different condition information on the monitoring objectives. Generally, these sensors can be divided into contact and non-contact sensors. Contact sensors need to be in contact with the monitoring object to collect information on its changes, including the physical quantities such as pressure, stress, strain, and temperature. When using a non-contact sensor, physical contact with the monitoring object is not required, and the data on the monitoring object can be measured by measuring changes in the electric field, magnetic field, light, sound, etc. In recent years, as a kind of non-contact sensor, the vision sensor has been widely used in civil engineering structure health monitoring [11,12,13,14,15,16]. In the field of leakage identification of underground structures, many computer vision techniques also have been proposed for the classification and location of leakage defects [17,18,19,20]. These computer vision techniques generally can be divided into four categories, including image classification [21], object detection [22], semantic segmentation [23], and instance segmentation [24]. The image classification method is used to identify whether there is a leakage defect in the images. Object detection methods can be used to detect the leakage defect in the images and locate the regions. The semantic segmentation method can classify the pixels of different defects in the images. At last, the instance segmentation method can classify the pixels of different items of each defect in the images. Compared with the other three methods, the image classification method generally needs lower computational cost and is ideal for fast leakage defect detection.
With the development of artificial intelligence, many deep learning-based computer vision methods have been proposed. These methods have been validated to be more powerful than conventional computer vision methods. The VGG model [25], AlexNet model [26], SqueezeNet model [27], and ResNet model [28] are four very famous deep learning-based classification models that can provide high-accuracy classification results for many datasets. However, most deep learning models need a lot of training data to ensure classification performance. If there is insufficient training data to train the deep neural network, the accuracy of leakage identification will be low. To deal with the insufficient training data problems, transfer learning [29,30,31], few-shot meta-learning [32,33], semi-supervised learning [34,35], active learning [36,37,38], and other advanced methods have been proposed recently. In these methods, the latter several methods usually use carefully designed network structures or training data selection strategies to improve the classification performance of the deep neural network. But, the training data are restricted to the limited objective dataset. However, the transfer learning methods take advantage of the related datasets, which may be large enough to pretrain the deep neural network of classification models. The pretrained deep neural network has a strong ability to extract the features of the images. Then, much smaller training data in the objective dataset is needed to retrain the deep neural network. Because the transfer learning method can use the information from both the training data of the objective dataset and the related datasets, and the size of the related datasets is not limited, the transfer learning method greatly saves the training data in objective dataset. Recently, transfer learning has attracted some attention in the field of defect detection of engineering structures [39,40,41]. However, most of the references only consider one classification model. How transfer learning affects the performance of different deep learning-based classification models for leakage identification of underground structures is unclear.
In this paper, a leakage identification method for underground structures is developed based on the classification of deep neural networks and transfer learning. The main contributions of this paper are as follows. (1) The performance of the classification models incorporating transfer learning is validated in the leakage identification of underground structures and (2) the influence of transfer learning on different classification models (including the VGG16 model, AlexNet model, SqueezeNet model, and ResNet model) and the influence of training data size on their classification performance are comparatively studied.

2. Methodology

2.1. Deep Neural Network for Leakage Identification

In this paper, four deep neural networks for leakage identification were studied, which were the VGG16 model, AlexNet model, SqueezeNet model, and ResNet model. All pretrained models were downloaded from the Torchvision library. The input size of the model was 224 × 224 × 3. To classify the leakage dataset, the output size of the models was changed to 1 × 1 × 2.
The VGG (Visual Geometry Group) model is a deep convolutional neural network architecture proposed by the Visual Geometry Group in 2014. The architecture of the modified VGG16 model is shown in Figure 1. The original VGG16 model had 16 weight layers, including 13 convolutional layers and 3 fully connected layers. To classify the leakage defect, a fully connected layer with two outputs was added to the pretrained model. VGG16 has a simple network structure and is easy to implement, making it one of the classic models in the field of deep learning-based computer vision. Due to the deep convolutional structure, the number of parameters in the VGG model is very large.
AlexNet is a classic convolutional neural network, and it was proposed in the 2012 ImageNet image classification competition. The architecture of the modified AlexNet model is shown in Figure 2. As shown in this figure, the modified AlexNet model had five convolutional layers and four fully connected layers, where the last fully connected layer was added to change the output classes. AlexNet was an early implementation of a deep convolutional neural network. By increasing the network depth, AlexNet was able to better learn the features of the dataset, thereby improving the accuracy of image classification.
SqueezeNet is a lightweight deep learning model that can achieve high prediction accuracy with fewer model parameters. Figure 3 shows the architecture of the modified SqueezeNet model. The modified SqueezeNet model had two convolutional layers, four pooling layers, and eight fire modules.
The architecture of the modified ResNet18 model is shown in Figure 4. As shown in this figure, the modified ResNet18 model had many convolutional layers and pooling layers. The last pooling layer was fully connected with the output layer. In the ResNet model, shortcut connections are built between skip layers to deal with the vanishing and exploding gradient problems, so ResNet can be very deep.

2.2. Transfer Learning Strategy

As mentioned above, all pretrained models were downloaded from the Torchvision library. The models were pretrained on large datasets (e.g., ImageNet). Therefore, these pretrained models had the ability to extract the features from the images. As shown in the examples in Figure 5, in transfer learning, the lower-level layer of the pretrained models was frozen, and the parameters of these layers were not updated in the retraining process. However, to classify the target leakage dataset, the high-level fully connected layers of the VGG16 model and AlexNet model and the resized output layers of the SqueezeNet model and ResNet18 model were retrained using the training data from the target leakage dataset.
In the training of all classification models, the cross-entropy between the prediction distribution and the real distribution was used as the loss function. The cross-entropy was calculated by
E x = i = 1 n P i x log Q i l x
where P i x is the predicted probability that the sample x belongs to the i th class and Q i l x is the real probability that the sample x belongs to the i th class. The mini-batch gradient descent and RMSProp (root mean square propagation) algorithm were used to update the deep neural networks. The mini-batch size was set to 5, and the learning rate was set to 10 5 . The number of training epochs was determined by k-fold cross-validation.

2.3. Dataset Preparation

The datasets used in the examples were created through three approaches, including an online search, on-site photography, and an open-source dataset [42]. As shown in Figure 6, Figure 7, Figure 8 and Figure 9, the datasets contained water leakage images of the underground garage, water leakage images of the underground equipment room, water leakage images of the underground tunnel lining, images of underground structures without water leakage, etc. There were 136 high-resolution leakage images collected through an online search and on-site photography. By cropping, flipping, and adding noise, 1431 small-sized leakage images were obtained. The other 4655 leakage images were directly collected from the open-source dataset. By cropping 100 high-resolution no-leakage images of the open-source dataset, 1200 small-sized no-leakage images were obtained. Then, by flipping in two directions and adding noise, the other 6000 small-sized no-leakage images were obtained. Finally, the original datasets were extended to a new dataset with 13,286 images in total. In the new dataset, the number of water leakage images was 6086, and the number of images without water leakage was 7200. All images were in RGB format and had a size of 224 × 224 × 3.
In the experiments, a certain proportion of data was randomly selected as the training set, and the remaining data were used to test the classification performance. To compare the performance of the pretrained models under the different sizes of the training set, the ratio of the training set changed from 0.05 to 0.30 with a step of 0.01, the size of the training set increased from 664 to 3986, and the size of the remaining test set changed from 12,622 to 9300. A detailed configuration of the training set and test set of 26 different experiments is shown in Table 1.

2.4. Identification Performance Evaluation Metrics

The classification metrics used in this paper included classification accuracy, classification precision, classification recall, and classification F1 score. As shown by the confusion matrix in Figure 10, T P represents the number of leakage images accurately classified as leakage and F P represents the number of no-leakage images mistakenly classified as leakage. Similarly, T N represents the number of no-leakage images accurately classified as no leakage and F N represents the number of leakage images mistakenly classified as no leakage.
The classification metrics were then calculated by
A c c u r a c y = T P + T N T P + F P + F N + T N
  P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 - s c o r e = 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l

3. Experimental Results

To test the impact of transfer learning on the classification performance of deep neural networks, the deep neural networks trained using different sizes of data were used to classify the remaining test set. For each test, both the training methods with and without transfer learning were used. In the method with transfer learning, the deep neural networks and their pretrained parameters were both downloaded from Torchvision library, and the parameters of the low-level layers were not updated in the training process. On the contrary, in the method without transfer learning, the deep neural networks were also downloaded from Torchvision library, but all model parameters were randomly initialized. It is worth noting that the models with and without transfer learning used the same hyperparameter configuration.

3.1. Example 1: Test Results of the VGG16 Model

In this section, the test results of the VGG16 model are discussed. Table 2 shows the prediction results of the VGG16 model with and without transfer learning. Figure 11, Figure 12, Figure 13 and Figure 14 compare classification accuracy, classification precision, classification recall, and classification F1 score by the VGG16 model with and without transfer learning. As shown in Figure 11, Figure 12, Figure 13 and Figure 14, the classification metrics by the VGG16 model with transfer learning were overall higher than the method without transfer learning. When the ratio of training data was 0.05, classification accuracy, classification precision, classification recall, and classification F1 score by the VGG16 model with transfer learning were 0.928, 0.927, 0.929, and 0.928, respectively, which were higher than 0.868, 0.867, 0.867, and 0.867 by the VGG16 model without transfer learning. The heatmap of the confusion matrix of the prediction results by the VGG16 models when the ratio of training data was 0.05 is shown in Figure 15. With the increase in the ratio of training data, both methods obtained higher classification metrics. However, the increasing process of the VGG16 model with transfer learning was more stable than the VGG16 model without transfer learning.
When the ratio of training data reached 0.2, the classification metrics by the VGG16 model without transfer learning were similar to those of the VGG16 model with transfer learning. When the ratio of training data was 0.3, classification accuracy, classification precision, classification recall, and classification F1 score by the VGG16 model with transfer learning were 0.962, 0.966, 0.960, and 0.962, respectively, and the results by the VGG16 model without transfer learning were 0.965, 0.966, 0.965, and 0.965, respectively. Due to the small randomness of training, the trained VGG16 model without transfer learning achieved slightly better results than the VGG16 model trained with transfer learning. The results showed that transfer learning could significantly improve classification performance of the VGG16 model on leakage defects when the ratio of training data was lower than 0.2 and could deal with the insufficient training data problem.

3.2. Example 2: Test Results of the AlexNet Model

The test results of the AlexNet model are discussed in this section. The detailed prediction results of the AlexNet model with and without transfer learning are shown in Table 3, and classification accuracy, classification precision, classification recall, and classification F1 score by the AlexNet model with and without transfer learning are shown in Figure 16, Figure 17, Figure 18 and Figure 19.
As shown in Figure 16, Figure 17, Figure 18 and Figure 19, the classification metrics on the leakage dataset by the AlexNet model with transfer learning were overall higher than the method without transfer learning. When the ratio of training data was 0.05, classification accuracy, classification precision, classification recall, and classification F1 score by the AlexNet model with transfer learning were 0.934, 0.933, 0.934, and 0.933, respectively, which were higher than 0.861, 0.866, 0.856, and 0.859 by the AlexNet model without transfer learning. The heatmap of the confusion matrix of the prediction results by the AlexNet models when the ratio of training data was 0.05 is shown in Figure 20. However, the AlexNet model with transfer learning reached very high classification metrics when the ratio of training data was 0.05. With the increase in the ratio of training data, the classification metrics by the AlexNet model with transfer learning still slightly increased. However, the classification metrics by the AlexNet model without transfer learning were not stable. It changed within a range of approximately 0.85 to 0.95 when the ratio of training data increased.
When the ratio of training data was 0.3, classification accuracy, classification precision, classification recall, and classification F1 score by the AlexNet model with transfer learning were 0.971, 0.973, 0.969, and 0.970, respectively, but the results by the VGG16 model without transfer learning were 0.912, 0.916, 0.917, and 0.912, respectively. The results showed that the AlexNet model with transfer learning could obtain higher classification performance on leakage defects than that without transfer learning. Because of the constant pretrained model parameters of the AlexNet model, the AlexNet model with transfer learning also obtained a more stable classification performance on leakage defects.

3.3. Example 3: Test Results of the SqueezeNet Model

In this section, the test results of the SqueezeNet model are discussed. Table 4 shows the prediction results of the SqueezeNet model with and without transfer learning. Figure 21, Figure 22, Figure 23 and Figure 24 compare classification accuracy, classification precision, classification recall, and classification F1 score by the SqueezeNet model with and without transfer learning. As shown in Figure 21, Figure 22, Figure 23 and Figure 24, the classification metrics by the SqueezeNet model with transfer learning were overall higher than the method without transfer learning. When the ratio of training data was 0.05, classification accuracy, classification precision, classification recall, and classification F1 score by the SqueezeNet model with transfer learning were 0.957, 0.959, 0.955, and 0.957, respectively, which were higher than 0.934, 0.934, 0.933, and 0.934 by the SqueezeNet model without transfer learning. The heatmap of the confusion matrix of the prediction results by the SqueezeNet models when the ratio of training data was 0.05 is shown in Figure 25.
With the increase in the ratio of training data, the classification metrics by the SqueezeNet model with transfer learning increased approximately monotonically. However, the increasing process of the SqueezeNet model without transfer learning was unstable. When the ratio of training data was 0.3, the SqueezeNet models with and without transfer learning obtained similar classification metrics on the leakage dataset, and classification accuracy, classification precision, classification recall, and classification F1 score by the SqueezeNet model with transfer learning were 0.977, 0.977, 0.976, and 0.977, respectively and the results by the SqueezeNet model without transfer learning were 0.969, 0.968, 0.970, and 0.969, respectively. The results showed that using transfer learning, the SqueezeNet model could provide higher and more stable classification metrics on leakage defects.

3.4. Example 4: Test Results of the ResNet18 Model

This section presents the test results of the ResNet18 model. The detailed prediction results of the ResNet18 model with and without transfer learning are shown in Table 5.
Classification accuracy, classification precision, classification recall, and classification F1 score by the ResNet18 model with and without transfer learning are shown in Figure 26, Figure 27, Figure 28 and Figure 29. As shown in Figure 26, Figure 27, Figure 28 and Figure 29, the classification metrics on the leakage dataset by the ResNet18 model with transfer learning were overall similar to those without transfer learning. When the ratio of training data was 0.05, classification accuracy, classification precision, classification recall, and classification F1 score by the ResNet18 model with transfer learning were 0.938, 0.940, 0.935, and 0.937, respectively. The prediction results by the ResNet18 model without transfer learning were 0.933, 0.935, 0.931, and 0.933, respectively. The heatmap of the confusion matrix of the prediction results by the ResNet18 model when the ratio of training data was 0.05 is shown in Figure 30. With the increase in the ratio of training data, the classification metrics by the ResNet18 model with transfer learning increased slightly. However, the classification metrics by the ResNet18 model without transfer learning were not stable. When the ratio of training data was 0.3, classification accuracy, classification precision, classification recall, and classification F1 score by the ResNet18 model with transfer learning were 0.961, 0.963, 0.959, and 0.961, respectively, which is similar to the 0.967, 0.969, 0.965, and 0.966 of the ResNet18 model without transfer learning. The results showed that the classification metrics on leakage defects by the ResNet18 model with transfer learning were overall similar to those without transfer learning. This may be because ResNet18 has a specific network structure (e.g., shortcut connection between skip layers), and ResNet18 has strong feature extraction capabilities and can achieve similar performance to ResNet18 with transfer learning using a small amount of training data. However, due to the pretrained model parameters, the prediction results by the ResNet18 model with transfer learning were more stable than the ResNet18 model without transfer learning.

3.5. Comparison of the Different Classification Models with Transfer Learning

In the last section, the classification performance of the different classification models with transfer learning was compared. Figure 31, Figure 32, Figure 33 and Figure 34 show classification accuracy, classification precision, classification recall, and classification F1 score on the leakage dataset by different classification models. As shown in these figures, overall, the SqueezeNet model obtained higher performance than the comparative models on the leakage dataset for all classification metrics.
When the train data ratio was 0.05, the prediction accuracy of the SqueezeNet model was 0.957, which was higher than 0.928, 0.934, and 0.938 of the VGG16, AlexNet, and ResNet18 models. For classification precision, the SqueezeNet model obtained a value of 0.959 when the train data ratio was 0.05, but the values for the VGG16, AlexNet, and ResNet18 models were 0.927, 0.933, and 0.940, respectively. When the train data ratio was 0.05, the obtained classification recall by the SqueezeNet model was 0.955, which was also higher than 0.929, 0.934, and 0.935 of the VGG16, AlexNet, and ResNet18 models. Considering classification F1 score, the prediction value of the SqueezeNet model was 0.957, but the results of the VGG16, AlexNet, and ResNet18 models were 0.928, 0.933, and 0.937, respectively. With the increase in the train data ratio, all models obtained higher classification performance. However, the increased process of classification performance of the SqueezeNet model was more stable than the other models. When the ratio of training data was 0.3, classification accuracy, classification precision, classification recall, and classification F1 score by the VGG16 model with transfer learning were, respectively, 0.962, 0.966, 0.960, and 0.962. Those of the AlexNet model with transfer learning were, respectively, 0.971, 0.973, 0.969, and 0.970. The four classification metrics of the ResNet18 model with transfer learning were, respectively, 0.961, 0.963, 0.959, and 0.961. However, when the ratio of training data was 0.3, classification accuracy, classification precision, classification recall, and classification F1 score by the SqueezeNet model with transfer learning were 0.977, 0.977, 0.976, and 0.977, respectively, which were higher than those of the other methods. The results showed that the SqueezeNet model with transfer learning has a higher classification performance on the leakage defects than the other comparative methods.
To test the computational efficiency of the models, the training time of different final models with and without transfer learning was compared, as shown in Table 6 and Figure 35. There are many factors that affect training time, such as computer equipment, the number of model parameters, training data size, the number of training epochs, etc. The computer equipment used in this paper was a personal computer with an AMD Ryzen 9 5950X 16-Core Processor CPU and an NVIDIA GeForce RTX 3090 GPU. For a specific ratio of training data, the training data sizes for different models were equal, and the number of model parameters and training epochs might be the main factors affecting the training time.
As shown in Table 6 and Figure 35, the average training time for the VGG16 model, AlexNet model, SqueezeNet model, and ResNet model with transfer learning were 5.2 min, 4.4 min, 18.8 min, and 16.5 min, respectively. The average training time for the VGG16 model, AlexNet model, SqueezeNet model, and ResNet model without transfer learning were 13.8 min, 8.8 min, 11.3 min, and 10.6 min, respectively. The results showed that AlexNet achieved the highest computational efficiency, regardless of whether transfer learning was used or not. The VGG16 and AlexNet models with transfer learning used less training time than those without transfer learning. That was because the VGG16 and AlexNet models with transfer learning required a smaller number of training epochs to provide high prediction accuracy. However, the SqueezeNet and ResNet18 models with transfer learning used more training time than those without transfer learning, which may be because the training process of the models without transfer learning was more unstable than the models with transfer learning, and the optimal number of train epochs determined by cross-validation was smaller. As shown in Figure 35, with the increase in the train data ratio, the training time of the models with transfer learning increased. However, the training time of the models without transfer learning was unstable when the train data ratio increased. This phenomenon may also be caused by the unstable training process of the models without transfer learning.

4. Conclusions and Discussion

In this paper, a transfer learning strategy was developed to deal with the problems of data shortage in training deep learning-based classification models for underground structure leakage identification. The classification performance of four famous classification models, including VGG16, AlexNet, SqueezeNet, and ResNet18, with the transfer learning on the leakage dataset was comparatively studied under different sizes of training data.
The results showed that the VGG16, AlexNet, and SqueezeNet models with transfer learning could overall provide higher and more stable classification performance on the leakage dataset than those without transfer learning. The ResNet18 model with transfer learning could overall provide a similar value of classification performance on the leakage dataset than that without transfer learning, but its classification performance was more stable than that without transfer learning. When the ratio of training data was 0.05, classification accuracy, classification precision, classification recall, and classification F1 score by the VGG16 model with transfer learning were 0.928, 0.927, 0.929, and 0.928, respectively, and those of the VGG16 model without transfer learning were 0.868, 0.867, 0.867, and 0.867 respectively. The classification metrics by the AlexNet model with transfer learning were 0.934, 0.933, 0.934, and 0.933, respectively, which were higher than 0.861, 0.866, 0.856, and 0.859 by the AlexNet model without transfer learning. The classification metrics by the SqueezeNet model with transfer learning were 0.957, 0.959, 0.955, and 0.957, respectively, and those of the SqueezeNet model without transfer learning were 0.934, 0.934, 0.933, and 0.934. The classification metrics by the ResNet18 model with transfer learning were 0.938, 0.940, 0.935, and 0.937, respectively, and the prediction results by the ResNet18 model without transfer learning were 0.933, 0.935, 0.931, and 0.933, respectively.
In addition, the SqueezeNet model obtained overall higher and more stable performance than the comparative models on the leakage dataset for all classification metrics. When the train data ratio was 0.05, the prediction accuracy of the SqueezeNet model was 0.957, which was higher than 0.928, 0.934, and 0.938 of the VGG16, AlexNet, and ResNet18 models. For the classification precision, the SqueezeNet model obtained a value of 0.959 when the train data ratio was 0.05, but the values for the VGG16, AlexNet, and ResNet18 models were 0.927, 0.933, and 0.940, respectively. The obtained classification recall by the SqueezeNet model was 0.955, which was also higher than 0.929, 0.934, and 0.935 of the VGG16, AlexNet, and ResNet18 models. Considering the classification F1 score, the prediction value of the SqueezeNet model was 0.957, but the results of the VGG16, AlexNet, and ResNet18 models were 0.928, 0.933, and 0.937, respectively.
In the methods with transfer learning, the classification models inherit a strong ability to extract the features of images. During the retraining process in this paper, the feature extraction layers were frozen, and only the weights in the classification layers were updated. The retraining process is actually learning the mapping from extracted features to the classification results. Therefore, the retraining process is generally easier than training the entire classification models without transfer, and less training data are required. In this paper, all the feature extraction layers of the models with transfer learning were fixed during the retraining process. However, if some of the convolution layers are retrained, the classification performance may change. The optimal retraining scheme for different models is worth further research.

Author Contributions

Conceptualization, Z.X. and Y.S.; methodology, Z.X. and Y.S.; software, Z.X. and Y.S.; validation, Z.X. and Y.S.; formal analysis, Z.X.; investigation, Z.X.; resources, W.W. and Q.C.; data curation, W.W. and Q.C.; writing—original draft preparation, Z.X.; writing—review and editing, Y.S.; visualization, Z.X.; supervision, Y.S.; project administration, W.W., Q.C. and Y.S.; funding acquisition, W.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by technology project funding from Shandong Zhiyuan Electric Power Design Consulting Co., Ltd. (ZY-2023-04), the National Natural Science Foundation of China (52208223), the Hunan Provincial Natural Science Foundation (2023JJ40754), and the China Postdoctoral Science Foundation (2022M713545).

Data Availability Statement

Data supporting the reported results can be provided by the corresponding author upon reasonable request.

Conflicts of Interest

The authors Wenyang Wang and Qingwei Chen are employed in State Grid Shandong Electric Power Company. The authors declare no conflicts of interest.

References

  1. Liu, W.; Shao, Y.; Li, C.; Li, C.; Jiang, Z. Development of a non-Gaussian copula Bayesian network for safety assessment of metro tunnel maintenance. Reliab. Eng. Syst. Saf. 2023, 238, 109423. [Google Scholar] [CrossRef]
  2. Chu, C.R.; Su, Z.Y. Natural ventilation design for underground parking garages. Build. Environ. 2023, 227, 109784. [Google Scholar] [CrossRef]
  3. Guizán, R.; Colominas, I.; París, J.; Couceiro, I.; Navarrina, F. Numerical analysis and safety design of grounding systems in underground compact substations. Electr. Power Syst. Res. 2022, 203, 107627. [Google Scholar] [CrossRef]
  4. Bao, Y.; Li, H. Machine learning paradigm for structural health monitoring. Struct. Health Monit. 2021, 20, 1353–1372. [Google Scholar] [CrossRef]
  5. Bao, Y.; Chen, Z.; Wei, S.; Xu, Y.; Tang, Z.; Li, H. The state of the art of data science and engineering in structural health monitoring. Engineering 2019, 5, 234–242. [Google Scholar] [CrossRef]
  6. Zhou, Z.; Zhou, Z.; Lu, C.; He, C. A tunnel structure health monitoring method based on surface strain monitoring. J. Civ. Struct. Health Monit. 2024, 14, 1339–1367. [Google Scholar] [CrossRef]
  7. Mishra, M.; Lourenço, P.B.; Ramana, G.V. Structural health monitoring of civil engineering structures by using the internet of things: A review. J. Build. Eng. 2022, 48, 103954. [Google Scholar] [CrossRef]
  8. Wang, M.; Yin, X. Construction and maintenance of urban underground infrastructure with digital technologies. Autom. Constr. 2022, 141, 104464. [Google Scholar] [CrossRef]
  9. Tan, X.; Chen, W.; Zou, T.; Yang, J.; Du, B. Real-time prediction of mechanical behaviors of underwater shield tunnel structure using machine learning method based on structural health monitoring data. J. Rock Mech. Geotech. Eng. 2023, 15, 886–895. [Google Scholar] [CrossRef]
  10. Xiang, Z.; He, X.; Zou, Y.; Jing, H. An active learning method for crack detection based on subset searching and weighted sampling. Struct. Health Monit. 2024, 23, 1184–1200. [Google Scholar] [CrossRef]
  11. Dong, C.Z.; Catbas, F.N. A review of computer vision-based structural health monitoring at local and global levels. Struct. Health Monit. 2021, 20, 692–743. [Google Scholar] [CrossRef]
  12. Bao, Y.; Tang, Z.; Li, H.; Zhang, Y. Computer vision and deep learning-based data anomaly detection method for structural health monitoring. Struct. Health Monit. 2019, 18, 401–421. [Google Scholar] [CrossRef]
  13. Jang, K.; An, Y.; Kim, B.; Cho, S. Automated crack evaluation of a high-rise bridge pier using a ring-type climbing robot. Comput.-Aided Civ. Infrastruct. Eng. 2021, 36, 14–29. [Google Scholar] [CrossRef]
  14. Kim, B.; Cho, S. Image-based concrete crack assessment using mask and region-based convolutional neural network. Struct. Control Health Monit. 2019, 26, e2381. [Google Scholar] [CrossRef]
  15. Guo, Y.; Cui, H.; Li, S. Excavator joint node-based pose estimation using lightweight fully convolutional network. Autom. Constr. 2022, 141, 104435. [Google Scholar] [CrossRef]
  16. Ai, D.; Jiang, G.; Lam, S.-K.; He, P.; Li, C. Computer vision framework for crack detection of civil infrastructure—A review. Eng. Appl. Artif. Intell. 2023, 117, 105478. [Google Scholar] [CrossRef]
  17. Chen, J.; Xu, X.; Jeon, G.; Camacho, D.; He, B.-G. WLR-Net: An Improved YOLO-V7 With Edge Constraints and Attention Mechanism for Water Leakage Recognition in the Tunnel. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 3105–3116. [Google Scholar] [CrossRef]
  18. Li, J.; He, Z.; Li, D.; Zheng, A. Research on water seepage detection technology of tunnel asphalt pavement based on deep learning and digital image processing. Sci. Rep. 2022, 12, 11519. [Google Scholar] [CrossRef]
  19. Xu, T.; Xu, L.; Li, X.; Yao, J. Detection of water leakage in underground tunnels using corrected intensity data and 3D point cloud of terrestrial laser scanning. IEEE Access 2018, 6, 32471–32480. [Google Scholar] [CrossRef]
  20. Huang, H.; Li, Q.; Zhang, D. Deep learning based image recognition for crack and leakage defects of metro shield tunnel. Tunn. Undergr. Space Technol. 2018, 77, 166–176. [Google Scholar] [CrossRef]
  21. Lu, D.; Weng, Q. A survey of image classification methods and techniques for improving classification performance. Int. J. Remote Sens. 2007, 28, 823–870. [Google Scholar] [CrossRef]
  22. Zou, Z.; Chen, K.; Shi, Z.; Guo, Y.; Ye, J. Object detection in 20 years: A survey. Proc. IEEE 2023, 111, 257–276. [Google Scholar] [CrossRef]
  23. Hao, S.; Zhou, Y.; Guo, Y. A brief survey on semantic segmentation with deep learning. Neurocomputing 2020, 406, 302–321. [Google Scholar] [CrossRef]
  24. Hafiz, A.M.; Bhat, G.M. A survey on instance segmentation: State of the art. Int. J. Multimed. Inf. Retr. 2020, 9, 171–189. [Google Scholar] [CrossRef]
  25. Qassim, H.; Verma, A.; Feinzimer, D. Compressed residual-VGG16 CNN model for big data places image recognition. In Proceedings of the 2018 IEEE 8th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 8–10 January 2018; IEEE: Las Vegas, NV, USA, 2018; pp. 169–175. [Google Scholar]
  26. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  27. Lee, H.J.; Ullah, I.; Wan, W.; Gao, Y.; Fang, Z. Real-time vehicle make and model recognition with the residual SqueezeNet architecture. Sensors 2019, 19, 982. [Google Scholar] [CrossRef]
  28. Targ, S.; Almeida, D.; Lyman, K. Resnet in resnet: Generalizing residual architectures. arXiv 2016, arXiv:1603.08029. [Google Scholar]
  29. Zhu, J.; Zhang, C.; Qi, H.; Lu, Z. Vision-based defects detection for bridges using transfer learning and convolutional neural networks. Struct. Infrastruct. Eng. 2020, 16, 1037–1049. [Google Scholar] [CrossRef]
  30. Pan, Q.; Bao, Y.; Li, H. Transfer learning-based data anomaly detection for structural health monitoring. Struct. Health Monit. 2023, 22, 3077–3091. [Google Scholar] [CrossRef]
  31. Liu, J.; Guo, F.; Gao, H.; Huang, Z.; Zhang, Y.; Zhou, H. Image classification method on class imbalance datasets using multi-scale CNN and two-stage transfer learning. Neural Comput. Appl. 2021, 33, 14179–14197. [Google Scholar] [CrossRef]
  32. Cui, Z.; Wang, Q.; Guo, J.; Lu, N. Few-shot classification of façade defects based on extensible classifier and contrastive learning. Autom. Constr. 2022, 141, 104381. [Google Scholar] [CrossRef]
  33. Xu, Y.; Bao, Y.; Zhang, Y.; Li, H. Attribute-based structural damage identification by few-shot meta learning with inter-class knowledge transfer. Struct. Health Monit. 2021, 20, 1494–1517. [Google Scholar] [CrossRef]
  34. Guo, J.; Wang, Q.; Li, Y. Semi-supervised learning based on convolutional neural network and uncertainty filter for façade defects classification. Comput.-Aided Civ. Infrastruct. Eng. 2021, 36, 302–317. [Google Scholar] [CrossRef]
  35. Shim, S.; Kim, J.; Lee, S.-W.; Cho, G.-C. Road damage detection using super-resolution and semi-supervised learning with generative adversarial network. Autom. Constr. 2022, 135, 104139. [Google Scholar] [CrossRef]
  36. Guo, J.; Wang, Q.; Su, S.; Li, Y. Informativeness-guided active learning for deep learning-based façade defects detection. Comput.-Aided Civ. Infrastruct. Eng. 2023, 38, 2408–2425. [Google Scholar] [CrossRef]
  37. Zheng, Y.; Gao, Y.; Lu, S.; Mosalam, K.M. Multistage semisupervised active learning framework for crack identification, segmentation, and measurement of bridges. Comput.-Aided Civ. Infrastruct. Eng. 2022, 37, 1089–1108. [Google Scholar] [CrossRef]
  38. Shu, J.; Li, J.; Zhang, J.; Zhao, W.; Duan, Y.; Zhang, Z. An active learning method with difficulty learning mechanism for crack detection. Smart Struct. Syst. 2022, 29, 195–206. [Google Scholar]
  39. Man, K.; Liu, R.; Liu, X.; Song, Z.; Liu, Z.; Cao, Z.; Wu, L. Water leakage and crack identification in tunnels based on transfer-learning and convolutional neural networks. Water 2022, 14, 1462. [Google Scholar] [CrossRef]
  40. Zoubir, H.; Rguig, M.; El Aroussi, M.; Chehri, A.; Saadane, R.; Jeon, G. Concrete bridge defects identification and localization based on classification deep convolutional neural networks and transfer learning. Remote Sens. 2022, 14, 4882. [Google Scholar] [CrossRef]
  41. Hou, S.; Dong, B.; Wang, H.; Wu, G. Inspection of surface defects on stay cables using a robot and transfer learning. Autom. Constr. 2020, 119, 103382. [Google Scholar] [CrossRef]
  42. Xue, Y.; Cai, X.; Shadabfar, M.; Shao, H.; Zhang, S. Deep learning-based automatic recognition of water leakage area in shield tunnel lining. Tunn. Undergr. Space Technol. 2020, 104, 103524. [Google Scholar] [CrossRef]
Figure 1. Modified VGG16 model.
Figure 1. Modified VGG16 model.
Sensors 24 05569 g001
Figure 2. Modified AlexNet model.
Figure 2. Modified AlexNet model.
Sensors 24 05569 g002
Figure 3. Modified SqueezeNet model.
Figure 3. Modified SqueezeNet model.
Sensors 24 05569 g003
Figure 4. Modified ResNet18 model.
Figure 4. Modified ResNet18 model.
Sensors 24 05569 g004
Figure 5. Transfer learning configurations of a deep neural network for water leakage identification.
Figure 5. Transfer learning configurations of a deep neural network for water leakage identification.
Sensors 24 05569 g005
Figure 6. Water leakage image of the underground garage.
Figure 6. Water leakage image of the underground garage.
Sensors 24 05569 g006
Figure 7. Water leakage image of the underground equipment room.
Figure 7. Water leakage image of the underground equipment room.
Sensors 24 05569 g007
Figure 8. Water leakage image of the underground tunnel lining.
Figure 8. Water leakage image of the underground tunnel lining.
Sensors 24 05569 g008
Figure 9. Images of the underground structure without water leakage.
Figure 9. Images of the underground structure without water leakage.
Sensors 24 05569 g009
Figure 10. Confusion matrix of leakage classification.
Figure 10. Confusion matrix of leakage classification.
Sensors 24 05569 g010
Figure 11. Example 1: prediction accuracy of different methods under different training data ratios.
Figure 11. Example 1: prediction accuracy of different methods under different training data ratios.
Sensors 24 05569 g011
Figure 12. Example 1: prediction precision of different methods under different training data ratios.
Figure 12. Example 1: prediction precision of different methods under different training data ratios.
Sensors 24 05569 g012
Figure 13. Example 1: prediction recall of different methods under different training data ratios.
Figure 13. Example 1: prediction recall of different methods under different training data ratios.
Sensors 24 05569 g013
Figure 14. Example 1: prediction F1 score of different methods under different training data ratios.
Figure 14. Example 1: prediction F1 score of different methods under different training data ratios.
Sensors 24 05569 g014
Figure 15. Example 1: heatmap of the confusion matrix of the prediction results when the ratio of training data was 0.05. (a) VGG16 with transfer and (b) VGG16 without transfer.
Figure 15. Example 1: heatmap of the confusion matrix of the prediction results when the ratio of training data was 0.05. (a) VGG16 with transfer and (b) VGG16 without transfer.
Sensors 24 05569 g015
Figure 16. Example 2: prediction accuracy of different methods under different training data ratios.
Figure 16. Example 2: prediction accuracy of different methods under different training data ratios.
Sensors 24 05569 g016
Figure 17. Example 2: prediction precision of different methods under different training data ratios.
Figure 17. Example 2: prediction precision of different methods under different training data ratios.
Sensors 24 05569 g017
Figure 18. Example 2: prediction recall of different methods under different training data ratios.
Figure 18. Example 2: prediction recall of different methods under different training data ratios.
Sensors 24 05569 g018
Figure 19. Example 2: prediction F1 score of different methods under different training data ratios.
Figure 19. Example 2: prediction F1 score of different methods under different training data ratios.
Sensors 24 05569 g019
Figure 20. Example 2: heatmap of the confusion matrix of the prediction results when the ratio of training data was 0.05. (a) AlexNet with transfer and (b) AlexNet without transfer.
Figure 20. Example 2: heatmap of the confusion matrix of the prediction results when the ratio of training data was 0.05. (a) AlexNet with transfer and (b) AlexNet without transfer.
Sensors 24 05569 g020
Figure 21. Example 3: prediction accuracy of different methods under different training data ratios.
Figure 21. Example 3: prediction accuracy of different methods under different training data ratios.
Sensors 24 05569 g021
Figure 22. Example 3: prediction precision of different methods under different training data ratios.
Figure 22. Example 3: prediction precision of different methods under different training data ratios.
Sensors 24 05569 g022
Figure 23. Example 3: prediction recall of different methods under different training data ratios.
Figure 23. Example 3: prediction recall of different methods under different training data ratios.
Sensors 24 05569 g023
Figure 24. Example 3: prediction F1 score of different methods under different training data ratios.
Figure 24. Example 3: prediction F1 score of different methods under different training data ratios.
Sensors 24 05569 g024
Figure 25. Example 3: heatmap of the confusion matrix of the prediction results when the ratio of training data was 0.05. (a) SqueezeNet with transfer and (b) SqueezeNet without transfer.
Figure 25. Example 3: heatmap of the confusion matrix of the prediction results when the ratio of training data was 0.05. (a) SqueezeNet with transfer and (b) SqueezeNet without transfer.
Sensors 24 05569 g025
Figure 26. Example 4: prediction accuracy of different methods under different training data ratios.
Figure 26. Example 4: prediction accuracy of different methods under different training data ratios.
Sensors 24 05569 g026
Figure 27. Example 4: prediction precision of different methods under different training data ratios.
Figure 27. Example 4: prediction precision of different methods under different training data ratios.
Sensors 24 05569 g027
Figure 28. Example 4: prediction recall of different methods under different training data ratios.
Figure 28. Example 4: prediction recall of different methods under different training data ratios.
Sensors 24 05569 g028
Figure 29. Example 4: prediction F1 score of different methods under different training data ratios.
Figure 29. Example 4: prediction F1 score of different methods under different training data ratios.
Sensors 24 05569 g029
Figure 30. Example 4: heatmap of the confusion matrix of the prediction results when the ratio of training data was 0.05. (a) ResNet18 with transfer and (b) ResNet18 without transfer.
Figure 30. Example 4: heatmap of the confusion matrix of the prediction results when the ratio of training data was 0.05. (a) ResNet18 with transfer and (b) ResNet18 without transfer.
Sensors 24 05569 g030
Figure 31. Comparison of mean prediction accuracy by different pretrained models.
Figure 31. Comparison of mean prediction accuracy by different pretrained models.
Sensors 24 05569 g031
Figure 32. Comparison of mean prediction precision by different pretrained models.
Figure 32. Comparison of mean prediction precision by different pretrained models.
Sensors 24 05569 g032
Figure 33. Comparison of mean prediction recall by different pretrained models.
Figure 33. Comparison of mean prediction recall by different pretrained models.
Sensors 24 05569 g033
Figure 34. Comparison of mean prediction F1 score by different pretrained models.
Figure 34. Comparison of mean prediction F1 score by different pretrained models.
Sensors 24 05569 g034
Figure 35. Comparison of the training time of different models under different training data ratios.
Figure 35. Comparison of the training time of different models under different training data ratios.
Sensors 24 05569 g035
Table 1. Configuration of the training set and test set of different experiments.
Table 1. Configuration of the training set and test set of different experiments.
Test NumberRatio of Training SetSize of Training SetSize of Test Set
10.0566412,622
20.0679712,489
30.0793012,356
40.08106312,223
50.09119612,090
60.10132911,957
70.11146111,825
80.12159411,692
90.13172711,559
100.14186011,426
110.15199311,293
120.16212611,160
130.17225911,027
140.18239110,895
150.19252410,762
160.20265710,629
170.21279010,496
180.22292310,363
190.23305610,230
200.24318910,097
210.2533229965
220.2634549832
230.2735879699
240.2837209566
250.2938539433
260.3039869300
Table 2. Example 1: prediction results of the VGG16 model with and without transfer learning.
Table 2. Example 1: prediction results of the VGG16 model with and without transfer learning.
RatioAccuracyPrecisionRecallF1 Score
TransferYesNoYesNoYesNoYesNo
0.05 0.928 0.868 0.927 0.867 0.929 0.867 0.928 0.867
0.06 0.881 0.846 0.892 0.874 0.890 0.834 0.881 0.839
0.07 0.944 0.833 0.945 0.856 0.942 0.822 0.943 0.826
0.08 0.948 0.855 0.949 0.870 0.948 0.846 0.948 0.850
0.09 0.939 0.725 0.944 0.761 0.936 0.739 0.938 0.722
0.10 0.938 0.917 0.944 0.916 0.934 0.916 0.937 0.916
0.11 0.934 0.882 0.934 0.890 0.938 0.889 0.934 0.882
0.12 0.947 0.827 0.950 0.852 0.945 0.838 0.947 0.826
0.13 0.949 0.914 0.952 0.914 0.946 0.912 0.948 0.913
0.14 0.957 0.934 0.960 0.934 0.954 0.933 0.956 0.934
0.15 0.948 0.912 0.952 0.914 0.945 0.916 0.947 0.912
0.16 0.951 0.814 0.955 0.869 0.948 0.798 0.950 0.800
0.17 0.961 0.937 0.962 0.937 0.960 0.936 0.961 0.936
0.18 0.944 0.930 0.951 0.929 0.939 0.933 0.942 0.930
0.19 0.958 0.939 0.957 0.946 0.959 0.934 0.958 0.938
0.20 0.949 0.954 0.948 0.953 0.951 0.955 0.948 0.954
0.21 0.955 0.937 0.958 0.944 0.952 0.933 0.954 0.936
0.22 0.961 0.942 0.961 0.946 0.961 0.939 0.961 0.942
0.23 0.954 0.942 0.953 0.941 0.956 0.941 0.954 0.941
0.24 0.948 0.960 0.948 0.960 0.951 0.959 0.948 0.959
0.25 0.967 0.963 0.968 0.964 0.967 0.962 0.967 0.963
0.26 0.960 0.962 0.962 0.962 0.959 0.962 0.960 0.962
0.27 0.961 0.951 0.963 0.950 0.959 0.952 0.961 0.950
0.28 0.968 0.973 0.969 0.973 0.967 0.972 0.968 0.973
0.29 0.965 0.923 0.966 0.934 0.964 0.917 0.965 0.921
0.30 0.962 0.965 0.966 0.966 0.960 0.965 0.962 0.965
Table 3. Example 2: prediction results of the AlexNet model with and without transfer learning.
Table 3. Example 2: prediction results of the AlexNet model with and without transfer learning.
RatioAccuracyPrecisionRecallF1 Score
Transfer YesNoYesNoYesNoYesNo
0.05 0.934 0.861 0.933 0.866 0.934 0.856 0.933 0.859
0.06 0.917 0.831 0.920 0.873 0.922 0.816 0.917 0.820
0.07 0.953 0.918 0.957 0.921 0.949 0.915 0.952 0.917
0.08 0.955 0.912 0.959 0.912 0.952 0.910 0.954 0.911
0.09 0.951 0.883 0.956 0.886 0.947 0.888 0.950 0.883
0.10 0.963 0.940 0.963 0.942 0.963 0.939 0.963 0.940
0.11 0.963 0.875 0.963 0.896 0.962 0.865 0.963 0.870
0.12 0.958 0.926 0.960 0.927 0.957 0.925 0.958 0.926
0.13 0.963 0.937 0.964 0.937 0.961 0.940 0.962 0.937
0.14 0.947 0.932 0.954 0.931 0.943 0.932 0.946 0.932
0.15 0.955 0.927 0.954 0.926 0.956 0.927 0.955 0.926
0.16 0.964 0.909 0.964 0.912 0.964 0.914 0.964 0.909
0.17 0.972 0.953 0.972 0.952 0.971 0.953 0.971 0.953
0.18 0.968 0.943 0.969 0.947 0.966 0.940 0.967 0.942
0.19 0.970 0.940 0.971 0.946 0.969 0.936 0.970 0.939
0.20 0.972 0.959 0.973 0.960 0.971 0.958 0.972 0.959
0.21 0.972 0.909 0.973 0.924 0.971 0.902 0.972 0.907
0.22 0.964 0.935 0.967 0.934 0.961 0.935 0.964 0.935
0.23 0.969 0.898 0.970 0.903 0.968 0.904 0.969 0.898
0.24 0.969 0.954 0.972 0.953 0.967 0.955 0.969 0.954
0.25 0.966 0.968 0.969 0.967 0.964 0.968 0.966 0.968
0.26 0.966 0.958 0.966 0.957 0.968 0.957 0.966 0.957
0.27 0.964 0.810 0.967 0.850 0.962 0.823 0.964 0.808
0.28 0.977 0.956 0.978 0.959 0.976 0.954 0.977 0.956
0.29 0.973 0.970 0.975 0.971 0.971 0.969 0.973 0.970
0.30 0.971 0.912 0.973 0.916 0.969 0.917 0.970 0.912
Table 4. Example 3: prediction results of the SqueezeNet model with and without transfer learning.
Table 4. Example 3: prediction results of the SqueezeNet model with and without transfer learning.
RatioAccuracyPrecisionRecallF1 Score
Transfer YesNoYesNoYesNoYesNo
0.05 0.957 0.934 0.959 0.934 0.955 0.933 0.957 0.933
0.06 0.963 0.935 0.964 0.934 0.961 0.937 0.962 0.935
0.07 0.963 0.921 0.963 0.922 0.962 0.925 0.962 0.921
0.08 0.964 0.923 0.965 0.922 0.962 0.925 0.963 0.923
0.09 0.966 0.964 0.966 0.963 0.965 0.964 0.965 0.964
0.10 0.966 0.956 0.966 0.955 0.965 0.957 0.965 0.955
0.11 0.967 0.945 0.969 0.945 0.965 0.949 0.966 0.945
0.12 0.967 0.956 0.968 0.956 0.965 0.956 0.966 0.956
0.13 0.969 0.944 0.969 0.947 0.968 0.941 0.969 0.943
0.14 0.972 0.909 0.972 0.913 0.971 0.914 0.971 0.909
0.15 0.970 0.967 0.970 0.966 0.969 0.968 0.969 0.967
0.16 0.970 0.959 0.970 0.959 0.969 0.958 0.970 0.959
0.17 0.969 0.959 0.970 0.961 0.968 0.957 0.969 0.959
0.18 0.971 0.967 0.972 0.969 0.971 0.965 0.971 0.966
0.19 0.971 0.974 0.971 0.974 0.970 0.973 0.971 0.973
0.20 0.970 0.960 0.971 0.962 0.968 0.958 0.969 0.960
0.21 0.975 0.956 0.976 0.957 0.975 0.955 0.975 0.956
0.22 0.973 0.971 0.973 0.971 0.972 0.970 0.972 0.971
0.23 0.974 0.968 0.975 0.970 0.973 0.965 0.974 0.967
0.24 0.976 0.946 0.978 0.947 0.975 0.950 0.976 0.946
0.25 0.977 0.978 0.977 0.978 0.976 0.978 0.976 0.978
0.26 0.974 0.953 0.976 0.953 0.973 0.951 0.974 0.952
0.27 0.975 0.970 0.976 0.970 0.975 0.969 0.975 0.969
0.28 0.976 0.973 0.976 0.973 0.976 0.973 0.976 0.973
0.29 0.977 0.976 0.977 0.978 0.976 0.975 0.976 0.976
0.30 0.977 0.969 0.977 0.968 0.976 0.970 0.977 0.969
Table 5. Example 4: prediction results of the ResNet18 model with and without transfer learning.
Table 5. Example 4: prediction results of the ResNet18 model with and without transfer learning.
RatioAccuracyPrecisionRecallF1 Score
Transfer YesNoYesNoYesNoYesNo
0.05 0.938 0.933 0.940 0.935 0.935 0.931 0.937 0.933
0.06 0.947 0.948 0.946 0.948 0.947 0.948 0.946 0.948
0.07 0.945 0.956 0.945 0.957 0.944 0.954 0.945 0.956
0.08 0.956 0.942 0.957 0.942 0.954 0.944 0.955 0.942
0.09 0.951 0.926 0.953 0.927 0.950 0.931 0.951 0.926
0.10 0.953 0.926 0.953 0.932 0.953 0.921 0.953 0.925
0.11 0.954 0.962 0.954 0.961 0.953 0.963 0.954 0.962
0.12 0.949 0.951 0.948 0.954 0.950 0.948 0.948 0.950
0.13 0.953 0.952 0.953 0.952 0.951 0.952 0.952 0.952
0.14 0.951 0.953 0.953 0.958 0.949 0.950 0.951 0.953
0.15 0.950 0.939 0.951 0.941 0.949 0.936 0.950 0.938
0.16 0.959 0.972 0.959 0.971 0.960 0.972 0.959 0.971
0.17 0.956 0.983 0.957 0.983 0.954 0.984 0.956 0.983
0.18 0.959 0.976 0.961 0.976 0.958 0.976 0.959 0.976
0.19 0.964 0.923 0.963 0.923 0.964 0.927 0.963 0.923
0.20 0.960 0.978 0.960 0.978 0.960 0.979 0.960 0.978
0.21 0.947 0.935 0.953 0.935 0.943 0.939 0.946 0.935
0.22 0.958 0.969 0.959 0.968 0.956 0.969 0.957 0.968
0.23 0.961 0.975 0.961 0.975 0.961 0.975 0.961 0.975
0.24 0.957 0.983 0.959 0.985 0.954 0.982 0.956 0.983
0.25 0.962 0.980 0.961 0.979 0.962 0.980 0.962 0.979
0.26 0.959 0.986 0.959 0.987 0.959 0.985 0.959 0.986
0.27 0.962 0.979 0.961 0.979 0.961 0.979 0.961 0.979
0.28 0.962 0.982 0.963 0.981 0.960 0.982 0.961 0.982
0.29 0.958 0.965 0.959 0.964 0.956 0.967 0.957 0.965
0.30 0.961 0.967 0.963 0.969 0.959 0.965 0.961 0.966
Table 6. Training time of different models under different training data ratios (min).
Table 6. Training time of different models under different training data ratios (min).
RatioVGG16AlexNetSqueezeNetResNet18
Transfer YesNoYesNoYesNoYesNo
0.05 212137465
0.06 28127467
0.07 24247465
0.08 28249585
0.09 28239983
0.10 31025101196
0.11 31034117113
0.12 37351171210
0.13 485131081212
0.14 3104101316157
0.15 41147159196
0.16 4134815111412
0.17 4214101581623
0.18 5165816111810
0.19 52541319171720
0.20 5216816141814
0.21 717512219207
0.22 61761322121810
0.23 8245162592117
0.24 81551529232620
0.25 91471729183411
0.26 91681431182715
0.27 877632142010
0.28 824682910168
0.29 81591638172320
0.30 10195442182810
Average5.213.84.48.818.811.316.510.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, W.; Chen, Q.; Shen, Y.; Xiang, Z. Leakage Identification of Underground Structures Using Classification Deep Neural Networks and Transfer Learning. Sensors 2024, 24, 5569. https://doi.org/10.3390/s24175569

AMA Style

Wang W, Chen Q, Shen Y, Xiang Z. Leakage Identification of Underground Structures Using Classification Deep Neural Networks and Transfer Learning. Sensors. 2024; 24(17):5569. https://doi.org/10.3390/s24175569

Chicago/Turabian Style

Wang, Wenyang, Qingwei Chen, Yongjiang Shen, and Zhengliang Xiang. 2024. "Leakage Identification of Underground Structures Using Classification Deep Neural Networks and Transfer Learning" Sensors 24, no. 17: 5569. https://doi.org/10.3390/s24175569

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop