Next Article in Journal
ASCEND-UNet: An Improved UNet Configuration Optimized for Rural Settlements Mapping
Previous Article in Journal
Non-Tectonic Geohazards of Guangdong Province, China, Monitored Using Sentinel-1A/B from 2015 to 2022
Previous Article in Special Issue
A Novel Intelligent Fault Diagnosis Method for Bearings with Multi-Source Data and Improved GASA
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DAEiS-Net: Deep Aggregation Network with Edge Information Supplement for Tunnel Water Stain Segmentation

1
Beijing Metro Construction Administration Co., Ltd., Beijing 100068, China
2
Beijing Key Laboratory of Fully Automatic Operation and Safety Monitoring for Urban Rail Transit, Beijing 100068, China
3
College of Civil Engineering, Tongji University, Shanghai 200092, China
4
Beijing MTR Corporation Ltd., Beijing 100068, China
5
School of Advanced Manufacturing Engineering, Chongqing University of Posts and Telecommunication Technology, Chongqing 400065, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(17), 5452; https://doi.org/10.3390/s24175452 (registering DOI)
Submission received: 10 July 2024 / Revised: 9 August 2024 / Accepted: 14 August 2024 / Published: 23 August 2024
(This article belongs to the Special Issue AI-Assisted Condition Monitoring and Fault Diagnosis)

Abstract

:
Tunnel disease detection and maintenance are critical tasks in urban engineering, and are essential for the safety and stability of urban transportation systems. Water stain detection presents unique challenges due to its variable morphology and scale, which leads to insufficient multiscale contextual information extraction and boundary information loss in complex environments. To address these challenges, this paper proposes a method called Deep Aggregation Network with Edge Information Supplement (DAEiS-Net) for detecting tunnel water stains. The proposed method employs a classic encoder–decoder architecture. Specifically, in the encoder part, a Deep Aggregation Module (DAM) is introduced to enhance feature representation capabilities. Additionally, a Multiscale Cross-Attention Module (MCAM) is proposed to suppress noise in the shallow features and enhance the texture information of the high-level features. Moreover, an Edge Information Supplement Module (EISM) is designed to mitigate semantic gaps across different stages of feature extraction, improving the extraction of water stain edge information. Furthermore, a Sub-Pixel Module (SPM) is proposed to fuse features at various scales, enhancing edge feature representation. Finally, we introduce the Tunnel Water Stain Dataset (TWS), specifically designed for tunnel water stain segmentation. Experimental results on the TWS dataset demonstrate that DAEiS-Net achieves state-of-the-art performance in tunnel water stain segmentation.

1. Introduction

Urban transportation systems have become the lifeline for billions in today’s bustling metropolises, driving the heartbeat of cities and fostering urban development. As urban areas continue to expand, the demand for reliable transportation infrastructure has reached unprecedented levels. Tunnels are critical components of these systems, providing commuters with fast and convenient transportation options [1,2]. The safety and stability of tunnels, as essential components of transportation and urban infrastructure, is directly linked to the sustainable development of the socioeconomic landscape. Due to complex geological conditions, diverse construction techniques, and natural and human factors in long-term operations, tunnel structures are prone to various diseases such as cracks, leaks, spalling, and deformation [3,4,5]. If these diseases are not detected and repaired promptly, they may pose serious safety risks and even lead to catastrophic accidents; therefore, disease detection and maintenance in tunnels have become critical and urgent tasks in the field of engineering.
Traditional structural disease detection methods often rely on on-site manual inspections, which are time-consuming and labor-intensive [6,7,8]. Various non-destructive methods, such as acoustic emission [9,10,11,12,13], visual imaging [14,15,16], and ultrasonic tomography [17], have been employed for tunnel detection to enhance efficiency. Additionally, Order–Frequency Holo-Hilbert spectral analysis has been used to demodulate fault information in time-varying vibration signals [18]. However, these methods require manual parameter adjustments, and their detection effectiveness is often suboptimal. To address these issues, researchers have increasingly focused on machine vision approaches. With advancements in information technology, automated detection techniques based on image processing and computer vision have become research hotspots. In particular, the rapid development of deep learning technologies has provided novel solutions for tunnel disease detection. Deep learning models, particularly Convolutional Neural Networks (CNNs) [19,20,21], have demonstrated outstanding performance in tasks such as image classification, object detection, and image segmentation [22,23,24]. These models can automatically learn and extract complex features from images, enabling accurate identification and segmentation of tunnel diseases.
As machine learning [25,26,27] and deep learning [28,29,30] technologies develop, deep learning-based tunnel disease detection tasks can be primarily divided into two categories, namely, object detection and semantic segmentation. In the domain of object detection, several noteworthy studies have been conducted; for instance, Cha et al. [31] proposed a vision-based method that utilizes CNNs to detect concrete cracks without computing disease features. The designed CNN network was trained on 40,000 images, achieving an accuracy of 98%. Zhou et al. [32] introduced a YOLOv4-based model that incorporates depthwise separable convolution, enhancing the recognition accuracy of disease targets under complex tunnel backgrounds and lighting conditions. Li et al. [33] proposed a multi-scale disease region proposal network capable of generating disease region proposals on multilayer feature maps, with each layer focusing on diseases within a certain scale range. This multiscale detector ultimately improves the accuracy of disease detection. Li et al. [34] leveraged deep learning and computer vision to identify underwater structural damage in hydraulic tunnels. A high-performance detector was constructed based on YOLOv5 and an adaptive spatial feature fusion module. Through experiments, they determined the sparsity ratio and pruning rate, balancing accuracy and efficiency. Liu et al. [35] introduced a belief capsule network with a multi-stream capsule routing strategy and a consistency-aware fusion strategy to generate high-quality pseudolabels to solve the problem of deep supervised significant object detection being heavily dependent on labeling.
In the field of semantic segmentation, Chun et al. [36] proposed a supervised dynamic crack detection method. By considering the pixel values of target pixels and the geometric features of cracks when they are linearly connected, they introduced a segmentation method based on pixel values and geometric shapes. This method is applicable for tunnel crack segmentation under adverse lighting conditions such as the presence of shadows and dirt. Zhou et al. [37] addressed the characteristics of cracks and linear seams by proposing a Mixed Attention (MA) module based on effective embedding channels and positional information. Unlike common spatial attention modules that aggregate information across the entire space, the MA module aggregates features directly along the spatial dimensions of height and width. Liao et al. [38] introduced a novel mobile tunnel inspection system comprising a unique mobile imaging module and an automatic crack detection module. A novel lightweight CNN was utilized for tunnel crack detection, incorporating effective spatial constraint strategies to maintain the continuity of cracks.
Tunnel water stain disease segmentation, a specific type of tunnel disease detection, is particularly challenging due to the complexity and variability of the tunnel environment. Insufficient lighting inside tunnels often leads to images with low contrast and with significant noise. Additionally, water stains vary widely in morphology and can appear on different materials and surfaces, resulting in diverse visual features that are challenging to capture. Consequently, accurate segmentation of tunnel water stain diseases faces two main challenges: (1) significant variation in the morphology and size of different types of disease, which leads to insufficient extraction of multiscale contextual information; and (2) complex environmental and lighting conditions, which cause the loss of boundary detail information.
To address these challenges, this paper proposes a novel approach leveraging deep aggregation features supplemented with edge information. The proposed method, named Deep Aggregation Network with Edge Information Supplement (DAEiS-Net), adopts a classic encoder–decoder architecture for tunnel water leakage detection. Specifically, we introduce the Deep Aggregation Module (DAM) in the encoder section to enhance the network’s feature representation capabilities. To extract multiscale contextual information of water leakage and mitigate the effects of variable morphology and size, a Multiscale Cross-Attention Module (MCAM) is proposed. This module integrates spatial and semantic information, reducing noise in shallow features and enhancing texture information extraction. Additionally, to address inadequacies in extracting water leakage edge information, the Edge Information Supplement Module (EISM) is designed to bridge the semantic gaps at different stages, effectively integrating precise detailed information with rich semantic content. Furthermore, in the Sub-Pixel Decoder (SPD) is utilized as a Sub-Pixel Module (SPM) to combine features at various scales, improving the representation of edge features. Finally, we introduce the Tunnel Water Stain Dataset (TWS), a dataset specifically designed for tunnel water stain segmentation, covering various tunnel water leakage scenarios. Experimental results on the TWS dataset demonstrate that DAEiS-Net achieves state-of-the-art performance. The main contributions of this paper are summarized as follows:
  • A novel tunnel water leakage segmentation network model called DAEiS-Net is proposed, focusing primarily on aggregating multiscale contextual information and extracting edge information. By enhancing multiscale fusion, the segmentation results are significantly improved.
  • The proposed DAM incorporates MCAM to adaptively extract the multiscale contextual information of water leakage. The EISM is designed to integrate features from various stages, effectively bridging semantic gaps and enhancing edge information extraction. The SPD module is used in the decoding process to enhance edge feature representation.
  • The TWS dataset is developed to encompass a wide range of tunnel water leakage scenarios. This dataset provides a solid foundation for training and evaluating models.
The remainder of this paper is organized as follows: in Section 2, we review related work on tunnel water leakage segmentation; Section 3 provides a detailed description of our proposed method; in Section 4, we evaluate the effectiveness of our method; finally, Section 5 concludes the paper.

2. Related Work

2.1. Semantic Segmentation

Semantic segmentation algorithms directly separate relevant pixels from the background and visually display the predicted diseases in the image. Long et al. [39] proposed the first Fully Convolutional Network (FCN), which converts all fully connected layers in CNNs into convolutional layers and achieves semantic segmentation through pixel-wise classification. To address the slow speed of multistage methods and the low accuracy of single-stage methods, Ronneberger et al. [40] introduced the classic U-Net, a U-shaped encoder–decoder network. This method utilizes an encoder for downsampling to extract high-level image features and a decoder for upsampling to restore image details and spatial dimensions, resulting in excellent outcomes. To address the issue of network spatial inconsistency, Zhao et al. [41] proposed PSPNet, which utilizes a Spatial Pyramid Pooling (SPP) module to fuse multiscale contextual information, thereby reducing the probability of mis-segmentation inherent in FCNs. In order to capture richer multiscale contextual semantic information, Chen et al. [42] introduced DeepLabv1, which employs atrous convolution to overcome the issues of reduced image resolution and irrecoverable position information caused by repeated pooling and down-sampling. Chen et al. [43] later proposed SoCo, which utilizes the concept of self-supervised contrastive learning. Self-supervised constrastive learning extracts features by learning the similarity relationships between different parts of an image, achieving more accurate and robust natural semantic segmentation in images. However, due to multiple downsampling operations during feature extraction, existing algorithms lose a great deal of detail in the features, resulting in absence of effective local contextual information. Additionally, object edges are often segmented inaccurately due to incomplete feature fusion and insufficient global contextual information.

2.2. Segmentation Models Applied to Tunnel Disease Detection

Yang et al. [44] proposed a method based on U-Net for welding disease segmentation, utilizing an attention-guided segmentation network. To mitigate the loss of contextual information caused by multiple convolution and pooling operations, a multiscale feature fusion module was integrated into the U-Net to capture more comprehensive information. In response to the challenges of complex environmental interference and multiscale target recognition in tunnel disease identification, Zhou et al. [45] introduced a novel segmentation algorithm called Multiscale Attention and Context Information Enhancement. This algorithm was designed to create a context-enhanced feature encoder to address the limitations of CNNs. It aims to fully extract global contextual information and reduce false detections and omissions caused by complex environmental interference. To address the issue of small crack segmentation, Chu et al. [46] proposed a multiscale feature fusion network with an attention mechanism, which they named Tiny-Crack-Net (TCN). This network utilizes an enhanced residual network to capture local features of small cracks, and incorporates dual attention modules into the architecture to effectively distinguish tiny cracks from the background. Qin et al. [47] introduced a disease segmentation model based on Vision Transformer (ViT), which significantly differs from traditional CNNs. An adapter and decoding head were proposed to enhance the training effectiveness of the transformer encoder, enabling it to adapt to small-scale datasets.

3. Methodology

The overall structure of the proposed method is illustrated in Figure 1. It adopts the encoder–decoder framework of U-Net, with ResNet-34 as the backbone network. In the encoder part, we introduce the DAM to enhance the representation capability of the network. The DAM incorporates an MCAM to fuse the semantic and spatial information, improving the model’s ability to extract multiscale information of tunnel water leakage and alleviating the impact of variable morphology and scale. Moreover, an EISM is employed to bridge the semantic gaps at different stages. The EISM effectively integrates precise detail information with rich semantic information to address the insufficiency in extracting the boundary information of tunnel water leakage. In the SPD decoder, an SPM is utilized to merge features at different scales, enhancing the representation of edge features. The specific convolution settings for each module in our proposed network are shown in Table 1.

3.1. Deep Aggregation Module

Different types of water leakage diseases vary significantly in morphology and scale. Therefore, enhancing the encoder’s ability to extract multiscale features is crucial to solving the problem of losing information on tunnel water leakage detection. The MCAM suppresses the noise in the shallow features and enhances the texture information of the high-level features by fusing the spatial and semantic information of different stages.
The structure of the DAM is illustrated in Figure 2. In the deep feature layers, the fourth-layer feature f 4 is first upsampled. The upsampling operation consists of two standard convolutions (with a kernel size of 3) and bilinear interpolation with a scale factor of 2. Then, f 4 is convolved with f 3 to obtain the fused features. This convolution operation includes a 1 × 1 convolution (with both kernel size and stride of 1), batch normalization (BN), and ReLU activation. Finally, f 3 is concatenated with the fused features to form a residual connection, preventing gradient vanishing and yielding the final output T 3 . This process is repeated for each layer in the encoder to obtain deep features T 2 . The computation process is described by the following formulas:
T 3 = t 3 + f 3 = C 3 1 f 3 , f 4 + f 3
where + denotes matrix addition, · indicates concatenation, ↑ represents upsampling, and C i j denotes the j-th convolution module in the i-th layer. Here, t i represents the fused features of f i and the upsampled f i + 1 of the i-th layer. The DAM fuses features from adjacent layers, where the deep feature is used to extract rich semantic information and the intermediate layer features contain edge texture information and semantic information. The network can extract the multiscale information of tunnel water leakage based on the fused features in an adaptive manner.
Specifically, as shown in Figure 2, we integrate the MCAM at the feature fusion stage of the first layer. We fuse f i + 1 with the upsampled f 2 , then the fused features are enhanced using the MCAM to obtain the fused feature t 1 1 . Subsequently, a second convolution operation and the MCAM are applied to further fuse f i + 1 , t 1 1 , and t 2 1 . Based on this, the final output T 1 is obtained. In this way, fused features with rich texture information and semantic information are achieved through continuous feature fusion operations.
Algorithmically, this can be defined as
t 1 1 = A C 1 1 f 1 , f 2 ,
t 1 2 = A C 1 2 f 1 , t 1 1 , t 2 1 ,
T 1 = t 1 3 + f 1 = C 1 3 f 1 , t 1 2 , t 2 2 + f 1 ,
where t 1 1 represents the feature map obtained by fusing f 1 and f 2 , T 1 denotes the output of the first layer, and A denotes the entire process of the MCAM. We propose the MCAM to suppress the noise in the shallow features and enhance the feature representation capability of the model. Additionally, the MCAM improves the connectivity structure of the network, allowing for more efficient utilization of the neurons in the neural network for enhanced computational efficiency and overall network performance.
As illustrated in Figure 3, the MCAM comprises multiple convolutional blocks of different scales along with spatial attention and channel attention mechanisms. The MCAM performs attention operations on the features at various scales in order to adaptively extract the relevant information while suppressing the noise. Specifically, the input features undergo convolution operations followed by 2 × and 4 × downsampling. These features are then fed into the channel attention mechanism at three different scales. Subsequently, a convolution operation is applied to integrate the multiscale features, obtaining the intermediate features. These intermediate features are downsampled by 2 × and 4 × and then processed by the spatial attention mechanism. The final output features are obtained through a convolution operation.
The process of channel attention is depicted in Figure 4a. Initially, the input features undergo global max pooling and global average pooling operations to obtain two different feature descriptor operators. These operators are then processed by a 1 × 1 convolution to perform dimensionality reduction and expansion, enhancing the feature representation in the attention map. Finally, the channel weights are normalized using the sigmoid function. The output feature map is produced by applying the channel weights to the input feature map through a residual connection.
The process of spatial attention is illustrated in Figure 4b. Initially, the feature map processed by channel attention undergoes max pooling and average pooling operations along the channel dimension. The two resulting feature tensors are concatenated and passed through a 1 × 1 convolution to adjust the number of channels. The spatial attention map is then normalized using the sigmoid function. Finally, the spatial attention operator is combined with the input feature map through a residual connection to produce the final output of the MCAM.
The MCAM enhances the network’s ability to perceive objects and resist interference by effectively integrating and filtering multiscale features. The MCAM can adaptively weight each feature of different scales using attention mechanisms, which improves the representational capability of the model. Furthermore, the MCAM captures detailed features of objects at various scales, enhancing the network’s detection accuracy and localization precision.

3.2. Edge Information Supplement Module

The interiors of tunnels often suffer from insufficient lighting, resulting in images with low contrast but significant noise. Additionally, diseases exhibit diverse shapes and appear on different materials and surfaces, making their visual characteristics varied and difficult to capture. Inadequate extraction of water leakage edge information leads to suboptimal segmentation results. To address this problem, we employ an EISM to enhance the extraction of edge information by directly supervising the exploration of edge semantic information.
In deep learning networks, shallow features contain extensive regional texture information, including edge interference; in contrast, deep features contain rich semantic information, which can be used to locate relevant edges within the shallow features. The designed EISM exploits the edge information from f 1 and f 4 . As shown in Figure 5, the structure of the edge information module is as follows: first, features from both layers are input into 1 × 1 convolutions to balance the proportion of high-level semantic information. The features of f 4 are upsampled to match the spatial resolution of f 2 . Then, the features from both layers are concatenated along the channel dimension and further integrated with a 3 × 3 convolution to enhance the semantic information of the feature map, which strengthens the edge information extraction capability. Finally, a 1 × 1 convolution followed by a sigmoid activation function produces the edge information segmentation result.

3.3. Sub-Pixel Decoder

In traditional encoder–decoder architectures, most methods rely on upsampling and skip connections to restore the resolution of the feature maps in the decoder. However, upsampling often leads to inevitable information loss. To address this problem, we propose an SPD to enhance feature representation and minimize information loss during feature fusion, which improves the accuracy of water stain segmentation.
The SPD is primarily composed of four sub-pixel convolution modules, as shown in Figure 6. Initially, D 4 is upsampled via sub-pixel operations to match the size of the third layer network. The T 3 features undergo a convolution for dimensionality reduction, which reduces the number of parameters and enhances the efficiency of the network. Subsequently, both features are fused after passing through separate 3 × 3 convolutions. The fused features are then added pixel-wise to the output from the sub-pixel upsampling. A final 3 × 3 convolution yields the decoder’s output D 3 . This process is repeated similarly to obtain the output D i for each subsequent layer in the decoder.
In algorithmic form, this is described as
E i , = C 3 C 1 E i ,
D i = C 3 C 3 S P M D i + 1 × E i , + S P M D i + 1 ) ,
where D i represents the output of the decoder, C 1 denotes the convolution operation with a kernel size of 1 × 1, E i , indicates the intermediate output, and S P M ( · ) represents the sub-pixel upsampling operation.
In conventional upsampling operations, bilinear interpolation is commonly used. This method calculates the target pixel value by taking the weighted average of the four nearest pixels in a 2 × 2 region around the corresponding source image position. However, bilinear interpolation has the drawback of causing blurring and distortion, especially in regions with high-frequency textures. This inadequacy can lead to insufficient extraction of tunnel water leakage boundary information. The reason for this is that bilinear interpolation only considers the local distribution of four neighboring pixels around the target pixel, neglecting the global information. As a result, when high-frequency textures surround the target pixel, bilinear interpolation can produce overly smooth results, leading to image distortion. In contrast, transposed convolution (deconvolution) upsampling often involves padding zeros in empty spaces when enlarging a small image, which introduces invalid information to the network.
In contrast to bilinear interpolation, sub-pixel upsampling performs convolution operations on the feature map and splits it into several groups along the channel dimension. The feature maps within each group are then rearranged to form a higher-resolution feature map. This method increases resolution while preserving more details, reducing information loss, and minimizing blurriness. In addition, it avoids the issues of high-frequency component damage and invalid padding, thereby enhancing the performance of semantic segmentation. The detailed process is as follows: for every four channels in the feature map, the channels are combined according to a specific pattern, resulting in a feature map that is twice the original size with one-fourth the number of channels, thereby achieving 2 × upsampling. This process is illustrated in Figure 7.

3.4. Loss Function

In pixel-level segmentation tasks, the cross-entropy loss is commonly used to measure the discrepancy between the predictions of the model and the ground truth labels. For binary classification tasks, such as determining whether each pixel belongs to tunnel water leakage or not, the binary cross-entropy loss is utilized to evaluate the model’s accuracy. The calculation process is as follows:
L b c e P , G = i = 1 H j = 1 W P i j l o g G i j + 1 P i j l o g G i j
where G i j and P i j represent the ground truth label and the predicted value at position ( i , j ) , respectively.
Unlike the cross-entropy loss function, the Intersection over Union ( I o U ) loss is primarily used to measure the overlap between the predicted segmentation map and the ground truth labels, thereby accelerating the network’s optimization. The calculation process is as follows:
L I o U P , G = 1 i = 1 H j = 1 W G i j , P i j i = 1 H j = 1 W G i j + P i j G i j × P i j .
The Structural Similarity (SSIM) loss is used to compare the Structural Similarity Index of two images. SSIM is a method for measuring the similarity between two images based on human visual perception, considering luminance, contrast, and structure. By calculating the SSIM index between two images, it is possible to assess their degree of similarity. The calculation process is as follows:
SSIM x , y = 2 μ x μ y + M 1 2 σ x y + M 2 μ x 2 + μ y 2 + M 1 σ x 2 + σ y 2 + M 2
where x and y represent the predicted result and the ground truth label, μ x is the mean value of x, μ y is the mean value of y, σ x 2 is the variance of x, σ y 2 is the variance of y, σ x y is the covariance between x and y, and C 1 and C 2 are constants used to stabilize the division, and are typically small values.
The overall loss function is calculated as follows:
L P , G = L b c e P , G + L I o U P , G + SSIM x , y .

4. Experiments

In this section, we present the experimental setup and discuss the comparative results of the proposed method against the current state-of-the-art algorithms.

4.1. Datasets

TWS Dataset: In this study, we created the TWS dataset, which was designed specifically for the task of tunnel water stain segmentation. This dataset aims to capture various instances of water stains within tunnels, providing a diverse set of training, validation, and testing samples to evaluate the effectiveness of our proposed method. The dataset details are shown in Table 2 and some samples are shown in Figure 8.

4.2. Evaluation Metrics

The segmentation of water stain within tunnels is fundamentally a pixel-level segmentation task; therefore, this study utilizes standard pixel-level evaluation metrics to assess the proposed method. These evaluation metrics include Accuracy ( A c c ), I o U , and F 1 -score [48,49,50,51]. A c c refers to the model’s pixel classification accuracy across all categories. The I o U measures the overlap between the predicted values of tunnel water stain and the ground truth labels, serving as a critical metric for evaluating pixel-level tasks. The F 1 -score is an important metric that balances the precision–recall relationship. The formulas for these metrics are as follows:
A c c = T P + T N T P + F N + F P + T N ,
I o U = P p P g P p P g ,
F 1 - score = 2 × P × R P + R ,
where True Positive ( T P ), True Negative ( T N ), False Positive ( F P ), and False Negative ( F N ) represent the counts of true positive, true negative, false positive, and false negative pixels, respectively, while P p and P g denote the respective pixel values of the prediction and ground truth.

4.3. Experimental Setup

The proposed method was implemented in Python3.8 using the PyTorch framework. The experiments were conducted on a computer running the Ubuntu operating system with the following configuration: an Intel i7-7800X CPU (Santa Clara, CA, USA) and two NVIDIA GeForce RTX 2080Ti GPUs (Santa Clara, CA, USA) with 11 GB of memory each. During training, the Adam optimizer was employed with an initial learning rate of 10 4 , a momentum of 0.9, and a weight decay of 10 5 . The batch size was set to 32. If the I o U did not improve for five consecutive epochs, the learning rate was reduced by a factor of 0.2 until it reached zero, at which point the training process was terminated. Due to GPU memory limitations, all images used in the experiments were resized to a size of 512 × 512 pixels. To enhance the randomness of the input images, data augmentation strategies such as random cropping and horizontal and vertical flipping were applied during preprocessing. All experiments were conducted on the TWS dataset.

4.4. Experimental Results and Discussion

To evaluate the performance of the proposed method, we compared it against several state-of-the-art methods. These methods included classic semantic segmentation networks such as U-Net [40], PANet [52], CCNet [53], DeepLabV3+ [54], SCDeepLab [55], ViT-Seg [47], TCDNet [37], and MTIS [38]. All experimental results were generated based on open-source model implementations or models provided by the authors. It is important to note that the implementation settings for our method were consistent with the experimental details of all reproduced networks, including data augmentation strategies.

4.4.1. Training Process

We focused on the variation of the loss function during the training process. By analyzing the loss curves, the convergence speed and stability of each model can be intuitively understood. As shown in Figure 9, the loss curves of each model during training reveal significant insights. DAEiS-Net demonstrates rapid convergence speed from the initial stages of training. Compared to other methods, its loss value decreases rapidly and reaches a stable state within a relatively small number of iterations. This indicates that DAEiS-Net effectively learns useful features from the training data, thereby accelerating the optimization process. Moreover, DAEiS-Net maintains the lowest loss value throughout the training process, further demonstrating the superiority of the model. A low loss value indicates a smaller error between the model’s predictions and the ground truth, suggesting that DAEiS-Net achieves high accuracy and robustness in handling the tunnel water stain segmentation task.
In contrast, the loss curves of the other six methods demonstrate slower convergence speeds, and display significant fluctuations even in the later stages of training. This suggests that these methods face greater optimization challenges and instability when applied to tunnel water stain segmentation. By comparing the loss curves during the training process, it is evident that DAEiS-Net significantly outperforms the other methods in terms of both convergence speed and loss value. These observations underscore the effectiveness of DAEiS-Net in quickly learning and optimizing features essential for accurate and reliable tunnel water stain segmentation, emphasizing its potential for practical applications in demanding environments.

4.4.2. Quantitative Comparison

In our experiments, we utilized common pixel-level metrics such as A c c , I o U , and F 1 -score to evaluate the proposed method. mAcc, mIoU and mF1 represent the mean values of A c c , I o U and F 1 -score respectively. The experimental results are presented in Table 3. Each row of the table represents the evaluation results of a different method, while each column corresponds to a specific evaluation metric.
The proposed model has a parameter count of 15.3 M, which is the lowest among all the listed models. In comparison, other models have parameter counts ranging from 25.8 M to 113.71 M, with the ViT-Seg model having the highest count at 113.71 M. A lower parameter count implies reduced computational resource requirements and faster inference speed, making the model suitable for deployment in resource-constrained environments. For the mean accuracy (mAcc) metric, the proposed model achieves an mAcc of 0.7923, which is second only to ViT-Seg’s mAcc of 0.8021. Despite having a lower parameter count, the proposed model achieves near-maximum accuracy, showcasing its strength in terms of precision. The proposed model also achieves a mean Intersection over Union (mIoU) of 0.7874, surpassing most other models. A higher mIoU indicates that the model excels in segmentation tasks by accurately predicting water stain regions. Additionally, the proposed model achieves an mF1-score of 0.77, matching DeepLabV3+, the highest among all models. This indicates that the proposed model not only effectively identifies water stain regions but also minimizes false positives.
The GAM in our model incorporates the MCAM to enhance feature extraction capabilities by integrating spatial and semantic information, reducing noise in shallow features, and enhancing texture extraction. This enhancement improves the model’s capability to detect water stains of different sizes. Furthermore, the proposed EISM bridges the semantic gap between different stages, effectively combining precise detailed information with rich semantic information to address deficiencies in extracting tunnel water stain boundary information. Despite having the fewest parameters, our model achieves near-optimal performance in mAcc and mIoU and attains the highest score in mF1, demonstrating its superiority in segmentation tasks. When compared to models with similar or even larger parameter counts, such as U-Net, PANet, and CCNet, the proposed model performs significantly better, highlighting its efficiency and effectiveness.

4.4.3. Qualitative Comparison

To further demonstrate the effectiveness of our method, we conducted a qualitative analysis of the results generated by classical semantic segmentation networks. Figure 10 shows examples of tunnel water stain extraction from our dataset. We selected representative samples, including tunnel water stains in complex environments and those with varying scales. The first and second rows of Figure 10 show the original images and their corresponding ground truth values, respectively. Rows (3)–(9) display the segmentation results from our proposed model and other classical models. Compared to the other methods, our proposed model demonstrates superior performance in extracting tunnel water stains from various environments, areas, and shapes.
As shown in Figure 10, when traditional segmentation models such as DeepLabV3+, U-Net, CCNet, and PANet are applied to the task of tunnel water stain segmentation, their performance is inadequate; specifically, DeepLabV3+ and U-Net exhibit significant segmentation errors over large areas. In contrast, Vit-Seg and SCDeepLab, which are specifically designed for tunnel defect segmentation, perform better at this task. They tend to overlook more detailed boundary and structural information when handling tunnel water stain segmentation, which could lead to a decrease in segmentation accuracy in practical applications. However, these models often overlook detailed boundary and structural information. In comparison, the proposed model introduces an EISM, which bridges the semantic gap between different stages and enhances the network’s sensitivity to edge information, resulting in optimal extraction performance. Moreover, the GAM integrated with the MCAM strengthens the network’s feature extraction capability by combining spatial and semantic information. This integration reduces noise in shallow features and improves texture information extraction. Additionally, the proposed SPD enhances the network’s feature integration capability within the decoder.
In summary, the superior overall performance of our model can be attributed to the application of the GAM and MCAM. These modules enhance the feature extraction capability by integrating spatial and semantic information, effectively reducing noise in shallow features while reinforcing texture information extraction. This comprehensive approach ensures better handling of edge information and multiscale feature extraction, making our model highly effective for tunnel water stain segmentation tasks.

4.4.4. Ablation Study

To further validate the impact of various modules in our proposed DAEiS-Net, we conducted an ablation study on the TWS dataset. It is important to highlight that compared to U-Net, our DAEiS-Net incorporates DAM, EISM, and SPD modules; the DAM includes the MCAM to enhance feature extraction in the encoder, the EISM is designed to improve the extraction of edges of tunnel water stains, and the SPD aims to enhance feature integration in the decoder. In order to systematically evaluate the effectiveness of each proposed module, we conducted several experiments. Initially, the baseline does not include the segmentation model based on the above functional modules.
As shown in Table 4, the baseline model has a parameter count of 13.2 M, achieving an mAcc of 0.7238, mIoU of 0.7001, and mF1 of 0.71. This baseline serves as the reference point for all improved models. Upon introducing the DAM module, the parameter count increases to 14.8 M, resulting in an mAcc improvement to 0.7652 and mIoU improvement to 0.7537, while mF1 remains at 0.71. This indicates that the DAM module significantly enhances the feature extraction capability of the encoder, thereby improving segmentation performance. With the inclusion of both the EISM and SPD modules in the baseline model the parameter count increases to 15.3 M, leading to an mAcc of 0.7856, mIoU of 0.7622, and mF1 of 0.75. Compared to solely introducing the DAM module, the inclusion of the EISM module notably enhances mAcc and mIoU, demonstrating its crucial role in improving tunnel water stain edge extraction. Introducing the DAM and SPD modules results in a parameter count of 15.3 M, resulting in an mAcc of 0.7731, mIoU of 0.7521, and mF1 of 0.74. Adding the SPD module alongside the DAM module enhances the feature integration capability of the decoder, significantly boosting mIoU and mF1 compared to solely introducing the DAM module. By sequentially introducing the DAM, EISM, and SPD modules on top of the baseline model, the parameter count remains at 15.3 M, with mAcc improving to 0.7923, mIoU to 0.7874, and mF1 to 0.77. Across all three metrics, the introduction of all modules results in performance improvements of 9.46%, 12.46%, and 8.45%, respectively.
The comprehensive integration of the DAM, EISM, and SPD modules in our proposed model leads to significant improvements across all performance metrics, with particularly noticeable enhancements in mIoU and mF1. These results indicate that the DAM module enhances feature extraction capability, the EISM module improves edge extraction performance, and the SPD module enhances feature integration capability. The combination of these three modules effectively addresses the challenges of multiscale boundary information loss and global information extraction in tunnel water stain segmentation tasks, outperforming both the baseline model and other individual or partial module combinations.

4.4.5. Feature Visualization

To further validate the effectiveness of DAEiS-Net, we conducted feature visualization experiments. These experiments intuitively demonstrate the model’s feature responses to different input images, helping to understand how the model identifies and segments areas with tunnel water stains. As illustrated in Figure 11, each row represents the processing of a specific input image, including the original image, segmentation result, and feature heatmap. The first column showcases the original tunnel water image used as input to the model. The second column displays the segmentation results produced by DAEiS-Net, with red areas indicating the recognized water stain regions. The third column presents the model’s feature responses to the input image, where darker colors (closer to blue) represent higher attention from the model to that region. Through analysis, we observed that DAEiS-Net accurately focuses its attention on water stain regions when the stain area is small, as depicted in Figure 11a,b. Conversely, when the stain area is large, as shown in Figure 11c,d, the model’s attention is more dispersed, indicating its strong global perception capability. The feature heatmap illustrates that the model can focus on the overall context when handling large stain areas, ensuring the complete identification of water stain regions. These findings suggest that DAEiS-Net not only possesses strong local feature extraction capabilities but also exhibits excellent global perception abilities. This allows it to flexibly adapt to different scales of water stain scenarios, thereby enhancing the overall performance of tunnel water stain segmentation tasks. In summary, through feature visualization experiments, DAEiS-Net demonstrates outstanding performance in tunnel water stain segmentation tasks, especially in focusing on key areas and handling large stain areas.
In summary, the method proposed in this paper significantly outperforms other relevant methods in addressing issues such as multiscale boundary information loss and global information extraction. Through innovative design of the model structure and optimization of feature extraction mechanisms, the proposed model demonstrates excellent performance in tunnel water stain segmentation tasks. It offers a more robust and efficient solution, highlighting its superiority in addressing challenges associated with tunnel water stain segmentation.

5. Conclusions

This paper proposes a novel tunnel water stain segmentation network architecture called Deep Aggregation Network with Edge Information Supplement (DAEiS-Net). Leveraging the classic encoder–decoder structure, the proposed network incorporates DAM and EISM, significantly enhancing the model’s feature representation capability and edge information extraction capacity. Additionally, the decoder part utilizes the SPD module to merge features from different scales, further enhancing the representation capability of edge features. Experimental results demonstrate that DAEiS-Net achieves satisfactory performance on a custom dataset of tunnel water stains. By effectively aggregating multiscale contextual information and extracting edge details, DAEiS-Net successfully tackles the key challenges in tunnel water stain segmentation, demonstrating its significant potential for practical applications.
Future work will focus on further optimizing this model and exploring its application and extension in other related fields. This includes refining the network architecture, improving computational efficiency, and adapting the model to different types of tunnel diseases and other segmentation tasks. Additionally, expanding the dataset to include more diverse tunnel environments and conditions will help to validate and enhance the robustness of DAEiS-Net.

Author Contributions

Methodology, Y.W.; Software, Y.W.; Investigation, Y.W. and K.H.; Data curation, K.H.; Writing—original draft, Y.W. and K.H.; Writing—review & editing, K.Z. and S.L.; Visualization, Y.W. and K.H.; Supervision, K.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by National Natural Science Foundation of China (51905065)and the Science and Technology Research Program of Chongqing Municipal Education Commission (Grant No. KJQN202000614, KJQN202000613).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors wish to acknowledge Sebastian Bader of Mid Sweden University for his significant contributions of improving the quality of this paper.

Conflicts of Interest

Author Yuliang Wang was employed by the Beijing Metro Construction Administration Co., Ltd. Author Kai Huang was employed by the Beijing MTR Corporation, Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Xu, Y.; Chen, X. The spatial vitality and spatial environments of urban underground space (UUS) in metro area based on the spatiotemporal analysis. Tunn. Undergr. Space Technol. 2022, 123, 104401. [Google Scholar] [CrossRef]
  2. Huang, M.; Ninić, J.; Zhang, Q. BIM, machine learning and computer vision techniques in underground construction: Current status and future perspectives. Tunn. Undergr. Space Technol. 2021, 108, 103677. [Google Scholar] [CrossRef]
  3. Tan, L.; Hu, X.; Tang, T.; Yuan, D. A lightweight metro tunnel water leakage identification algorithm via machine vision. Eng. Fail. Anal. 2023, 150, 107327. [Google Scholar] [CrossRef]
  4. Tan, L.; Chen, X.; Hu, X.; Tang, T. DMDSNet: A Computer Vision-based Dual Multi-task Model for Tunnel Bolt Detection and Corrosion Segmentation. In Proceedings of the 2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC), Bilbao, Spain, 24–28 September 2023; pp. 4827–4833. [Google Scholar]
  5. Ren, Q.; Wang, Y.; Xu, J. A DL method to detect multi-type hidden objects in tunnel linings using a comprehensive GPR dataset. Measurement 2024, 238, 115379. [Google Scholar] [CrossRef]
  6. Zhang, Y.; Bader, S.; Oelmann, B. A Lightweight Convolutional Neural Network Model for Concrete Damage Classification using Acoustic Emissions. In Proceedings of the 2022 IEEE Sensors Applications Symposium (SAS), Sundsvall, Sweden, 1–3 August 2022; pp. 1–6. [Google Scholar]
  7. Arena, A.; Delle Piane, C.; Sarout, J. A new computational approach to cracks quantification from 2D image analysis: Application to micro-cracks description in rocks. Comput. Geosci. 2014, 66, 106–120. [Google Scholar] [CrossRef]
  8. Hu, X.; Cao, Y.; Sun, Y.; Tang, T. Railway automatic switch stationary contacts wear detection under few-shot occasions. IEEE Trans. Intell. Transp. Syst. 2021, 23, 14893–14907. [Google Scholar] [CrossRef]
  9. Adın, V.; Zhang, Y.; Oelmann, B.; Bader, S. Tiny Machine Learning for Damage Classification in Concrete Using Acoustic Emission Signals. In Proceedings of the 2023 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), Kuala Lumpur, Malaysia, 22–25 May 2023; pp. 1–6. [Google Scholar]
  10. Falls, S.D.; Young, R.P. Acoustic emission and ultrasonic-velocity methods used to characterise the excavation disturbance associated with deep tunnels in hard rock. Tectonophysics 1998, 289, 1–15. [Google Scholar] [CrossRef]
  11. Zhang, Y.; Adin, V.; Bader, S.; Oelmann, B. Leveraging Acoustic Emission and Machine Learning for Concrete Materials Damage Classification on Embedded Devices. IEEE Trans. Instrum. Meas. 2023, 72, 2525108. [Google Scholar] [CrossRef]
  12. Zhang, C.; Wang, R.; Yu, L.; Xiao, Y.; Guo, Q.; Ji, H. Localization of cyclostationary acoustic sources via cyclostationary beamforming and its high spatial resolution implementation. Mech. Syst. Signal Process. 2023, 204, 110718. [Google Scholar] [CrossRef]
  13. Ren, Q.; Wang, Y.; Xu, J.; Hou, F.; Cui, G.; Ding, G. REN-GAN: Generative adversarial network-driven rebar clutter elimination network in GPR image for tunnel defect identification. Expert Syst. Appl. 2024, 255, 124395. [Google Scholar] [CrossRef]
  14. Lee, C.H.; Chiu, Y.C.; Wang, T.T.; Huang, T.H. Application and validation of simple image-mosaic technology for interpreting cracks on tunnel lining. Tunn. Undergr. Space Technol. 2013, 34, 61–72. [Google Scholar] [CrossRef]
  15. Huang, C.; Sun, X.; Zhang, Y. Tiny-Machine-Learning-Based Supply Canal Surface Condition Monitoring. Sensors 2024, 24, 4124. [Google Scholar] [CrossRef]
  16. Xie, Y.; Nie, Y.; Lundgren, J.; Yang, M.; Zhang, Y.; Chen, Z. Cervical Spondylosis Diagnosis Based on Convolutional Neural Network with X-ray Images. Sensors 2024, 24, 3428. [Google Scholar] [CrossRef]
  17. Schabowicz, K. Ultrasonic tomography–The latest nondestructive technique for testing concrete members—Description, test methodology, application example. Arch. Civ. Mech. Eng. 2014, 14, 295–303. [Google Scholar] [CrossRef]
  18. Ying, W.; Zheng, J.; Huang, W.; Tong, J.; Pan, H.; Li, Y. Order-frequency Holo-Hilbert spectral analysis for machinery fault diagnosis under time-varying operating conditions. ISA Trans. 2024, 16, 472–483. [Google Scholar] [CrossRef] [PubMed]
  19. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 60, 84–90. [Google Scholar] [CrossRef]
  20. Hu, X.; Tan, L.; Tang, T. M2BIST-SPNet: RUL prediction for railway signaling electromechanical devices. J. Supercomput. 2024, 80, 16744–16774. [Google Scholar] [CrossRef]
  21. Zhang, X.; Gong, Y.; Li, Z.; Gao, X.; Jin, D.; Li, J.; Liu, H. SkipcrossNets: Adaptive Skip-cross Fusion for Road Detection. arXiv 2023, arXiv:2308.12863. [Google Scholar]
  22. Li, Y.; Zhu, Z.; Li, Y.; Zhang, J.; Li, X.; Shang, S.; Zhu, D. CTMU-Net: An Improved U-Net for Semantic Segmentation of Remote-Sensing Images Based on the Combined Attention Mechanism. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 10148–10161. [Google Scholar] [CrossRef]
  23. Niu, P.; Gu, J.; Zhang, Y.; Zhang, P.; Cai, T.; Xu, W.; Han, J. MDCGA-Net: Multi-Scale Direction Context-Aware Network with Global Attention for Building Extraction from Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 8461–8476. [Google Scholar] [CrossRef]
  24. Gong, Y.; Jiang, X.; Wang, L.; Xu, L.; Lu, J.; Liu, H.; Lin, L.; Zhang, X. TCLaneNet: Task-Conditioned Lane Detection Network Driven by Vibration Information. IEEE Trans. Intell. Veh. 2024, 1–14. [Google Scholar] [CrossRef]
  25. Fang, Z.; Zhang, J.; Yao, L.; Wang, H.; Wang, H. Improved spectral amplitude modulation for tacholess estimation of rotation speed. Measurement 2024, 236, 115056. [Google Scholar] [CrossRef]
  26. Liu, Y.; Huo, M.; Li, Q.; Zhao, H.; Xue, Y.; Yang, J.; Qi, N. Imbalanced Source-Free Adaptation Diagnosis for Rotating Machinery. IEEE Trans. Instrum. Meas. 2024, 73, 1–11. [Google Scholar] [CrossRef]
  27. Zhang, Y. Tiny Machine Learning for Structural Health Monitoring with Acoustic Emissions. Licentiate Thesis, Mid Sweden University, Sundsvall, Sweden, 2024. [Google Scholar]
  28. Yao, L.; Wang, H.; Tao, L.; Fang, Z.; Wang, H.; Liu, Y.; Wang, H. Bearing fault diagnosis based on transfer learning with dual-flow manifold ResNet and improved CapsNet. Meas. Sci. Technol. 2024, 35, 076123. [Google Scholar] [CrossRef]
  29. Hu, X.; Zhang, X.; Wang, Z.; Chen, Y.; Xia, J.; Du, Y.; Li, Y. Railway Switch Machine Fault Diagnosis Considering Sensor Abnormality Scenarios. In Proceedings of the 2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC), Bilabo, Spain, 24–28 September 2023; pp. 4834–4839. [Google Scholar]
  30. Liu, Y.; Han, J.; Zhang, Q.; Shan, C. Deep salient object detection with contextual information guidance. IEEE Trans. Image Process. 2019, 29, 360–374. [Google Scholar] [CrossRef] [PubMed]
  31. Cha, Y.J.; Choi, W.; Büyüköztürk, O. Deep learning-based crack damage detection using convolutional neural networks. Comput. Aided Civ. Infrastruct. Eng. 2017, 32, 361–378. [Google Scholar] [CrossRef]
  32. Zhou, Z.; Zhang, J.; Gong, C. Automatic detection method of tunnel lining multi-defects via an enhanced You Only Look Once network. Comput.-Aided Civ. Infrastruct. Eng. 2022, 37, 762–780. [Google Scholar] [CrossRef]
  33. Li, R.; Yuan, Y.; Zhang, W.; Yuan, Y. Unified vision-based methodology for simultaneous concrete defect detection and geolocalization. Comput. Aided Civ. Infrastruct. Eng. 2018, 33, 527–544. [Google Scholar] [CrossRef]
  34. Li, Y.; Bao, T.; Li, T.; Wang, R. A robust real-time method for identifying hydraulic tunnel structural defects using deep learning and computer vision. Comput. Aided Civ. Infrastruct. Eng. 2023, 38, 1381–1399. [Google Scholar] [CrossRef]
  35. Liu, Y.; Dong, X.; Zhang, D.; Xu, S. Deep unsupervised part-whole relational visual saliency. Neurocomputing 2024, 563, 126916. [Google Scholar] [CrossRef]
  36. Chun, P.j.; Izumi, S.; Yamane, T. Automatic detection method of cracks from concrete surface imagery using two-step light gradient boosting machine. Comput. Aided Civ. Infrastruct. Eng. 2021, 36, 61–72. [Google Scholar] [CrossRef]
  37. Zhou, Q.; Qu, Z.; Li, Y.X.; Ju, F.R. Tunnel crack detection with linear seam based on mixed attention and multiscale feature fusion. IEEE Trans. Instrum. Meas. 2022, 71, 5014711. [Google Scholar] [CrossRef]
  38. Liao, J.; Yue, Y.; Zhang, D.; Tu, W.; Cao, R.; Zou, Q.; Li, Q. Automatic tunnel crack inspection using an efficient mobile imaging module and a lightweight CNN. IEEE Trans. Intell. Transp. Syst. 2022, 23, 15190–15203. [Google Scholar] [CrossRef]
  39. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  40. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. pp. 234–241. [Google Scholar]
  41. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  42. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected crfs. arXiv 2014, arXiv:1412.7062. [Google Scholar]
  43. Islam, A.; Lundell, B.; Sawhney, H.; Sinha, S.N.; Morales, P.; Radke, R.J. Self-supervised learning with local contrastive loss for detection and semantic segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 5624–5633. [Google Scholar]
  44. Yang, L.; Song, S.; Fan, J.; Huo, B.; Li, E.; Liu, Y. An automatic deep segmentation network for pixel-level welding defect detection. IEEE Trans. Instrum. Meas. 2021, 71, 5003510. [Google Scholar] [CrossRef]
  45. Zhou, Z.; Yan, L.; Zhang, J.; Zheng, Y.; Gong, C.; Yang, H.; Deng, E. Automatic segmentation of tunnel lining defects based on multiscale attention and context information enhancement. Constr. Build. Mater. 2023, 387, 131621. [Google Scholar] [CrossRef]
  46. Chu, H.; Wang, W.; Deng, L. Tiny-Crack-Net: A multiscale feature fusion network with attention mechanisms for segmentation of tiny cracks. Comput. Aided Civ. Infrastruct. Eng. 2022, 37, 1914–1931. [Google Scholar] [CrossRef]
  47. Qin, S.; Qi, T.; Deng, T.; Huang, X. Image segmentation using Vision Transformer for tunnel defect assessment. Comput. Aided Civ. Infrastruct. Eng. 2024. [Google Scholar] [CrossRef]
  48. Yang, J.; Wang, Z.; Guo, Y.; Gong, T.; Shan, Z. A novel noise-aided fault feature extraction using stochastic resonance in a nonlinear system and its application. IEEE Sensors J. 2024, 24, 11856–11866. [Google Scholar] [CrossRef]
  49. He, C.; Huo, X.; Zhu, C.; Chen, S. Minimum Redundancy Maximum Relevancy-based Multiview Generation for Time Series Sensor Data Classification and Its Application. IEEE Sensors J. 2024, 24, 12830–12839. [Google Scholar] [CrossRef]
  50. He, C.; Huo, X.; Zhu, C.; Chen, S. Feature Selection-based Multi-view Concentration for Multivariate Time Series Classification and Its Application. IEEE Sensors J. 2023, 24, 4798–4806. [Google Scholar] [CrossRef]
  51. He, W.; Mao, J.; Li, Z.; Wang, Y.; Fang, Q.; Wu, H. Fault identification of rotating machinery based on dynamic feature reconstruction signal graph. IEEE/ASME Trans. Mechatronics 2023, 29, 2056–2066. [Google Scholar] [CrossRef]
  52. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  53. Huang, Z.; Wang, X.; Huang, L.; Huang, C.; Wei, Y.; Liu, W. Ccnet: Criss-cross attention for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 603–612. [Google Scholar]
  54. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  55. Zhou, Z.; Zhang, J.; Gong, C. Hybrid semantic segmentation for tunnel lining cracks based on Swin Transformer and convolutional neural network. Comput. Aided Civ. Infrastruct. Eng. 2023, 38, 2491–2510. [Google Scholar] [CrossRef]
Figure 1. The schematic diagram of the DAEiS-Net structure, which is composed of encoding blocks, SPD blocks, EISM, and DAM.
Figure 1. The schematic diagram of the DAEiS-Net structure, which is composed of encoding blocks, SPD blocks, EISM, and DAM.
Sensors 24 05452 g001
Figure 2. The schematic diagram of the DAM structure. The fused features are obtained through the upsampling operation. The final output is obtained through the MCAM.
Figure 2. The schematic diagram of the DAM structure. The fused features are obtained through the upsampling operation. The final output is obtained through the MCAM.
Sensors 24 05452 g002
Figure 3. The schematic diagram of the MCAM structure. The MCAM comprises multiple convolutional blocks of different scales along with channel attention and spatial attention mechanisms.
Figure 3. The schematic diagram of the MCAM structure. The MCAM comprises multiple convolutional blocks of different scales along with channel attention and spatial attention mechanisms.
Sensors 24 05452 g003
Figure 4. The structure of channel attention (a) and spatial attention (b). The attention operators are obtained through max pooling and the global average pooling.
Figure 4. The structure of channel attention (a) and spatial attention (b). The attention operators are obtained through max pooling and the global average pooling.
Sensors 24 05452 g004
Figure 5. The structure diagram of the edge information module. The fused features are concatenated along the channel dimension.
Figure 5. The structure diagram of the edge information module. The fused features are concatenated along the channel dimension.
Sensors 24 05452 g005
Figure 6. The SPD structure diagram; D i is updated by the SPM and convolution operation.
Figure 6. The SPD structure diagram; D i is updated by the SPM and convolution operation.
Sensors 24 05452 g006
Figure 7. The process of subpixel upsampling. The channels are combined according to a specific pattern, resulting in a feature map that is twice the original size with one-fourth the number of channels.
Figure 7. The process of subpixel upsampling. The channels are combined according to a specific pattern, resulting in a feature map that is twice the original size with one-fourth the number of channels.
Sensors 24 05452 g007
Figure 8. Different types of tunnel water stain samples in our dataset.
Figure 8. Different types of tunnel water stain samples in our dataset.
Sensors 24 05452 g008
Figure 9. The loss curves during the training process; the horizontal axis represents the epoch, while the vertical axis represents the loss value.
Figure 9. The loss curves during the training process; the horizontal axis represents the epoch, while the vertical axis represents the loss value.
Sensors 24 05452 g009
Figure 10. Qualitative analysis of the results. The first row shows the original images, the second row shows their corresponding ground truth values, and rows (3)–(9) show the segmentation results of our proposed model and the other classical models. The purple part is the segmentation area, and the yellow dotted line box represents the key contrast part of different methods.
Figure 10. Qualitative analysis of the results. The first row shows the original images, the second row shows their corresponding ground truth values, and rows (3)–(9) show the segmentation results of our proposed model and the other classical models. The purple part is the segmentation area, and the yellow dotted line box represents the key contrast part of different methods.
Sensors 24 05452 g010
Figure 11. Visualization results of the feature map. Each row (ad) represents the processing of different samples. The first column showcases the original tunnel water image that is fed to the model. The second column displays the segmentation results produced by DAEiS-Net, with red areas indicating the recognized water stain regions. The third column presents the model’s feature responses to the input image, with darker colors (closer to blue) representing higher attention of the model on that region.
Figure 11. Visualization results of the feature map. Each row (ad) represents the processing of different samples. The first column showcases the original tunnel water image that is fed to the model. The second column displays the segmentation results produced by DAEiS-Net, with red areas indicating the recognized water stain regions. The third column presents the model’s feature responses to the input image, with darker colors (closer to blue) representing higher attention of the model on that region.
Sensors 24 05452 g011
Table 1. The specific convolutional settings of the network structure.
Table 1. The specific convolutional settings of the network structure.
ModuleSpecific Convolutional Settings
Encoding Block 1 [7 × 7, 64, s = 2 ]; [3 × 3, MaxPooling, s = 2 ]
Encoding Block 2 [3 × 3, 64] × 3
Encoding Block 3 [3 × 3, 128] × 4
Encoding Block 4 [3 × 3, 256] × 6
Encoding Block 5 [3 × 3, 512] × 3
DAM [1 × 1, 256]; [1 × 1, 128] × 2; [1 × 1, 64] × 3
MCAM[1 × 1, AvgPooling]; [1 × 1, MaxPooling]; [1 × 1, 64] ×2
EISM [1 × 1, 64]; [1 × 1, 128]; [3 × 3, 64]; [1 × 1, 64]
Table 2. Introduction to the TWS dataset.
Table 2. Introduction to the TWS dataset.
ItemDescription
Dataset NameTWS (Tunnel Water Stain)
Dataset PurposeTunnel water stain segmentation
Training set: 2000 images
Dataset SizeValidation set: 400 images
Test set: 100 images
Image Resolution2330 × 1747; 1944 × 2592
Scenarios and ConditionsCaptures various scenarios and conditions of tunnel water leakage
AnnotationsEach image has been meticulously annotated,
detailing the contours and locations of water leakage areas
Provides valuable training data for the model,
Annotation Purposehelping it learn about various shapes,
and sizes of water stain features
Table 3. Experimental results of different models.
Table 3. Experimental results of different models.
ModelPara (M)mAccmIoUmF1
UNet33.20.57090.56080.54
PANet34.90.64820.66490.62
CCNet40.20.76060.75560.68
Deeplabv325.80.77720.77390.77
SCDeepLab42.30.78220.72740.73
ViT-Seg113.710.80210.73370.72
TCDNet140.10.70830.68170.65
MTISs19.00.72280.70050.71
DAEiS-Net (Ours)15.30.79230.78740.77
Table 4. Results of the ablation experiments.
Table 4. Results of the ablation experiments.
BaselineDAMEISMSPDPara (M)mAccGain (%)mIoUGain (%)mF1Gain (%)
---13.20.7238/0.7001/0.71/
--14.80.7652+5.720.7537+7.650.710
-15.30.7792+7.650.7418+5.960.73+2.82
-15.30.7856+8.540.7622+8.870.75+5.63
-15.30.7731+6.810.7521+7.430.74+4.23
15.30.7923+9.460.7874+12.460.77+8.45
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Huang, K.; Zheng, K.; Liu, S. DAEiS-Net: Deep Aggregation Network with Edge Information Supplement for Tunnel Water Stain Segmentation. Sensors 2024, 24, 5452. https://doi.org/10.3390/s24175452

AMA Style

Wang Y, Huang K, Zheng K, Liu S. DAEiS-Net: Deep Aggregation Network with Edge Information Supplement for Tunnel Water Stain Segmentation. Sensors. 2024; 24(17):5452. https://doi.org/10.3390/s24175452

Chicago/Turabian Style

Wang, Yuliang, Kai Huang, Kai Zheng, and Shuliang Liu. 2024. "DAEiS-Net: Deep Aggregation Network with Edge Information Supplement for Tunnel Water Stain Segmentation" Sensors 24, no. 17: 5452. https://doi.org/10.3390/s24175452

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop