Next Article in Journal
SoftVein-WELM: A Weighted Extreme Learning Machine Model for Soft Biometrics on Palm Vein Images
Next Article in Special Issue
An AIoT-Based Assistance System for Visually Impaired People
Previous Article in Journal
Neural Machine Translation of Electrical Engineering Based on Integrated Convolutional Neural Networks
Previous Article in Special Issue
Efficient Reversible Data Hiding Using Two-Dimensional Pixel Clustering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AI-Driven High-Precision Model for Blockage Detection in Urban Wastewater Systems

by
Ravindra R. Patil
1,*,
Rajnish Kaur Calay
1,
Mohamad Y. Mustafa
1 and
Saniya M. Ansari
2
1
Faculty of Engineering Science and Technology, UiT the Arctic University of Norway, 8514 Narvik, Norway
2
Department of E & TC Engineering, Ajeenkya D Y Patil School of Engineering, Pune 411047, India
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(17), 3606; https://doi.org/10.3390/electronics12173606
Submission received: 10 August 2023 / Revised: 22 August 2023 / Accepted: 24 August 2023 / Published: 26 August 2023
(This article belongs to the Special Issue Advances of Artificial Intelligence and Vision Applications)

Abstract

:
In artificial intelligence (AI), computer vision consists of intelligent models to interpret and recognize the visual world, similar to human vision. This technology relies on a synergy of extensive data and human expertise, meticulously structured to yield accurate results. Tackling the intricate task of locating and resolving blockages within sewer systems is a significant challenge due to their diverse nature and lack of robust technique. This research utilizes the previously introduced “S-BIRD” dataset, a collection of frames depicting sewer blockages, as the foundational training data for a deep neural network model. To enhance the model’s performance and attain optimal results, transfer learning and fine-tuning techniques are strategically implemented on the YOLOv5 architecture, using the corresponding dataset. The outcomes of the trained model exhibit a remarkable accuracy rate in sewer blockage detection, thereby boosting the reliability and efficacy of the associated robotic framework for proficient removal of various blockages. Particularly noteworthy is the achieved mean average precision (mAP) score of 96.30% at a confidence threshold of 0.5, maintaining a consistently high-performance level of 79.20% across Intersection over Union (IoU) thresholds ranging from 0.5 to 0.95. It is expected that this work contributes to advancing the applications of AI-driven solutions for modern urban sanitation systems.

1. Introduction

Computer vision is a field of artificial intelligence (AI) with its own conventional algorithms that extract required information from various visual forms such as photos and videos, and based on that information form, perform actions, or make recommendations in order to detect and identify distinct objects. Thus, the large datasets should increase the performance properties of computer vision.
Object detection techniques of computer vision detect the occurrence of objects in an image or video with bounding boxes and identify their classes. Initially, machine learning was mainly used for object detection tasks but when deep neural networks, i.e., deep learning methods emerged, they became popular due to automatic representative feature extraction from large datasets for training purposes [1]. Occlusion, clutter, and low resolution are some of the sub-problems that are handled very efficiently by deep learning-based detection frameworks [2,3]. It has two method types such as single-stage, which works for inference speed and real-time use, and two-stage, which works for model performance, i.e., detection accuracy. The single-stage detectors remove the process of region of interest (ROI) extraction and moves for classification and regression whereas two-stage detectors extract ROI and then apply classification and regression. The YOLO detection model (YOLOv2 [4], YOLOv3 [5], YOLOv4 [6], and YOLOv5 [7]), SSD [8], CenterNet [9], CornerNet [10], etc., are some single stage detectors. Region proposal models (R-CNN [11], Fast-RCNN [12], Faster RCNN [13], Cascade R-CNN [14], and R-FCN [15]) are two-stage detectors. Classification and localization accuracy and inference speed are two important metrics for object detectors. In the advancement of detection models, transfer learning techniques with quality datasets meet the requirements with a minimum training time [16,17]. Transfer learning harnesses prior knowledge to enhance performance on novel tasks. By fine-tuning, pre-trained deep neural models are adapted to new contexts with certain layers preserved and others refined. This leads to many advantages such as achieving quick convergence, good performance, and adaptability in real-world scenarios. As the applications of AI evolve, such as video surveillance, military applications, security aspects, health monitoring, and critical detection tasks, the AI techniques are being enhanced to suit these needs.
Addressing the application-based needs to produce sensible and accurate results, detection models need to be adapted and modified, which usually have heavy computational demands. However, there are methods such as the embedded vision approach with AI that has an ability to enable real-time, efficient, and intelligent visual processing directly on edge devices, which reduces dependency on cloud computing and enhances privacy and responsiveness in many applications [18,19].
Detecting various sewer blockages is a major challenge due to their complex and heterogeneous nature. Moreover, their locations in the sewer network may vary, including main lines, lateral connections, and junctions. Blockages can exhibit varying levels of severity, from partial restrictions that gradually reduce flow to complete blockages that cause sewer overflows. The dynamic and unpredictable nature of urban wastewater systems, influenced by factors such as climate, wastewater composition, and hydraulic conditions adds another layer of complexity. In this research work, transfer learning and fine-tuning techniques are utilized to achieve a high precision rate in the detection of blockages within urban wastewater systems. This approach is intended for real-time implementation on mobile devices and other environments with limited resources, with the goal of effectively removing such blockages. Our primary emphasis is on the training of the single-stage YOLOv5 model using the S-BIRD dataset [20,21], which contains representative and critical multi-class images depicting prevalent sewer blockage scenarios.
The study implements all computer vision and model training procedures using Python programming, OpenCV, PyTorch framework, and other machine learning libraries. These operations are carried out on a DGX GPU workstation system running on the Linux platform, ensuring a robust and efficient experimental environment. The results are analyzed and discussed to demonstrate the effectiveness of the methodology used.

2. Structural Insights of YOLOv5 Model

YOLOv5 is an anchor-based single-stage detection model, which is built on the PyTorch framework. It focuses on simplicity, model scaling, and transfer learning, making it versatile for a wide range of object detection tasks. The model’s backbone is CSP Darknet-53, which incorporates Cross Stage Partial (CSP) connections to enhance information flow and feature representation.
To create feature pyramids for effective object scaling and generalization, YOLOv5 employs the Path Aggregation Network (PAN) as its neck. The head design utilizes anchor boxes to generate output vectors that contain class probabilities, objectness scores, and bounding box coordinates (center_x, center_y, height, and width). The model parameters are updated during training using the following loss function:
L o s s = λ 1 L _ c l s + λ 2 L _ o b j + λ 3 L _ l o c
where L_cls represents the Binary Cross Entropy loss for predicted classes, L_obj represents the Binary Cross Entropy loss for objectness scores, and L_loc represents the Complete Intersection over Union loss for bounding box locations. Here, λ1, λ2, and λ3 are hyperparameters controlling the contribution of each component to the overall loss. The employed auto anchor automatically determines and generates anchor boxes based on the distribution of bounding boxes in the custom dataset using K-means clustering and a genetic learning algorithm. In this, SiLU (Sigmoid Linear Unit) activation function in hidden layers acquire intricate details and Sigmoid activation function in the output layer functions for binary classification.
As shown in Figure 1, the backbone employs Convolutional and C3 layers to extract image features, which are then combined at various levels using Conv, Upsample, Concat, and C3 layers in the head. The object detection process is facilitated by a Detect layer that uses anchor boxes and the indicated class count. Particularly, each C3 (CSP-3) block consists of two parallel convolutional layers, the first layer channels input features through a bottleneck layer, compressing the information and the second layer directly outputs feature. These streams are then concatenated and processed through pooling and convolutional layers. The C3 blocks also use skip connections and attention mechanisms to enhance information flow and reduce noisy features.

3. Details of Training Instances in Critical Multi-Class S-BIRD

The dataset comprises a total of 14,765 training frames of classes (grease, plastics, and tree roots), which are meticulously annotated with 69,061 objects as shown in Figure 2, resulting in an average of 4.7 annotations per frame. Specifically, the dataset comprises 26,847 annotations for grease, 21,553 annotations for tree roots, and 20,661 annotations for plastics. To ensure uniformity and standardization, the frames were preprocessed and augmented, resulting in an average frame size of 0.173 Megapixels. The frames were resized to a square aspect ratio of 416 × 416 pixels, thereby maintaining a 1:1 aspect ratio class. The angle of the diagonal was calculated to be 0.785 radians (equivalent to 45 degrees), with the diagonal length measuring 588 pixels.
Regarding pixel density, the dataset exhibits a density of 12 pixels per millimeter or 290 pixels per inch. These specific computational details are vital for understanding the characteristics and intricacies of the S-BIRD dataset, which plays a crucial role in effectively training the deep neural network. Figure 3 illustrates the distribution of object classes in each training frame based on the center x for the S-BIRD dataset. Figure 3 shows the relative distribution of center x coordinates across different classes during training. Each segment is color-coded and displays data values and percentiles, providing a clear understanding of object positions along the x-axis. This section provides valuable insights into the dataset’s dimensions, resolutions, and geometric properties, which contribute to the successful implementation of transfer learning and fine-tuning techniques for the deep neural detection model.

4. Training Method and Evaluation

The training process for the YOLOv5-s model (Based on PyTorch 1.10.0a0 with CUDA support) on the S-BIRD dataset involved a series of steps aimed at achieving the highest precision in detecting sewer blockages. Through the application of transfer learning and fine-tuning techniques, the model’s formulation was optimized to suit the specific characteristics of the representative dataset, enabling its effective adaptation for real-world scenarios. To facilitate the training process, annotations for object classes were applied in PyTorch TXT format, as needed. The training process was performed over 6000 epochs, using the stochastic gradient descent (SGD) optimizer with specified hyperparameters. The training process utilized the configurations listed in Table 1. The DGX-1 (utilized 32 GB GPU Card) available at UiT, Narvik, running a Docker container with a defined image served as the training platform, leveraging GPU parallelization for faster computations. Overfitting was mitigated using Early Stopping with a patience of 100 epochs.
The training progression concluded at 933 epochs due to a lack of improvement in the last 100 epochs. The most promising results were obtained at epoch 832, leading to the selection of the corresponding model for practical applications. The evaluation metrics are essential for quantifying the model’s performance, and they are computed using the following formulas:
P r e c i s i o n = T P / ( T P + F P )
R e c a l l = T P / ( T P + F N )
m A P = ( A P   f o r   e a c h   c l a s s ) / N u m b e r   o f   c l a s s e s
F 1   s c o r e = 2 ( P r e c i s i o n R e c a l l ) / ( P r e c i s i o n + R e c a l l )
Here, TP—true positive, FP—false positive, FN—false negative, and mAP—mean average precision.
During the training, at epoch 832, the model exhibited impressive precision (P) and recall (R) values of 94.40% and 93.90%, respectively, across all classes. Notably, Figure 4 illustrates that the developed detection model achieved outstanding average precision values of 95.90% for grease blocks, 98.40% for plastic blocks, and 94.50% for tree root blocks. These high precision values are indicative of the model’s ability to accurately detect and classify instances belonging to these specific classes. The overall mean average precision (mAP) for all classes, as indicated in Table 2, is remarkably high at 96.30% with a confidence threshold of 0.5. This highlights the model’s proficiency in making precise detections across all classes within the dataset. Moreover, the calculated mAP over various Intersection over Union (IoU) thresholds, ranging from 0.5 to 0.95 with an increment of 0.05, yielded a consistent performance of 79.20%. This demonstrates that the model maintains accurate localization of objects across a broad range of IoU thresholds. The timing results in Table 3 show that the model has efficient inference times, with an average forward time of 0.2 ms, average NMS time of 1.1 ms, and average inference time of 11 ms. These low inference times make the model suitable for real-time applications.
The confusion matrix in Figure 5, provides an overview of the model’s performance in correctly classifying instances of grease, plastic, and tree roots. This visualization provides a clear breakdown of correct and incorrect classifications for each category.
Figure 6 shows correlation connections within the frames of the dataset, demonstrating the exact connection between instances and their labels among discrete views. It is also evident that a majority of instances in the dataset are situated towards the outer edges of both the top and bottom sides of the images in the dataset. This indicates the efficiency of the trained model to detect and classify multiple objects in various real-world scenarios.
The scatter diagram, Figure 7, displays the instances in the dataset and their corresponding labels. This visualization helps with understanding the distribution of instances across different classes and assists with identifying potential clustering patterns.
The graph in Figure 8 illustrates the relationship between precision (P) and confidence (C) that informs concerning changes in the model’s precision at different confidence levels, providing insights into the model’s ability to make accurate detections at various confidence thresholds.
Figure 9 displays the correlation between recall (R) and confidence (C), which clarifies how well the model can recall positive instances at different confidence levels, giving sensitivity details to detection of true positives.
Figure 10 showcases the mean average precision (mAP) of the model, comparing the truth bounding box and the detection box. A higher mAP indicates better overall performance in detecting and localizing objects across all classes.
Figure 11 exhibits the F1 score at a 94% threshold with a confidence level of 0.566. The F1 score considers both precision and recall, making it a valuable metric for assessing model performance.
Figure 12 exhibits the training and validation losses of the detection model over 932 epochs on the S-BIRD dataset. This graph helps in understanding the model’s learning progress during training and validation phases. A decrease in loss indicates that the model is learning to make better predictions.
Figure 13 exhibits the detection outcomes obtained by deploying the trained model on Google Source frames [22,23,24,25,26,27] as input data. The outcomes include the location of objects and corresponding class labels (tree roots, grease, or plastic) predicted by the model. These results are of utmost importance as they enable a thorough evaluation of the model’s performance and adaptability when dealing with new and diverse data in real-world scenarios. Additionally, the model has been specifically optimized to handle multiple sewer blockages within the same frame, making it highly suitable for real-time detection in various practical situations.

5. Comparing AI-Driven Approach to MOEAs

The AI-driven approach presented in this research offers several advantages over Multi-Objective Evolutionary Algorithms (MOEAs) [28] commonly used in wastewater system management. While MOEAs such as NSGA-II, SPEA2, MOPSO, and MODE are effective at optimizing multiple objectives, they often come with the burden of complex mathematical models and high computational requirements [29,30]. In contrast, the AI approach leverages advanced computer vision and deep learning techniques to detect sewer blockages promptly and accurately. The model achieves a remarkable mean average precision (mAP) of 96.30% at a confidence threshold of 0.5, highlighting its exceptional precision in sewer blockage detection, which in turn enhances the reliability and efficiency of wastewater management systems.
Furthermore, the AI approach relies on labelled training data and lightweight deep learning models, enhancing its efficiency and real-time capabilities. This aligns well with the urgent need to address sewer blockages swiftly and prevent disruptions and overflows. The model’s accuracy, speed, and specialized focus on sewer blockage detection make it a highly promising solution for immediate and effective urban wastewater system management. In comparison, MOEAs such as the sensitivity-based adaptive procedure (SAP) [31], optimal control algorithms [32], and novel methodologies [33] have shown efficiency in various aspects of wastewater management, such as sewer rehabilitation and optimal scheduling. However, their computational demands and reliance on complex algorithms might hinder their real-time applicability. The AI-driven approach’s ability to process data in real-time, coupled with its high accuracy in detection, gives it a distinct edge for addressing dynamic and critical scenarios like sewer blockages.
Overall, while both AI-driven approaches and MOEAs contribute to the advancement of wastewater management, the AI approach’s ability to quickly detect and respond to sewer blockages makes it particularly well-suited for immediate, on-the-ground applications in modern urban sanitation systems.

6. Conclusions

This research highlights the potential of artificial intelligence, by employing the YOLOv5 single-stage detection model and transfer learning on the critical S-BIRD image dataset in sewer blockage detection. By harnessing the power of AI, we achieved a high precision rate suitable for real-time deployment on resource-constrained mobile devices.
Based on the current work, the following specific conclusions may be made.
  • The developed model demonstrated noticeable precision and recall rates, achieving 94.50%, 95.90%, and 98.40% average precision for tree roots, grease, and plastics, respectively. The mean average precision (mAP) reached an outstanding 96.30% at a confidence threshold of 0.5 and maintained consistent performance at mAP of 79.20% across IoU thresholds ranging from 0.5 to 0.95, indicating the model’s proficiency in handling different sewer blockage scenarios. The inference times were efficient, making the model suitable for real-time applications. The detection outcomes on Google Source frames further validated the model’s adaptability to diverse data.
  • The results emphasize the effectiveness of transfer learning and fine tuning, reducing training time, enhancing performance, and in adapting deep neural network models to new contexts.
  • The presented model’s ability to accurately detect sewer blockages holds promise for its application in modern wastewater management systems. The AI-driven sewer blockage detection system showcased in this research has significant implications for real-world applications, ranging from urban infrastructure management to environmental conservation.
As AI technologies continue to advance, the integration of computer vision and deep learning models will pave the way for more efficient and intelligent solutions in various new domains.

Author Contributions

Conceptualization, R.R.P., M.Y.M. and R.K.C.; methodology, R.R.P.; software, R.R.P.; dataset creation, R.R.P.; validation, R.R.P., M.Y.M. and R.K.C.; formal analysis, R.R.P., M.Y.M. and R.K.C.; investigation, R.R.P.; writing—original draft preparation, R.R.P.; writing—review and editing, R.K.C. and R.R.P.; visualization, R.K.C. and R.R.P.; project administration, R.K.C., M.Y.M. and S.M.A.; and funding acquisition, R.K.C. All authors have read and agreed to the published version of the manuscript.

Funding

The research visit of R.R.P. is funded by project PEERS (UTF 2020/10131). The publication charges for this article have been funded by the publication fund of UiT The Arctic University of Norway.

Data Availability Statement

The research data will be made available on the request.

Acknowledgments

Authors acknowledge the support from SPRING EU-India Project (No. 821423 and GOI No. BT/IN/EU-WR/60/SP/2018) and UiT The Arctic University of Norway, Narvik, Norway, for the Ph.D. studies of Ravindra R. Patil. We extend our thanks to ADY Patil School of Engineering, Pune, India.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhao, Z.Q.; Zheng, P.; Xu, S.T.; Wu, X. Object detection with deep learning: A review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [PubMed]
  2. Kaur, R.; Singh, S. A comprehensive review of object detection with deep learning. Digit. Signal Process. 2022, 132, 103812. [Google Scholar] [CrossRef]
  3. Zou, Z.; Chen, K.; Shi, Z.; Guo, Y.; Ye, J. Object detection in 20 years: A survey. Proc. IEEE 2023, 111, 257–276. [Google Scholar] [CrossRef]
  4. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  5. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  6. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  7. Ultralytics/yolov5. Available online: https://github.com/ultralytics/yolov5 (accessed on 9 June 2023).
  8. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single shot multibox detector. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  9. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6569–6578. [Google Scholar]
  10. Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
  11. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  12. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  13. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  14. Cai, Z.; Vasconcelos, N. Cascade R-CNN: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  15. Dai, J.; Li, Y.; He, K.; Sun, J. R-FCN: Object detection via region-based fully convolutional networks. In Proceedings of the Advances in Neural Information Processing Systems 29 (NIPS 2016), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  16. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? In Proceedings of the Advances in Neural Information Processing Systems 27 (NIPS 2014), Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  17. Long, M.; Cao, Y.; Wang, J.; Jordan, M. Learning transferable features with deep adaptation networks. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 97–105. [Google Scholar]
  18. Vaidya, O.S.; Patil, R.; Phade, G.M.; Gandhe, S.T. Embedded Vision Based Cost Effective Tele-operating Smart Robot. Int. J. Innov. Technol. Explor. Eng. 2019, 8, 1544–1550. [Google Scholar]
  19. Patil, R.R.; Vaidya, O.S.; Phade, G.M.; Gandhe, S.T. Qualified Scrutiny for Real-Time Object Tracking Framework. Int. J. Emerg. Technol. 2020, 11, 313–319. [Google Scholar]
  20. Patil, R.R.; Ansari, S.M.; Calay, R.K.; Mustafa, M.Y. Review of the State-of-the-art Sewer Monitoring and Maintenance Systems Pune Municipal Corporation-A Case Study. TEM J. 2021, 10, 1500–1508. [Google Scholar] [CrossRef]
  21. Patil, R.R.; Mustafa, M.Y.; Calay, R.K.; Ansari, S.M. S-BIRD: A Novel Critical Multi-Class Imagery Dataset for Sewer Monitoring and Maintenance Systems. Sensors 2023, 23, 2966. [Google Scholar] [CrossRef] [PubMed]
  22. Google Source Images. Available online: https://www.drainmasterohio.com/red-flags-of-tree-root-intrusion-in-your-drain-pipes/ (accessed on 30 June 2023).
  23. Google Source Images. Available online: https://arboriculture.files.wordpress.com/2016/02/treerootpipe.jpg (accessed on 30 June 2023).
  24. Google Source Images. Available online: https://spunout.ie/wp-content/uploads/elementor/thumbs/Plastic_bottles_in_the_sea-q0ubkb8pkwa5boeuhpaj6o0v1e8l43mla862l6488o.jpg (accessed on 30 June 2023).
  25. Google Source Images. Available online: https://bbwsd.com/wordpress/wp-content/uploads/2018/03/FOG-850x425.jpg (accessed on 30 June 2023).
  26. Google Source Images. Available online: https://images.squarespace-cdn.com/content/v1/55e97d2de4b0a47f46957437/1499308890029-VM48EFRJJMCSOFFHFETV/iStock-482437666.jpg?format=1000w (accessed on 30 June 2023).
  27. Google Source Images. Available online: https://www.istockphoto.com/photo/plastic-bottles-isolated-on-white-gm1202347223-345153972 (accessed on 30 June 2023).
  28. Wang, Z.; Pei, Y.; Li, J. A Survey on Search Strategy of Evolutionary Multi-Objective Optimization Algorithms. Appl. Sci. 2023, 13, 4643. [Google Scholar] [CrossRef]
  29. Jiang, L.; Geng, Z.; Gu, D.; Guo, S.; Huang, R.; Cheng, H.; Zhu, K. RS-SVM machine learning approach driven by case data for selecting urban drainage network restoration scheme. Data Intell. 2023, 5, 413–437. [Google Scholar] [CrossRef]
  30. Yazdi, J. Rehabilitation of urban drainage systems using a resilience-based approach. Water Resour. Manag. 2018, 32, 721–734. [Google Scholar] [CrossRef]
  31. Cai, X.; Shirkhani, H.; Mohammadian, A. Sensitivity-based adaptive procedure (SAP) for optimal rehabilitation of sewer systems. Urban Water J. 2022, 19, 889–899. [Google Scholar] [CrossRef]
  32. Rathnayake, U. Migrating storms and optimal control of urban sewer networks. Hydrology 2015, 2, 230–241. [Google Scholar] [CrossRef]
  33. Draude, S.; Keedwell, E.; Kapelan, Z.; Hiscock, R. Multi-objective optimisation of sewer maintenance scheduling. J. Hydroinform. 2022, 24, 574–589. [Google Scholar] [CrossRef]
Figure 1. Architectural perception of YOLOv5 model.
Figure 1. Architectural perception of YOLOv5 model.
Electronics 12 03606 g001
Figure 2. Labelling details of training instances from dataset.
Figure 2. Labelling details of training instances from dataset.
Electronics 12 03606 g002
Figure 3. Object classes in each training frame by center x.
Figure 3. Object classes in each training frame by center x.
Electronics 12 03606 g003
Figure 4. Obtained higher precision rate for each class.
Figure 4. Obtained higher precision rate for each class.
Electronics 12 03606 g004
Figure 5. Confusion matrix details for all classes.
Figure 5. Confusion matrix details for all classes.
Electronics 12 03606 g005
Figure 6. Correlogram for frames detailing.
Figure 6. Correlogram for frames detailing.
Electronics 12 03606 g006
Figure 7. Scatter chart for instances and linked labels.
Figure 7. Scatter chart for instances and linked labels.
Electronics 12 03606 g007
Figure 8. Precision (P) versus confidence (C) chart.
Figure 8. Precision (P) versus confidence (C) chart.
Electronics 12 03606 g008
Figure 9. Recall (R) versus confidence (C) chart.
Figure 9. Recall (R) versus confidence (C) chart.
Electronics 12 03606 g009
Figure 10. Precision (P) versus recall (R) chart.
Figure 10. Precision (P) versus recall (R) chart.
Electronics 12 03606 g010
Figure 11. F1 score versus confidence (C) chart.
Figure 11. F1 score versus confidence (C) chart.
Electronics 12 03606 g011
Figure 12. Detailing of losses in training and validation.
Figure 12. Detailing of losses in training and validation.
Electronics 12 03606 g012
Figure 13. Identification and localization outcomes.
Figure 13. Identification and localization outcomes.
Electronics 12 03606 g013
Table 1. Principal training configurations.
Table 1. Principal training configurations.
AttributesImplications
learning modelYOLOv5-s
Annotation data typePyTorch TXT
max_epoch6000
patience100
batch_size16
fp16True
num_classes3
Params7.2 M
Gflops15.9
depth0.33
width0.5
input_size(416, 416)
workers8
anchor_t4.0
scale0.5
hsv_h, hsv_s, hsv_v0.015, 0.7, 0.4
warmup_epochs3
weight_decay0.0005
momentum0.937
translate0.1
Table 2. Temporal evaluation details.
Table 2. Temporal evaluation details.
Timing AttributesOutturns (Milliseconds)
Average forward time0.2 ms
Average NMS time1.1 ms
Average inference time11 ms
Table 3. Precision assessment details.
Table 3. Precision assessment details.
Object ClassAverage Precisionmap_5095map_50
tree roots0.9450.7920.9630
grease0.959
plastic0.984
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Patil, R.R.; Calay, R.K.; Mustafa, M.Y.; Ansari, S.M. AI-Driven High-Precision Model for Blockage Detection in Urban Wastewater Systems. Electronics 2023, 12, 3606. https://doi.org/10.3390/electronics12173606

AMA Style

Patil RR, Calay RK, Mustafa MY, Ansari SM. AI-Driven High-Precision Model for Blockage Detection in Urban Wastewater Systems. Electronics. 2023; 12(17):3606. https://doi.org/10.3390/electronics12173606

Chicago/Turabian Style

Patil, Ravindra R., Rajnish Kaur Calay, Mohamad Y. Mustafa, and Saniya M. Ansari. 2023. "AI-Driven High-Precision Model for Blockage Detection in Urban Wastewater Systems" Electronics 12, no. 17: 3606. https://doi.org/10.3390/electronics12173606

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop