Next Article in Journal
Evaluation of Microlenses, Color Filters, and Polarizing Filters in CIS for Space Applications
Next Article in Special Issue
Progressive Model-Driven Approach for 3D Modeling of Indoor Spaces
Previous Article in Journal
Hang-Time HAR: A Benchmark Dataset for Basketball Activity Recognition Using Wrist-Worn Inertial Sensors
Previous Article in Special Issue
Occlusion Handling for Mobile AR Applications in Indoor and Outdoor Scenarios
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quantification of Structural Defects Using Pixel Level Spatial Information from Photogrammetry †

1
School of Civil and Environmental Engineering, University of New South Wales, Sydney, NSW 2052, Australia
2
Linke & Linke Surveys, 34-36 Byrnes St, Botany, Sydney, NSW 2019, Australia
*
Author to whom correspondence should be addressed.
This paper is an extended version of published paper in Guo, Y.; Wang, Z.; Shen, X.; Barati, K.; Linke, J. Automatic Detection and Dimensional Measurement of Minor Concrete Cracks with Convolutional Neural Network. In Proceedings of the 7th International Conference on Smart Data and Smart Cities (SDSC), Sydney, Australia, 19–21 October 2022.
Sensors 2023, 23(13), 5878; https://doi.org/10.3390/s23135878
Submission received: 31 March 2023 / Revised: 12 June 2023 / Accepted: 19 June 2023 / Published: 25 June 2023
(This article belongs to the Special Issue Smart Data Smart Cities & 3D GeoInfo)

Abstract

:
Aging infrastructure has drawn increased attention globally, as its collapse would be destructive economically and socially. Precise quantification of minor defects is essential for identifying issues before structural failure occurs. Most studies measured the dimension of defects at image level, ignoring the third-dimensional information available from close-range photogrammetry. This paper aims to develop an efficient approach to accurately detecting and quantifying minor defects on complicated infrastructures. Pixel sizes of inspection images are estimated using spatial information generated from three-dimensional (3D) point cloud reconstruction. The key contribution of this research is to obtain the actual pixel size within the grided small sections by relating spatial information. To automate the process, deep learning technology is applied to detect and highlight the cracked area at the pixel level. The adopted convolutional neural network (CNN) achieves an F1 score of 0.613 for minor crack extraction. After that, the actual crack dimension can be derived by multiplying the pixel number with the pixel size. Compared with the traditional approach, defects distributed on a complex structure can be estimated with the proposed approach. A pilot case study was conducted on a concrete footpath with cracks distributed on a selected 1500 mm × 1500 mm concrete road section. Overall, 10 out of 88 images are selected for validation; average errors ranging from 0.26 mm to 0.71 mm were achieved for minor cracks under 5 mm, which demonstrates a promising result of the proposed study.

1. Introduction

The problem with aging infrastructure has become one of the most concerning issues globally, especially in developed countries. Infrastructure varies from roads, railways, and buildings to bridges, power plants, and dams. Any potential failure would cause severe impacts to both economic and social life. As most of the crucial infrastructure is made of reinforced concrete, the inspection of these structures is considered a priority. With increasing age, defects such as cracking, spalling, and corrosion could inevitably appear. Wasim, et al. [1] reviewed the durability of geopolymer concrete in the last 20 years. It is reported that 46% of collapsed bridges were already categorized as structurally deficient before the collapse took place, and the collapse rate is estimated as 1/1200 in the New York State of the United States (US) [2]. According to Reagan, et al. [3], the typical designed lifespan for a bridge is 50 years. Structural health monitoring is of great importance to improve infrastructure safety, reduce downtime cost, and prevent catastrophic failure. Currently, 42% of all US bridges are over 50 years old, and 46,154 of the nation’s bridges are considered to be in poor condition [4]. To prevent potential structure collapse, any defects that might lead to major structural failure should be identified at an early stage. Once a certain amount of defects are identified, a decision will be made to repair the defects or even abandon the asset by considering the cost. A comprehensive inspection is indispensable to support the decision-making process.
High cost could be one of the issues hindering the execution of infrastructure inspection. The investment for the maintenance stage is much lower compared to the designing and construction stages. Khan, et al. [5] mentioned that scaffolding, lifting, and protective equipment are needed to conduct remote inspection, which would inevitably increase costs. The cost of crew, traffic control, and involved devices, such as snooper truck and man lift [6,7], could be significant. According to a report from ASCE [4], the investment gap for infrastructure has increased from USD 2.1 to 2.59 trillion every 10 years. It is estimated that the cost for bridge repair is USD 125 billion and the annual budget to improve the bridge condition has increased from USD 14.4 to 22.7 billion. It is noted that the Australian governments and industry have been trying to improve investment in infrastructure gaps since 2015 [8].
Conventional structural inspection is usually completed manually, which requires highly experienced skills and could be risky when accessing dangerous areas (such as working from heights or inside tunnels). Another shortcoming of manual inspection could be inconsistent record keeping. It is hard to record the defect’s exact location on a curved surface such as a power plant or chimney. Moreover, it is time-consuming for one inspector to conduct the examination. If multiple inspectors are involved, errors might appear due to the difference in recognition. Qureshi, et al. [9] also pointed out surface condition rating systems and characteristics to evaluate conditions are not unified.
Recently, some researchers have conducted studies on performing crack assessment using unmanned aerial vehicles (UAVs) [10] and applying deep learning techniques for crack segmentation [11]. Wasim and Djukic [12] reviewed the external corrosion of buried pipelines and up-to-date management methods. Semantic segmentation is one of the emerging technologies that has been adopted in various industrial applications. With the introduction of semantic segmentation for asset inspection, traditional time-consuming and tedious inspection work can be performed automatically. Moreover, as the data is stored in a digital format, a time-based inspection approach can be used to track defect changes over time.
The objective of this paper is to develop an integrated methodology for efficiently measuring minor cracks on concrete structures. Pixel size of an inspection image can be estimated by combining distance information obtained through 3D reconstruction. The first step is separating cracks from the original image with a convolutional neural network and counting the number of pixels for the crack width. Tang, et al. [13] proposed a complete solution including U-Net-based crack segmentation, light and stable backbone extraction, and distribution determination. The pixel sizes are then differentiated by griding the image into small areas and utilizing spatial photogrammetry. Finally, the actual crack dimension is determined by multiplying the pixel number with pixel size. This method has the advantage of being able to estimate defects on complex structures. Although the goal of this method is to monitor the surface condition of large concrete infrastructures accurately and efficiently, access to similar sites is limited. Therefore, to test the feasibility of this method at the initial stage, a pilot case study was conducted on a concrete footpath with cracks ranging from 0.7 to 10 mm. The performance of the proposed method was evaluated using absolute error. Most of the cracks can be identified, and the introduced error was no greater than 0.5 mm.
The structure for the remaining parts of this paper is as follows. In Section 2, previous work on CNN-based semantic segmentation, image-based crack quantification, and 3D reconstruction is reviewed. In Section 3, an integrated methodology is presented to conduct crack measurement utilizing semantic segmentation and computer vision, including data capturing, crack detection, and crack measurement. In Section 4, a case study on the cracks on a concrete footpath is performed to prove the applicability of obtaining grid pixel size. In Section 5, the gaps and limitations of the study are discussed. Finally, a conclusion is drawn in Section 6.

2. Related Work

As the traditional inspection is conducted by humans [14], the number of cracks may be underestimated due to limited access and unavoidable human error. To overcome these drawbacks and minimize the budget for infrastructure inspections, a series of advancements have been made to optimize the non-contact workflow from data capture and defect detection to defect quantification. Initially, cameras and light detection and ranging (LiDAR) sensors were used to capture structural defects. However, due to the balance between capability, compatibility, and cost, cameras are considered the best option for inspecting minor defects such as cracks. Therefore, two-dimensional red, green and blue (RGB) images will be the primary data source in this study. Moreover, the detection process was automated using machine learning techniques.

2.1. Defect Detection Based on Machine Learning

Since the introduction of AlexNet [15], CNN has been widely applied in many industries. In recent years, researchers have been introducing deep learning-based techniques for crack detection.
Manual inspection has been widely conducted in recent years due to the crucial role of experience in this field. However, the introduction of artificial intelligence (AI) has the potential to reduce the burden on inspectors by limiting the area of interest. In particular, AI has rapidly developed in the realm of two-dimensional imaging. Semantic segmentation algorithms, which are typically based on convolutional neural networks, have become a mature technology in computer science. For instance, Krizhevsky, Sutskever and Hinton [15] trained a deep CNN network with 60 million parameters, achieving a significant score in the ImageNet contest. Zeiler and Fergus [16] further explained the workings of the AlexNet model and developed a superior architecture. Szegedy, et al. [17] proposed an Inception network that optimized the utilization of computing resources. Ronneberger, et al. [18] also presented an efficient strategy for maximizing the use of annotated samples. Furthermore, Szegedy, Vanhoucke, Ioffe, Shlens and Wojna [19] established principles for designing high-performance networks with low computational costs.
Based on CNN, many researchers have developed different structures for different purposes. Defect detection is one of the semantic segmentation applications in the Architecture, Engineering, and Construction (AEC) industry. Semantic segmentation-driven crack detection is more objective and reliable compared to the traditional manual inspection [20]. Oliveira and Correia [21] proposed an automatic system for crack detection and characterization, and the algorithm could detect multiple cracks from 56 images in about two minutes. Chen, et al. [22] suggested a simple and improved structure of convolutional neural networks achieving high accuracy. The authors believed that a large convolution and pooling methodology with fewer network layers could be utilized to obtain a better result for simple crack identification. By setting the learning rate to 0.01, Li and Zhao [23] developed an algorithm with high accuracy based on CNN structure and AlexNet. Liu, et al. [24] adopted U-Net for high efficiency and robustness. Dung [25] proposed a crack detection method based on FCN for semantic segmentation on concrete crack images. Bang, et al. [26] proposed a deep convolutional encoder-decoder network-based method to identify road cracks from black-box images. The automated crack identification and visualization algorithm used by Jang, et al. [27] is enabled by transfer learning from GoogleNet. Qu, et al. [28] applied LeNet-5 to classify the cracks and optimized VGG16 to extract concrete crack characteristics. Chow, et al. [29] provided an artificial intelligence-based inspection workflow for anomaly detection and reduced the search space of defects up to 80% for minor defect regions. Dais, et al. [30] firstly applied deep learning techniques on masonry images with pixel-level segmentation. Miao and Srimahachota [31] combined a trained CNN and an image processing method to detect and quantify cracks in a semi-automatic way. Fu, Meng, Li and Wang [6] proposed an algorithm based on Dense-DeepLabv3+ network to segment bridge crack images. Ali, et al. [32] reviewed the applications of CNN on civil crack detection. Wang and Su [33] suggested the SegCrack model, including a hierarchically structured transformer encoder to output features and a top-down pathway with lateral connections to up-sample and fuse features. Moreover, Xu, et al. [34,35] applied deep neural networks for 3D object detection over as-built reconstruction and automated scan-to-BIM. Although much research has been conducted on crack detection, the labeled area is still not accurate enough for minor cracks.

2.2. Defect Measurement with Image Processing and Photogrammetry

To measure the actual dimensions of cracks, researchers have performed experiments to extract the information from images. However, lens distortion and projective transformation can result in inaccuracies in the measurements. Therefore, reconstructing cracks in three-dimensional space is one of the most reliable ways for quantifying cracks.
Cho, et al. [36] presented a five-step method to improve the accuracy and consistency of measuring crack width. Albareda-Valls, et al. [37] tested an image post-processing method to quantify cracks on concrete elements. Vashpanov, et al. [38] developed a method to determine crack dimensions based the pixel intensity distribution of images and achieved an accuracy of less than ±15%. Liu, Nie, Fan and Liu [10] concluded that the assessment of cracks can be concluded as filtering noise and extracting parameters. Bang, et al. [39] used structured lights and depth cameras to quantify structural damage. Wang, et al. [40] proposed a key point method for crack characterization and established a crack model based on anchor points. Shi, et al. [41] reconstructed 3D images based on structured illumination. Fan, et al. [42] proposed a method to measure crack dimensions by extracting crack skeletons from images. Parente, et al. [43] proposed a machine learning-based method that only requires a single image for training and provides accurate outputs.
The issue with the aforementioned approaches is that the quantification process is only based on a two-dimensional image, which neglects the information of the third dimension. The influence of lens distortion and projective transformation should also be considered. One of the most recent solutions is adopting computer vision and close-range photogrammetry to reconstruct the defects in three dimensions.
Jahanshahi and Masri [44] proposed a contactless quantification method for cracks based on computer vision and image processing. Liu, et al. [45] proposed a solution to locate cracks by combining 2D image and 3D scene reconstruction. Yang, et al. [46] proposed a damage-indexing method that integrates image-based crack measurement and crack quantification methods. Kalfarisi, et al. [47] used a 3D reality mesh for quantitative assessment. Wu, et al. [48] combined UAV-taken photos and Mask-RCNN to construct a 3D water tower model with highlighted cracks. Building upon previous results, Liu, Nie, Fan and Liu [10] presented a new crack assessment approach using UAVs and 3D scene reconstruction to inspect bridge piers. The authors also presented a method of projecting cracks onto a 3D mesh surface, which eliminates distortion on non-flat surfaces. Chaiyasarn, et al. [49] detected a large range of cracks on a 3D mesh model by creating an artificial camera position. Shokri, et al. [50] proposed a planar and matching method for 3D crack reconstruction with higher accuracy and faster speed. Zhao, et al. [51] presented a system of camera and laser rangefinder to measure the width of cracks from different angles and distances. Woo, et al. [52] used relative objects in the image to rectify the location of cracks without GPS information, which can potentially improve the accuracy of the measurement results.
Although much research has been conducted to measure the dimensions of cracks from 2D images, some limitations remain. The distance between the target and camera is typically fixed or manually measured, which is time-consuming, especially when multiple images are needed for photogrammetry. Additionally, traditional crack quantification can only be performed on simple flat surfaces.

3. Research Methodology

In this paper, a method based on 3D reconstruction and semantic segmentation will be adopted to acquire the pixel size information as shown in Figure 1. To measure the actual dimension of cracks from an image, distance information is needed. The most direct way is to measure the distance while taking photos. However, this process could be time-consuming and inaccurate. The obtained image in this study will be processed in three directions: automatic pixel dimension extraction, three-dimensional reconstruction, and grid point location.

3.1. Pixel Level Semantic Segmentation for Defect

To derive the crack pixel dimension from obtained images, the first step is to separate cracks from the background automatically. With the development of deep learning technology, CNN-based semantic segmentation technology is utilized to automatically detect the cracked area. This section will present a practical workflow to implement the two-dimensional artificial intelligence for crack detection.
Referring to Simonyan and Zisserman [53], VGG16 is adopted as the encoder. It initially has 16 weight layers, and each layer consists of Maxpool and Convolution + Batch Norm + ReLU. After testing, it was found that adding Batch Norm between Convolution and ReLU could improve performance. The proposed convolutional neural network model is adjusted for concrete cracks [54].
The input image was cut into 448 × 448 and fed into the algorithm as a [448 × 448 × 3] matrix. The decoder is designed for crack detection, each layer consists of Bilinear Interpolate and Convolution Kernel Size 3 + ReLU. The dimensions of each layer are listed below (see Table 1).
The final layer consists of Convolution (kernel size 3) + ReLU, Convolution Kernel Size 1, and Log SoftMax. The dimension of the output image is [448 × 448 × 1], and the value of each pixel is either 0 or 1, representing crack or non-crack. However, since the number of crack and non-crack images is not equal, the traditional Binary Cross Entry (BCE) loss tends to regard the image as not having cracks. To overcome this issue, a focal loss method based on the structure proposed by Lin, et al. [55] is applied for classification. Backpropagation is then performed to adjust the parameters. The VGG16 + Focal Loss model was trained on a smaller dataset with fewer epochs. Although ResNet is commonly used for crack detection, U-Net [27], which is one of the recent developments based on ResNet, will also be compared with VGG16 + Focal Loss and VGG16 + BCE Loss (see Table 2). The overall performance of VGG16 + Focal Loss, with an F1 Score of 0.613, is better than the other two models.
The results of annotated images for some obvious cracks and thin cracks are presented in Figure 2.
Next, the pixel dimensions of the cracks can be determined. The number of pixels within the highlighted area will be counted to represent the crack dimension. Crack measuring is based on the labelled crack image created in the previous section. Since the pixels in the crack area are set to 1 and the rest of the pixels are labelled as 0 for non-crack, the “skimage” package was applied to draw the boundary of the cracked area. Then, a skeleton of the crack will be created at the center of two longitudinal lines. After that, a line of width will be created perpendicularly to the skeleton, as proposed by Cho, Yoon and Jung [39].
To obtain the real dimension of the cracks, multiplying the pixel dimension with the scale factor will be needed. In this scenario, pixel size as the third-dimension information is the key issue. In the follow-up sections, scale factors will be obtained from photogrammetry-enabled 3D reconstruction.

3.2. Pixel Size Quantification Using Spatial Information from Photogrammetry

For most of the crack measurement process, pixel sizes are normally regarded as the same. Therefore, the main contribution for this research is to differentiate pixel sizes in different areas. Ideally, this method will make it possible to measure the defects on images taken from different angles. The workflow will be presented in the following sections.
To overcome the limitation of traditional inspection methods, photogrammetry algorithms are selected to obtain the spatial information by reconstructing the defects in 3D. As the exported point cloud model is not scaled, control points or references will be needed to convert the model to actual size. Moreover, the connecting information between 2D and 3D tie points will also be of great importance for the next procedures.
Grids are applied to divide the image into multiple small sections with different pixel sizes. The number of grids depends on the desired accuracy for crack measurement. In this paper, each image is divided into 8 × 8 areas to differentiate pixel sizes. Grid points are used to assist calculations within different areas.
The flowchart of calculating grid pixel size is presented in Figure 3. To obtain the actual size of corner grid pixels, the first step is to locate the 2D pixel in scaled 3D point clouds. However, the corner pixel at the grid point in the image does not usually have a corresponding tie point because the tie point cloud is relatively sparse. Therefore, instead of using the exact grid point pixel, the nearest three 2D tie points around the corner point will be adopted.
This paper adopts the perimeter method instead of the area method to calculate the grid pixel size, as the calculation can overcome the scenario when three points are located on the same line and computation speed is faster. The grid pixel size can be calculated using Equation (1). The distances between three points can be labelled as Pa, Pb, and Pc (pixels). Then, using the exported corresponding information, relate 2D coordinates to 3D coordinates in point clouds. The distances between three points can be labelled as Ra, Rb, and Rc (mm).
G r i d   P i x e l   S i z e = R a + R b + R c P a + P b + P c m m / P i x e l
Knowing the pixel size at the grid point, the next step is to derive the area pixel size. By averaging the values of four grid points at corners, area pixel sizes within that area can be obtained. For better visualization, pixel sizes are represented by different color brightness (the larger pixel size has a brighter color). Since pixel dimension and pixel size are known, the actual dimension of the cracks in different sections can then be calculated by multiplying the two values. To validate the accuracy, the exported results will be compared with the gauge measured value and dimensions measured from the point cloud.

4. Experiments and Validations

To validate the feasibility of the proposed methodology, a pilot case study was performed over a small-scale site which is a concrete footpath (approximately 1500 mm × 1500 mm) with a long crack distributed on the flat surface. Instead of a flying UAV, an Apple iPhone 7 was used to capture images for testing purposes. The average crack width is about 3.15 mm. Approximately 88 images with a resolution of 4032 × 3024 pixels were taken of the crack. The raw images, processed images, and part of the measured spots on the crack are labelled in Figure 4. Ten images took from different angles were selected to validate the proposed methodology (see Figure 5).

4.1. Crack Width in Pixel

As mentioned in the methodology, the proposed semantic segmentation algorithm was applied to the images to highlight the cracked area with yellow. The labelled crack on the original image and mask are presented in Figure 4b and Figure 4c, respectively. The pixel dimension is automatically labelled, as can be seen from Figure 4d,e.

4.2. Spatial Information

The next step is to obtain the spatial information. photogrammetry is applied to perform 3D reconstruction based on the captured images. A sparse point cloud (see Figure 6a) is created in this case study as the accuracy is good enough.
As the model exported from COLMAP is in an arbitrary unit system, a referenced length or ground control point (GCP) is needed to scale the model to actual dimension. In this case study, the arbitrary length of the crack is measured as 2.26 from the point cloud as can be seen in Figure 6a and the real length is measured as 1430 mm (see Figure 6b). By dividing the real length by the arbitrary value, the 3D point cloud scaled factor can be derived as 632, which will be applied to calculate the width of each crack.
The information generated and exported from photogrammetry is crucial as it contains the corresponding information between 2D tie points on the image and 3D point clouds. By selecting a pixel on the image, the spatial information can be derived through the exported data file.

4.3. Grided Image

The third stage is to calculate various pixel sizes along the crack. To achieve this goal, the image is divided into small sections. Theoretically, the smaller the division, the more accurate the result will be. In this case study, the image is segmented into an 8 × 8 grid as a preliminary test. As the original size of the image is 4032 × 3024, it is divided into 64 pieces of 504 × 378 elements.
The intersections are labelled with blue dots in Figure 7a. The blue dots are corner points of the rectangular area that represent the pixel sizes within the area. The process is described as follows.
The first step is obtaining the size of the corner point. As displayed in the figure, the corner point has a specific coordinate on the image. The nearest three tie points are found and labelled as orange triangles (see Figure 7b). The perimeter of the formed triangular area can be calculated in pixel unit. Then, the 3D coordinates of the three tie points can be found by linking information generated from photogrammetry. From there, the 3D perimeter in arbitrary units can be calculated, and the pixel size within that area can be calculated by dividing the 3D perimeter with the 2D pixel perimeter. Therefore, the blue dot size can be derived as the corner point size is subjected to the pixel sizes within the triangular area.
Since the pixel size at the corner points is known, the pixel size within the rectangular area can be calculated by averaging the sizes of the four corner points. After that, a chart showing different pixel sizes in different areas can be mapped as seen in Figure 8a. Since further pixels correspond to larger real dimensions in the image, they are represented by lighter colors on the map. However, the transition is not 100% correct. One of the main issues is that the noise generated along the point cloud can lead to inaccurate results as seen in Figure 8b.

4.4. Derivation of the Actual Size

Based on the automatically counted pixel width and pixel size map as displayed in Figure 9. The actual dimension can be calculated in Equation (2).
C a l c u l a t e d   W i d t h = 3 D   S c a l e   F a c t o r × P i x e l   W i d t h × P i x e l   S i z e

4.5. Validation

Statistical results will be created by comparing the calculated results with the gauge measured results.
According to the Australian Standard 2870 [56], damage levels for different crack widths are presented as below (see Table 3). Most of the cracks in this paper can be regarded as wide cracks.
IMG_5403 is selected to present the results of proposed crack measurement. The distribution of errors in this image is typical among the ten images. As can be seen from Table 4, most errors are around 1 mm. However, several errors are larger than 1.5 mm (e.g., 1.84 mm, 2.51 mm, and 3.01 mm). Several factors could cause the large errors, such as an inaccurate pixel scale factor. One of the most common issues is that the detection algorithm might export the wrong pixel number, as can be seen from Figure 10.
As the outliers could affect the accuracy, when calculating the average error, two sets of data are presented in Table 5. The first row of errors is calculated including errors less than 1.5 mm and the second row are calculated with all data. It can be found that the proposed workflow of crack measurement could lead to an error of 0.48 mm for 3.32 mm mean widths.

5. Discussions and Limitations

In this research, an innovative approach is proposed to determine the dimension of minor defects. With the application of CNN and photogrammetry, the shape of a crack can be automatically extracted, and the pixel size information can be determined. Compared to the conventional methods, the new method makes it possible to accurately quantify the defect from image.
Although the accuracy of the proposed methodology has been validated in the previous section, some issues were found that might affect the result, including point cloud noise, cast of shadows, irregular shape, and shooting angle.
Further research will be focusing on estimating individual pixel sizes while saving computational resources and increasing processing speed. Moreover, by deploying UAV, large-scale experiments will be performed on more complicated infrastructures, such as bridges, dams, and cooling towers to prove the feasibility of the proposed pixel-level method.

6. Conclusions

Many research efforts have been devoted to quantifying the dimension of cracks from images for aging infrastructures, such as bridges, roads, dams, and tunnels. Depth information derived from the close-range photogrammetry is omitted in most studies. This paper discussed the relationship between the real dimension and corresponding pixels on a 2D image. It provides an efficient solution to automatically detect and accurately measure the dimensions of minor cracks. The pixel size is obtained by leveraging spatial data exported from point cloud reconstruction. A case study was performed on a concrete footpath with cracks distributed on the surface in Sydney, and the results proved the feasibility of the proposed methodology.
Some improvements can be made in future research, such as enhancing the accuracy of the semantic segmentation algorithm, denoising the surface, and reducing the size of the grided section. The integration of LiDAR with images could provide an alternative approach to simplify and speed up the process of pixel size determination. Ideally, the proposed technology can be applied to provide accurate defect quantification and realize real-time asset inspection, ultimately improving the safety of public infrastructure.

Author Contributions

Conceptualization, X.S. and Y.G.; methodology, Y.G. and X.S.; software, Z.W.; validation, Y.G.; formal analysis, Y.G.; investigation, Y.G.; resources, J.L.; data curation, Y.G. and Z.W.; writing—original draft preparation, Y.G.; writing—review and editing, X.S. and Y.G.; visualization, Y.G. and Z.W.; supervision, X.S.; project administration, J.L. and X.S.; funding acquisition, J.L., K.B. and X.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Australian Research Council (ARC) Industry Transformation Research Hub for Resilient and Intelligent Infrastructure Systems (RIIS) in Urban, Resources and Energy Sectors (IH210100048).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We sincerely thank Yincai Zhou from the University of New South Wales for providing instructions on photogrammetry. We thank Alastair Linke and Samuel Yu from Linke & Linke Surveys for providing equipment and suggestions for image capturing.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wasim, M.; Ngo, T.D.; Law, D. A state-of-the-art review on the durability of geopolymer concrete for sustainable structures and infrastructure. Constr. Build. Mater. 2021, 291, 123381. [Google Scholar] [CrossRef]
  2. Cook, W.; Barr, P.J. Observations and trends among collapsed bridges in New York state. J. Perform. Constr. Facil. 2017, 31, 04017011. [Google Scholar] [CrossRef]
  3. Reagan, D.; Sabato, A.; Niezrecki, C. Feasibility of using digital image correlation for unmanned aerial vehicle structural health monitoring of bridges. Struct. Health Monit. 2018, 17, 1056–1072. [Google Scholar] [CrossRef]
  4. ASCE. 2021 Report Card for America’s Infrastructure: A comprehensive Assessment of America’s Infrastructure. Available online: https://infrastructurereportcard.org/wp-content/uploads/2020/12/National_IRC_2021-report.pdf (accessed on 4 November 2022).
  5. Khan, F.; Ellenberg, A.; Mazzotti, M.; Kontsos, A.; Moon, F.; Pradhan, A.; Bartoli, I. Investigation on Bridge Assessment Using Unmanned Aerial Systems. In Proceedings of the Structures Congress 2015, Portland, OR, USA, 23–25 April 2015; pp. 404–413. [Google Scholar]
  6. Fu, H.; Meng, D.; Li, W.; Wang, Y. Bridge crack semantic segmentation based on improved Deeplabv3+. J. Mar. Sci. Eng. 2021, 9, 671. [Google Scholar] [CrossRef]
  7. Galdelli, A.; D’Imperio, M.; Marchello, G.; Mancini, A.; Scaccia, M.; Sasso, M.; Frontoni, E.; Cannella, F. A Novel Remote Visual Inspection System for Bridge Predictive Maintenance. Remote Sens. 2022, 14, 2248. [Google Scholar] [CrossRef]
  8. Infrastructure Australia. An Assessment of Australia’s Future Infrastructure Needs. Available online: https://www.infrastructureaustralia.gov.au/sites/default/files/2019-08/Australian%20Infrastructure%20Audit%202019%20-%200.%20Executive%20Summary.pdf (accessed on 5 November 2022).
  9. Qureshi, W.S.; Hassan, S.I.; McKeever, S.; Power, D.; Mulry, B.; Feighan, K.; O’Sullivan, D. An Exploration of Recent Intelligent Image Analysis Techniques for Visual Pavement Surface Condition Assessment. Sensors 2022, 22, 9019. [Google Scholar] [CrossRef]
  10. Liu, Y.F.; Nie, X.; Fan, J.S.; Liu, X.G. Image-based crack assessment of bridge piers using unmanned aerial vehicles and three-dimensional scene reconstruction. Comput.-Aided Civ. Infrastruct. Eng. 2020, 35, 511–529. [Google Scholar] [CrossRef]
  11. Wang, J.J.; Liu, Y.F.; Nie, X.; Mo, Y.L. Deep convolutional neural networks for semantic segmentation of cracks. Struct. Control Health Monit. 2022, 29, e2850. [Google Scholar] [CrossRef]
  12. Wasim, M.; Djukic, M.B. External corrosion of oil and gas pipelines: A review of failure mechanisms and predictive preventions. J. Nat. Gas Sci. Eng. 2022, 100, 104467. [Google Scholar] [CrossRef]
  13. Tang, Y.; Huang, Z.; Chen, Z.; Chen, M.; Zhou, H.; Zhang, H.; Sun, J. Novel visual crack width measurement based on backbone double-scale features for improved detection automation. Eng. Struct. 2023, 274, 115158. [Google Scholar] [CrossRef]
  14. Chaiyasarn, K. Damage Detection and Monitoring for Tunnel Inspection Based on Computer Vision. Ph.D. Thesis, University of Cambridge, Cambridge, UK, 2014. [Google Scholar]
  15. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  16. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. In Proceedings of the 13th European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 818–833. [Google Scholar]
  17. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  18. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the 18th International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  19. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  20. Kim, J.J.; Kim, A.-R.; Lee, S.-W. Artificial neural network-based automated crack detection and analysis for the inspection of concrete structures. Appl. Sci. 2020, 10, 8105. [Google Scholar] [CrossRef]
  21. Oliveira, H.; Correia, P.L. Automatic road crack detection and characterization. IEEE Trans. Intell. Transp. Syst. 2012, 14, 155–168. [Google Scholar] [CrossRef]
  22. Chen, K.; Yadav, A.; Khan, A.; Meng, Y.; Zhu, K. Improved crack detection and recognition based on convolutional neural network. Model. Simul. Eng. 2019, 2019, 8796743. [Google Scholar] [CrossRef] [Green Version]
  23. Li, S.; Zhao, X. Image-based concrete crack detection using convolutional neural network and exhaustive search technique. Adv. Civ. Eng. 2019, 2019, 6520620. [Google Scholar] [CrossRef] [Green Version]
  24. Liu, Z.; Cao, Y.; Wang, Y.; Wang, W. Computer vision-based concrete crack detection using U-net fully convolutional networks. Autom. Constr. 2019, 104, 129–139. [Google Scholar] [CrossRef]
  25. Dung, C.V. Autonomous concrete crack detection using deep fully convolutional neural network. Autom. Constr. 2019, 99, 52–58. [Google Scholar] [CrossRef]
  26. Bang, S.; Park, S.; Kim, H.; Kim, H. Encoder–decoder network for pixel-level road crack detection in black-box images. Comput.-Aided Civ. Infrastruct. Eng. 2019, 34, 713–727. [Google Scholar] [CrossRef]
  27. Jang, K.; Kim, N.; An, Y.-K. Deep learning–based autonomous concrete crack evaluation through hybrid image scanning. Struct. Health Monit. 2019, 18, 1722–1737. [Google Scholar] [CrossRef]
  28. Qu, Z.; Mei, J.; Liu, L.; Zhou, D.-Y. Crack detection of concrete pavement with cross-entropy loss function and improved VGG16 network model. IEEE Access 2020, 8, 54564–54573. [Google Scholar] [CrossRef]
  29. Chow, J.K.; Su, Z.; Wu, J.; Li, Z.; Tan, P.S.; Liu, K.-f.; Mao, X.; Wang, Y.-H. Artificial intelligence-empowered pipeline for image-based inspection of concrete structures. Autom. Constr. 2020, 120, 103372. [Google Scholar] [CrossRef]
  30. Dais, D.; Bal, I.E.; Smyrou, E.; Sarhosis, V. Automatic crack classification and segmentation on masonry surfaces using convolutional neural networks and transfer learning. Autom. Constr. 2021, 125, 103606. [Google Scholar] [CrossRef]
  31. Miao, P.; Srimahachota, T. Cost-effective system for detection and quantification of concrete surface cracks by combination of convolutional neural network and image processing techniques. Constr. Build. Mater. 2021, 293, 123549. [Google Scholar] [CrossRef]
  32. Ali, R.; Chuah, J.H.; Talip, M.S.A.; Mokhtar, N.; Shoaib, M.A. Structural crack detection using deep convolutional neural networks. Autom. Constr. 2022, 133, 103989. [Google Scholar] [CrossRef]
  33. Wang, W.; Su, C. Automatic concrete crack segmentation model based on transformer. Autom. Constr. 2022, 139, 104275. [Google Scholar] [CrossRef]
  34. Xu, Y.; Shen, X.; Lim, S.; Li, X. Three-Dimensional Object Detection with Deep Neural Networks for Automatic As-Built Reconstruction. J. Constr. Eng. Manag. 2021, 147, 04021098. [Google Scholar] [CrossRef]
  35. Xu, Y.; Shen, X.; Lim, S. CorDet: Corner-Aware 3D Object Detection Networks for Automated Scan-to-BIM. J. Comput. Civil Eng. 2021, 35, 04021002. [Google Scholar] [CrossRef]
  36. Cho, H.; Yoon, H.-J.; Jung, J.-Y. Image-based Crack Detection Using Crack Width Transform (CWT) Algorithm. IEEE Access 2018, 6, 60100–60114. [Google Scholar] [CrossRef]
  37. Albareda-Valls, A.; Bustos Herrera, A.; Zamora Mestre, J.L.; Zaribaf, S.S. Image Post-Processing Method for Quantification of Cracking in RC Precast Beams under Bending. Buildings 2018, 8, 158. [Google Scholar] [CrossRef] [Green Version]
  38. Vashpanov, Y.; Son, J.-Y.; Heo, G.; Podousova, T.; Kim, Y.S. Determination of geometric parameters of cracks in concrete by image processing. Adv. Civ. Eng. 2019, 2019, 2398124. [Google Scholar] [CrossRef] [Green Version]
  39. Bang, H.; Min, J.; Jeon, H. Deep Learning-Based Concrete Surface Damage Monitoring Method Using Structured Lights and Depth Camera. Sensors 2021, 21, 2759. [Google Scholar] [CrossRef]
  40. Wang, D.; Cheng, J.; Cai, H. Detection Based on Crack Key Point and Deep Convolutional Neural Network. Appl. Sci. 2021, 11, 11321. [Google Scholar] [CrossRef]
  41. Shi, T.; Qi, Y.; Zhu, C.; Tang, Y.; Wu, B. Three-Dimensional Microscopic Image Reconstruction Based on Structured Light Illumination. Sensors 2021, 21, 6097. [Google Scholar] [CrossRef]
  42. Fan, Z.; Lin, H.; Li, C.; Su, J.; Bruno, S.; Loprencipe, G. Use of Parallel ResNet for High-Performance Pavement Crack Detection and Measurement. Sustainability 2022, 14, 1825. [Google Scholar] [CrossRef]
  43. Parente, L.; Falvo, E.; Castagnetti, C.; Grassi, F.; Mancini, F.; Rossi, P.; Capra, A. Image-Based Monitoring of Cracks: Effectiveness Analysis of an Open-Source Machine Learning-Assisted Procedure. J. Imaging 2022, 8, 22. [Google Scholar] [CrossRef]
  44. Jahanshahi, M.R.; Masri, S.F. A new methodology for non-contact accurate crack width measurement through photogrammetry for automated structural safety evaluation. Smart Mater. Struct. 2013, 22, 035019. [Google Scholar] [CrossRef]
  45. Liu, Y.-F.; Cho, S.; Spencer Jr, B.; Fan, J.-S. Concrete crack assessment using digital image processing and 3D scene reconstruction. J. Comput. Civ. Eng. 2016, 30, 04014124. [Google Scholar] [CrossRef]
  46. Yang, Y.-S.; Chang, C.-H.; Wu, C.-l. Damage Indexing Method for Shear Critical Tubular Reinforced Concrete Structures Based on Crack Image Analysis. Sensors 2019, 19, 4304. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Kalfarisi, R.; Wu, Z.Y.; Soh, K. Crack detection and segmentation using deep learning with 3D reality mesh model for quantitative assessment and integrated visualization. J. Comput. Civ. Eng. 2020, 34, 04020010. [Google Scholar] [CrossRef]
  48. Wu, Z.; Kalfarisi, R.; Kouyoumdjian, F.; Taelman, C. Applying deep convolutional neural network with 3D reality mesh model for water tank crack detection and evaluation. Urban Water J. 2020, 17, 682–695. [Google Scholar] [CrossRef]
  49. Chaiyasarn, K.; Buatik, A.; Mohamad, H.; Zhou, M.; Kongsilp, S.; Poovarodom, N. Integrated pixel-level CNN-FCN crack detection via photogrammetric 3D texture mapping of concrete structures. Autom. Constr. 2022, 140, 104388. [Google Scholar] [CrossRef]
  50. Shokri, P.; Shahbazi, M.; Nielsen, J. Semantic Segmentation and 3D Reconstruction of Concrete Cracks. Remote Sens. 2022, 14, 5793. [Google Scholar] [CrossRef]
  51. Zhao, S.; Kang, F.; Li, J. Non-Contact Crack Visual Measurement System Combining Improved U-Net Algorithm and Canny Edge Detection Method with Laser Rangefinder and Camera. Appl. Sci. 2022, 12, 10651. [Google Scholar] [CrossRef]
  52. Woo, H.-J.; Seo, D.-M.; Kim, M.-S.; Park, M.-S.; Hong, W.-H.; Baek, S.-C. Localization of Cracks in Concrete Structures Using an Unmanned Aerial Vehicle. Sensors 2022, 22, 6711. [Google Scholar] [CrossRef]
  53. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  54. Guo, Y.; Wang, Z.; Shen, X.; Barati, K.; Linke, J. Automatic Detection and Dimensional Measurement of Minor Concrete Cracks with Convolutional Neural Network. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, 10. [Google Scholar] [CrossRef]
  55. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  56. AS 2870; Residential Slabs and Footing. Standards Australia: Sydney, Australia, 2011.
Figure 1. The workflow of crack quantification.
Figure 1. The workflow of crack quantification.
Sensors 23 05878 g001
Figure 2. The results of different models for obvious and minor cracks: (a) original image of obvious cracks, (b) VGG16 + Focal Loss with more training data sets for obvious cracks, (c) original image of minor cracks, and (d) VGG16 + Focal Loss with more training data sets for minor cracks.
Figure 2. The results of different models for obvious and minor cracks: (a) original image of obvious cracks, (b) VGG16 + Focal Loss with more training data sets for obvious cracks, (c) original image of minor cracks, and (d) VGG16 + Focal Loss with more training data sets for minor cracks.
Sensors 23 05878 g002
Figure 3. Flowchart of calculating grid pixel size: (a) grid, (b) corner pixel, (c) nearest three tie points on image, (d) pixel perimeter, (e) corresponded three 3D tie points, (f) actual perimeter.
Figure 3. Flowchart of calculating grid pixel size: (a) grid, (b) corner pixel, (c) nearest three tie points on image, (d) pixel perimeter, (e) corresponded three 3D tie points, (f) actual perimeter.
Sensors 23 05878 g003
Figure 4. The process of measuring crack pixel dimension: (a) raw image of the crack; (b) labelled crack; (c) and crack mask; (d) automatically counted crack width in pixel; (e) partially zoomed in crack; (f) part of labelled measurement spots on the target crack.
Figure 4. The process of measuring crack pixel dimension: (a) raw image of the crack; (b) labelled crack; (c) and crack mask; (d) automatically counted crack width in pixel; (e) partially zoomed in crack; (f) part of labelled measurement spots on the target crack.
Sensors 23 05878 g004
Figure 5. Ten images shot from different angles.
Figure 5. Ten images shot from different angles.
Sensors 23 05878 g005
Figure 6. The process of defining 3D scale factor: (a) the referenced length of the crack; (b) the referenced length of the crack.
Figure 6. The process of defining 3D scale factor: (a) the referenced length of the crack; (b) the referenced length of the crack.
Sensors 23 05878 g006
Figure 7. Calculate the pixel size of corner grid: (a) corner point; (b) nearest three tie points.
Figure 7. Calculate the pixel size of corner grid: (a) corner point; (b) nearest three tie points.
Sensors 23 05878 g007
Figure 8. (a) Heat map on pixel size of an image; (b) noise of the created point cloud surface.
Figure 8. (a) Heat map on pixel size of an image; (b) noise of the created point cloud surface.
Sensors 23 05878 g008
Figure 9. Automatically counted pixel number.
Figure 9. Automatically counted pixel number.
Sensors 23 05878 g009
Figure 10. Some obvious errors caused by the detection algorithm.
Figure 10. Some obvious errors caused by the detection algorithm.
Sensors 23 05878 g010
Table 1. The dimensions of each layer in the encoder.
Table 1. The dimensions of each layer in the encoder.
No. of Layer
in Encoder
DimensionNo. of Layer
in Decoder
Dimension
1224 × 224 × 64528 × 28 × 256
2112 × 112 × 128456 × 56 × 256
356 × 56 × 2563112 × 112 × 64
428 × 28 × 5122224 × 224 × 32
514 × 14 × 5121448 × 448 × 32
Table 2. The comparison between three models.
Table 2. The comparison between three models.
Baseline U-NetVGG16 + BCE LossVGG16 + Focal Loss
Average Precision0.6160.4320.566
Average Recall0.5820.6030.670
F1 Score0.5980.5030.613
Table 3. Categories for damage on slab.
Table 3. Categories for damage on slab.
Description of
Typical Damage
Approximate Crack Width LimitChange in Offset in
3 m Straight Edge
Damage Category
Hairline crack<0.3 mm<8 mm0
Fine crack<1.0 mm<10 mm1
Distinct crack<2.0 mm<15 mm2
Wide crack2–4 mm15–25 mm3
Gaps in slab4–10 mm>25 mm4
Table 4. Validation result from IMG_5403.
Table 4. Validation result from IMG_5403.
Crack
No.
Ground Truth (mm)Pixel Scale FactorAutomatically Counted Pixel NumberCalculated Width (mm)Error (mm)Absolute Error (mm)
302.500.000329122.500.000.00
342.500.000392102.48−0.020.02
331.400.00033271.470.070.07
243.000.000384133.150.150.15
257.000.000337326.82−0.180.18
283.000.000329132.70−0.300.30
293.000.000329132.70−0.300.30
263.500.000337183.830.330.33
322.500.000332102.10−0.400.40
232.000.00033271.47−0.530.53
272.500.000337153.190.690.69
193.000.00039592.25−0.750.75
204.000.000395112.75−1.251.25
352.500.00039251.24−1.261.26
163.000.000511144.521.521.52
143.000.000511154.841.841.84
182.500.000528155.012.512.51
173.000.000528186.013.013.01
Table 5. Average error for each image (mm).
Table 5. Average error for each image (mm).
IMG
5338
IMG
5339
IMG
5341
IMG
5343
IMG
5347
IMG
5350
IMG
5399
IMG
5403
IMG
5407
IMG
5425
Average errors for cracks less than 1.5 mm0.260.490.530.450.530.420.360.450.710.60
Average errors for all cracks0.900.740.570.560.720.520.660.840.801.01
Average widths3.863.352.722.513.544.043.632.992.284.27
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, Y.; Shen, X.; Linke, J.; Wang, Z.; Barati, K. Quantification of Structural Defects Using Pixel Level Spatial Information from Photogrammetry. Sensors 2023, 23, 5878. https://doi.org/10.3390/s23135878

AMA Style

Guo Y, Shen X, Linke J, Wang Z, Barati K. Quantification of Structural Defects Using Pixel Level Spatial Information from Photogrammetry. Sensors. 2023; 23(13):5878. https://doi.org/10.3390/s23135878

Chicago/Turabian Style

Guo, Youheng, Xuesong Shen, James Linke, Zihao Wang, and Khalegh Barati. 2023. "Quantification of Structural Defects Using Pixel Level Spatial Information from Photogrammetry" Sensors 23, no. 13: 5878. https://doi.org/10.3390/s23135878

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop