Next Article in Journal
A Model of Multi-Finger Coordination in Keystroke Movement
Next Article in Special Issue
A Multi-Sensing IoT System for MiC Module Monitoring during Logistics and Operation Phases
Previous Article in Journal
Thermal-Adaptation-Behavior-Based Thermal Sensation Evaluation Model with Surveillance Cameras
Previous Article in Special Issue
Study on Improvement of Radio Propagation Characteristics of Cast Iron Boxes for Water Smart Meters
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Skeleton-Based Activity Recognition for Process-Based Quality Control of Concealed Work via Spatial–Temporal Graph Convolutional Networks

1
Department of Building and Real Estate, The Hong Kong Polytechnic University, Hong Kong, China
2
School of Civil and Environmental Engineering, Harbin Institute of Technology, Shenzhen 518055, China
3
Technology Department, The Second Construction Company Ltd. of China Construction Second Bureau, Shenzhen 518000, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(4), 1220; https://doi.org/10.3390/s24041220
Submission received: 14 December 2023 / Revised: 25 January 2024 / Accepted: 12 February 2024 / Published: 14 February 2024
(This article belongs to the Special Issue AIoT for Building Construction and Maintenance Engineering)

Abstract

:
Computer vision (CV)-based recognition approaches have accelerated the automation of safety and progress monitoring on construction sites. However, limited studies have explored its application in process-based quality control of construction works, especially for concealed work. In this study, a framework is developed to facilitate process-based quality control utilizing Spatial–Temporal Graph Convolutional Networks (ST-GCNs). To test this model experimentally, we used an on-site collected plastering work video dataset to recognize construction activities. An ST-GCN model was constructed to identify the four primary activities in plastering works, which attained 99.48% accuracy on the validation set. Then, the ST-GCN model was employed to recognize the activities of three extra videos, which represented a process with four activities in the correct order, a process without the activity of fiberglass mesh covering, and a process with four activities but in the wrong order, respectively. The results indicated that activity order could be clearly withdrawn from the activity recognition result of the model. Hence, it was convenient to judge whether key activities were missing or in the wrong order. This study has identified a promising framework that has the potential to the development of active, real-time, process-based quality control at construction sites.

1. Introduction

A certain type of construction work usually consists of several sequential activities, and the quality of work highly depends on the right order of activities. Subsequent activities may eventually cover the outcome of one activity, and the covered activity is called concealed work. It is difficult to assess in a non-destructive manner whether the concealed work has been carried out or not once the whole job has been completed. Thus, timely auditing of the process and results of the concealed work is vital to ensure that the work can be completed in steps, is compliant with procedures, and can be completed on time. Take the plastering work in decoration engineering as an example, which mainly contains four activities: surface preparation, fiberglass mesh covering, screeding, and coat application. The quality of each activity affects the quality of the follow-up activities. In particular, the fiberglass meshes can efficiently prevent cracks in the coat. However, the finishing coat will cover the fiberglass meshes after the whole process is completed, making it difficult to non-destructively detect whether the fiberglass meshes have been applied or not. Based on activity identification technologies, the order of activities can be identified and assessed automatically, which significantly facilitates progress management and quality control within the construction industry, especially for concealed work.
Due to the advancements in emerging technologies, automated real-time activity recognition systems for the construction industry have been developed in the last two decades [1]. Based on the types of sensors implemented, from both hardware and software perspectives, automated activity recognition methods could be divided into three major categories: kinematic-based methods, computer vision (CV)-based methods, and audio-based methods [2]. Kinematic-based methods use sensors such as accelerometers, gyroscopes, and inertial measurement units to recognize activities [3]. Audio-based methods mainly rely on recording sound patterns of equipment performing certain tasks [4,5]. CV-based methods identify activities by processing images or videos captured by different types of cameras [6]. Compared with the other two methods, CV-based methods are free of sensor–human contact and provide visual perception. Moreover, a large amount of visual data is affordable and easy to collect, since video surveillance is widely used for progress recording and site security.
Even though CV-based techniques are increasingly being used in construction, there is still significant untapped potential in this area. By far, CV-based techniques within the construction industry are mainly applied in safety [7,8,9,10,11,12,13] and progress monitoring [14,15,16]. Regarding construction quality control, most of these studies focus on surface quality evaluation, such as cracks, and few focus on the quality management of concealed work [17].
This study developed a CV-based framework that automatically identifies construction work activities and evaluates their orders. The framework is expected to be of particular value in process-based quality control, especially for concealed work. More specifically, a skeleton-based action recognition method, the Spatial–Temporal Graph Convolutional Network (ST-GCN), was innovatively employed for process-based quality control at construction sites. ST-GCN can effectively capture both spatial and temporal information by operating on skeleton joints along both space and time, and its efficiency and robustness in activity recognition have been proven [18]. ST-GCN and its variants have achieved many encouraging results [19,20,21] and have become one of the most frequently used methods for activity recognition [22].
In summary, the main contributions of this paper are as follows:
(1)
Provide a practical activity-order evaluation framework based on ST-GCNs which can facilitate process-based construction quality control. The experimental results highlight the accuracy of the approach and the feasibility of the framework.
(2)
Release a new plastering work video dataset containing both RGB and skeleton data to verify our model.
The rest of the paper is organized as follows. In Section 2, we introduce related works in the literature. Methodologies are described in Section 3. Section 4 presents the model to identify four primary activities in the plastering work. Section 5 covers an application example of the constructed model in order evaluation of the four activities, followed by discussions and conclusions in Section 6.

2. Related Works

This section first introduces existing research works focusing on CV applications in construction. Then, the state-of-the-art methods for the considered problem, especially in skeleton-based human activity recognition, are reviewed.

2.1. Computer Vision in Construction

CV techniques can be grouped into three categories: object detection, object tracking, and action recognition [23]. These techniques have accelerated the automation of several construction management tasks that require extensive human involvement, including safety monitoring, progress monitoring, productivity analysis, and quality control [24].
CV plays a vital role in site safety management. It can help monitor unsafe conditions and actions continuously at a job site using visual data [25]. Seo et al. [23] derived a general framework for CV-based safety monitoring comprising three categories of approaches: scene-based, location-based, and action-based risk identification. Zhang et al. [26] further divided the application of CV techniques in workers’ health and safety supervision into two aspects, i.e., workers themselves and workers’ interactions with the site. Guo et al. [27] classified the application of CV in site safety management into three different levels: (1) detection, recognition, and tracking; (2) assessment; and (3) prediction. They suggested the second and third levels as future research directions. In light of advances made with deep learning, Fang et al. [28] reviewed its integration with CV to support behavior-based safety programs in construction.
Automated progress monitoring comprises three stages, i.e., (1) data acquisition, (2) as-built modeling, and (3) progress assessment [29]. Literature reviews in this domain have focused on at least one of the three stages. Omar et al. [30] presented a review of the acquisition of as-built data. Typical vision-based technologies used for site data collection include digital cameras, video cameras, laser scanners, and range imaging (or RGB-D cameras) [29]. Ma et al. [31] comprehensively studied techniques to generate as-built models. According to the algorithm and method employed, the as-built models can be classified into point-cloud models, voxel models, mesh models, surface models, and BIM models [29]. Progress assessment compares the as-built and as-planned models [32]. By far, three methods have been established in comparison with the as-planned status, including point cloud comparison, feature detection, and surface matching [33].
Productivity is defined as the ratio of output (e.g., the soil amount excavated and transported) to input resources (e.g., equipment hours on the worksites) [34]. Productivity analysis measures construction input and output, while progress monitoring focuses on construction output. Most of the literature on automated productivity analysis dives particularly into productivity analysis of heavy equipment. Kim et al. [35] combined vision-based context reasoning and construction process simulation for productivity analysis of an earthmoving activity. Chen et al. [36] proposed a framework that automatically recognizes activities and analyzes the productivity of multiple excavators.
The use of CV in quality control is quite limited compared with its application in safety management and progress monitoring. By far, most studies have focused on surface quality assessment [37,38,39,40], but process-based quality control is understudied [41]. Luo et al. [17] provided a comprehensive literature review on digital technology for construction management from the perspective of the quality inspection process. They concluded that little attention was paid to the quality control of concealed works apart from surface quality assessment and measurements on concrete and steel. Quality control of concealed works in early construction projects was achieved mainly through enhanced collaboration and communication between project participants. Construction contractors submit applications to the quality inspection department for quality review and approval by the relevant personnel before concealed work commencement [42], which requires complex manual records and is very labor-intensive. Although Zhong et al. [43] proposed a method of ensuring the quality of construction by completing task lists, it is still difficult to ensure compliance with workers’ work behavior. Therefore, the existing research lacks a convenient and quick method for monitoring the quality of concealed works. This paper proposed a novel framework utilizing skeleton-based activity recognition, which has the potential to achieve a standardized inspection of the construction process and, thus, facilitate quality control of concealed works.

2.2. Skeleton-Based Human Activity Recognition

Human activity recognition based on human skeleton sequences has been extensively explored in recent years. Researchers usually represent the skeleton data with 2D or 3D human joint coordinates of all frames. Early studies on human skeletal action recognition methods generally focused on hand-crafted features [44].
Due to the abundant data and increased computing power, deep learning has become increasingly popular in the past few years. Deep learning has facilitated skeleton-based human action recognition by leveraging different aspects of skeleton data [44]. More specifically, recurrent neural network (RNN)-based methods enhance the order of joint coordinates by treating input skeleton data as time series [45]. Convolutional neural network (CNN)-based methods utilize spatial information by mapping skeleton sequences into pseudo-images representing temporal dynamics and skeleton joints in rows and columns, respectively [46]. Graph neural network (GNN)-based methods make use of both spatial and temporal data by exploiting information contained in the natural topological graph structure of the human skeleton [18]. GNN-based methods, in particular graph convolutional networks (GCNs), have become the most widespread method of performing skeleton-based action recognition. Models making use of GCN were first introduced in skeleton-based action recognition by Yan et al. [18], and they are usually referred to as Spatial–Temporal Graph Convolutional Networks (ST-GCNs).
In terms of skeleton-based activity recognition in construction, previous studies have mainly focused on identifying workers’ unsafe or non-ergonomic behavior [47]. Based on extracted human postures and pre-defined rules, Ray and Teizer [48] presented a real-time system that could determine whether a worker had the correct posture while executing characteristic construction work tasks. Guo et al. [49] use skeleton-based image technology to improve construction safety management by reducing dynamic movements to static poses in order to identify and warn workers of unsafe behavior. To the best of the authors’ knowledge, the application of skeleton-based activity recognition in construction quality control has not been reported. Moreover, the application of skeleton-based activity recognition can now benefit from the rapid advancement of image processing and skeleton extraction algorithms.

3. Methodology

3.1. Framework Overview

To further explore the application of CV-based activity recognition techniques in process-based quality control in the construction industry, a methodology framework is proposed in this study. The method has the potential to automatically identify activities in particular construction works and assess their orders based on human skeletons without the need for previous knowledge of these activities. As shown in Figure 1, the real-time activity recognition and order assessment framework mainly consists of two algorithms for skeleton extraction and activity recognition, respectively. The video data collected from surveillance cameras are pre-processed before being put into the skeleton extraction algorithm to estimate the skeleton in every video frame. The extracted skeletons and labels of video clips are then used to train an ST-GCN model for action classification. Then, with each category of actions identified, the order of activities in an action sequence can be assessed automatically. Each procedure involved in the process shown in Figure 1 is described in detail in the following subsections. Specifically, this framework can be applied to different video scenes, regardless of the image background, the angle, and the distance of the camera.

3.2. Dataset Preparation

There are four steps to prepare the dataset, i.e., video collection, video segmentation, data cleaning, and annotation, which are described as follows:
(1)
Video collection: To feasibly automatically recognize activities of particular construction works in real scenarios with the proposed method, the video dataset was collected based on surveillance videos from real construction sites.
(2)
Video segmentation: Video segmentation is the process of slicing a continuous video into discrete portions for feature extraction. The videos were first converted to a frame rate of 30 frames per second (FPS), and then divided into segments of consecutive frames according to the categories of activities to build a dataset for each activity.
(3)
Data cleaning: Invalid frames where the body or the operation of the worker was not fully captured by the camera or where irrelevant people were captured were removed.
(4)
Annotation: The activity class labels were assigned to video segments according to the categories of activities. This step was intended to ensure that the skeletons over a period would correctly represent actual construction activities. In addition, this serves as the ground truth for the learning algorithm [3].

3.3. Skeleton Extraction

A skeleton graph comprises human joints and joint connectivity, representing human body structures, as illustrated in Figure 2. There are multiple tools to extract skeleton graphs, such MediaPipe, YOLO, OpenPose, etc. Of these, OpenPose, as the first pose estimation method in the deep learning era, has gained widespread usage in various applications. Its simplicity and specific design for pose estimation make it efficient and effective, which allows for real-time skeleton-based human detection [50], particularly in the context of construction sites. OpenPose outputs a skeleton graph as an array with values (X, Y, C) of eighteen joints for two-dimensional (2D) cases. X and Y represent the coordinates on the X-axis and Y-axis of the video frame, respectively, and C is the confidence score of the detected joint. Thus, the estimated joint locations and confidence scores were used as the input, and the raw RGB frames were discarded. Figure 2 shows a graphical explanation of the whole-body joints extracted from OpenPose for 2D cases. Figure 3 illustrates the skeletons corresponding to an action sequence.

3.4. ST-GCN-Based Activity Recognition

Spatial–Temporal Graph Convolutional Networks (ST-GCNs) were introduced by Yan et al. [18]. A spatial–temporal graph consists of a sequence of skeletons with connections between the same joint in consecutive frames to track joint movement over time, as shown in Figure 4. Formally, a spatial–temporal graph can be represented mathematically as G = (V, E), which is a graph extending in both the spatial and the temporal dimensions. V =   { Vti|t = 1, …, T, i = 1, …, N} denotes the node set of all the joints in a skeleton sequence, where T represents the number of frames and N = 18 is the number of joints in each skeleton. The connection set E consists of two subsets, including the intra-skeleton connection at each frame and the intra-frame connections. The intra-skeleton connections are denoted as ES = {(Vti, Vtj)∣(i, j) ∈ H}, where H is the set of connections that represent human body structures. The intra-frame connections EF = {(Vti, V(t+1) i) ∣t = 1, …, T, i = 1, …, N} include all the connections of the same joints in consecutive frames, as shown in Figure 4. The intra-skeleton connections (ES) of each frame and the intra-frame connections (EF) comprise the spatial and temporal features of the skeleton sequence, respectively.
An ST-GCN comprises alternative spatial convolution blocks followed by temporal convolution blocks. Spatial convolution blocks are direct applications of the GCNs, which extend standard image convolution to the cases of graphs. A comparison between convolution on 2D images and on skeleton graphs is vividly shown in Figure 5. On images, the convolution is achieved by adopting a rectangular n × n receptive field, i.e., a so-called kernel, which defines the neighbors of a central location. On graphs, the nodes within a specific distance D compose the neighbor set B ( v t i ) = { v t j | d ( v t j , v t i ) D } of a node v t i . The concerning node v t i is called the root node. In this work, the neighbor set of a root node is defined as nodes within a distance of one to the root node. The neighbor set is further divided into a fixed number of K subsets according to partition strategies, where each subset has a numeric label. Then, the weight function, which provides the weights of each node in the neighbor set for the purpose of computing the convolution, is implemented by mapping nodes in the neighbor set to their subset labels. One of the three typical partition strategies, i.e., uni-labeling, distance partitioning, and spatial configuration partitioning, can be employed according to the modeling capacity and recognition performance, as shown in Figure 6. A more detailed explanation can be found in Ref. [18]. In this study, utilizing the distance portioning strategy, the neighbor set was partitioned into two subsets, one referring to the root node and another comprising the remaining neighbor nodes, as illustrated in Figure 6b. The spatial GCN was extended to the spatial–temporal domain by including temporally connected joints in the neighbor set by B ( v t i ) = { v q j | d ( v t j , v t i ) D ,   | q t | Γ / 2 } , where Γ is the temporal kernel size, which controls the frames to be included in the neighbor set.

4. Preliminary Experiments for Activity Recognition

To validate the effectiveness of the proposed framework, computational experiments exemplified by the plastering work were performed. As a preliminary step, an ST-GCN-based model was trained to recognize different categories of activities in plastering works, which laid the foundation for the action order assessments in Section 5.

4.1. Data Collection and Pre-Process

The plastering work usually consists of the following four activities:
(1)
Surface area preparation, including cleaning the wall, removing all dust, and applying an interface agent;
(2)
Covering fiberglass meshes to prevent cracks;
(3)
Screeding to guide the even application of plaster;
(4)
Applying the coat to make the surface uniform and level.
So far, there is no particular video dataset related to plastering works. The dataset in this study was collected from cameras at construction sites. It consisted of videos covering the four primary activities described above, as exemplified in Figure 7. Specifically, in each collection setup, the angle, height, and distance at which the camera was placed were not fixed, since, in real life, the camera angle and distance are not fixed. Moreover, the positions of the workers were not static during the plastering work process, which led to various positions of workers within the frame. Hence, the collected videos included a variety of perspectives. Moreover, the collected video covered multiple workers to exclude the influence of personal operating habits. After the videos were collected, the data pre-processing procedures stated in Section 3.2 were employed to prepare the dataset. Finally, we collected 89 valid videos in total. The numbers of videos for each category of activities and duration range are shown in Table 1.

4.2. Feature Extraction

As stated in Section 3.3, the publicly available OpenPose toolbox was employed in this study to extract 2D coordinates (X, Y) and confidence scores (C) of 18 human joints in every video frame. Then, each joint was represented with a tuple of (X, Y, C), and a skeleton was recorded as an array of 18 tuples. Figure 8 vividly displays the histogram of confidence scores of each joint extracted from the dataset, in which the result of each joint is plotted at the corresponding position in the skeleton. It can be seen that the algorithm can successfully track most of the joints in the main body, like the left and right shoulders, hips, and knees, while the track effect for the left and right wrist is comparatively poor. It is noteworthy that, to be consistent with the reality that some joint recognition results of the surveillance videos can be invalid, the video frames in which the joint features were extracted with low confidence scores were kept.

4.3. Model Training Details

A huge number of videos can efficiently prevent overfitting. However, collecting and annotating a huge dataset is time-consuming and labor-intensive. The dataset was extended to fully utilize the collected videos by extracting different subsequences from a sequence using the sliding window method. Each subsequence composed a video clip, intended to be long enough to ensure the activity was recognizable. In this study, a window of T = 128 was used, moving ahead one frame each time. The feature extraction is summarized in Figure 9. The dataset consisted of 26,243 clips for surface preparation, 26,411 clips for mesh covering, 14,532 clips for screeding, and 74,141 clips for coat application.
In this study, we employed L2-norm as the loss function, since it was preferred in most of the activity recognition and classification cases. The objective of the model training was to minimize the loss function by calculating the gradient with respect to the model parameter. In this study, Adam was used as the optimizer for the loss function, which is a popular algorithm because of its widespread usage and effectiveness in various deep learning tasks [51]. To assess the performance of the AI models during the training process, the dataset was split randomly into two subsets according to fivefold cross-validation: 80% of the video clips for training and 20% for validation.
The training process was deployed on a cloud CentOS system with a memory of 36 GB and an NVIDIA Tesla V100-SXM2 GPU with a memory of 32 GB. The critical code was programmed by Python 3 and its popular deep learning library, named Pytorch. Prior to training, the hyper-parameter initial learning rate was determined to be 10 5 for mini-batch learning.

4.4. Network Architecture

The ST-GCN shown in Figure 10 comprised ten layers of ST-GCN units. The first four layers had 64 output channels, the second three layers had 128 output channels each, and the last three layers had 256 output channels. The convolutional kernel size was set to one. Two techniques were used to avoid overfitting: the residual network mechanism and a dropout layer with a 0.5 drop rate. The input data were represented as a tensor of (N, C, T, V, M) dimensions, where N = 128 was the batch size, C = 3 was the joint data (X, Y, C), T = 128 was the number of frames in one video clip, V = 18 was the number of skeletal joints for each subject, and M represented the number of subjects in one clip (here, M = 1). Next, the data were fed into the spatial convolutional layer to extract the spatial features of each frame. The spatial convolutional output was then fed into a temporal convolutional layer to extract the temporal features of the same joint throughout successive frames. After the stacks of ST-GCN blocks, the resulting features were fed into a fully connected layer. The model’s output was a vector representing the probability that a given input belonged to a specific class.

4.5. Performance Metrics and Evaluation

As the ST-GCN model serves as an activity classifier, four metrics were adopted to evaluate its performance, including Accuracy, Precision, Recall, and F1-score, which are typical metrics for classification problems. In a classification task, the class prediction of an item can be either True-Positive (TP), False-Positive (FP), True-Negative (TN), or False-Negative (FN). TP includes items correctly labeled as belonging to the class by the model, while FP includes those incorrectly labeled as belonging to the class. TN consists of items correctly labeled as not belonging to the class, while FN represents items incorrectly labeled as not belonging to the class by the model. Accuracy is the proportion of correct predictions, including TP and TN, among the total items examined.
Accuracy = TP + TN TP + FP + TN + FN
Precision is the number of TP divided by the total number of elements labeled as belonging to the class, i.e., the sum of TP and FP, as
Precision = TP TP + FP
Recall is defined as the number of TP divided by the total number of elements that actually belong to the class, i.e., the sum of TP and FN, as
Recall = TP TP + FN
F1-score is a function of Precision and Recall, as
F 1 = 2 1 Precision + 1 Recall
The evolution of the four performance metrics over model training epochs is shown in Figure 11. It can be noticed that the model reached stable results. The results of training and validation accuracy are listed in Table 2. In particular, the Accuracy, Precision, Recall, and F1-score of the validation set were 0.9948, 0.9890, 0.9931, and 0.9909, respectively, which reveals that the trained model was able to accurately recognize different categories of activities in the plastering work.

5. Experiments for Activity Order Assessment

Three independent video clips were used to validate the effectiveness of the constructed ST-GCN model in activity recognition and to illustrate its application in activity order assessment. The three videos represented a process with four activities in the correct order, a process without the activity of mesh covering, and a process in the wrong order, respectively. The three videos consisted of 6540 frames, 4723 frames, and 6540 frames, respectively.
The confusion matrices of the three independent testing video clips and the corresponding comparison of the prediction by the ST-GCN model and ground truth throughout the frames are shown in Figure 12, Figure 13 and Figure 14, respectively. It can be observed in the confusion matrix for each testing video that the activities of surface preparation and coat application were detected with very high accuracy (over 98%). The activity of screeding was recognized with satisfying accuracy ranging from 67% to 77%. The accuracy in identifying mesh covering was relatively low, but acceptable. Moreover, it can be seen that the recognition accuracy for mesh covering was good enough, since the processes with (Figure 12b and Figure 14b) and without mesh covering (Figure 13b) obviously showed distinguished characteristics. Hence, it was easy to judge whether the covering mesh activity was missing. Even though the constructed model might have recognized mesh covering as coat application, the actual corresponding activity order can be identified clearly from Figure 12b and Figure 14b, respectively, which proves the feasibility of this application in activity order assessment.

6. Discussion and Conclusions

6.1. Performance of Skeleton-Based Activity Order Assessment

The experiment in Section 5 has validated the feasibility of the proposed framework for activity order assessment. The constructed ST-GCN model can efficiently recognize the four categories of activities, even though the model might recognize mesh covering as coat application sometimes. In fact, in the plastering work, before covering the mesh, in order to enhance the cohesion between the fiberglass mesh and the wall, the worker needs to cover the base coat before and after covering the fiberglass mesh. And the operation of covering the base coat is almost the same as that of coat covering. Hence, it is reasonable that the model sometimes recognizes mesh covering as coat application. This deficiency can be improved by a finer activity partition of the whole process. However, the activity order can still be easily identified from the presented results. Thus, the proposed framework can contribute to the realization of contactless, cost-effective, and real-time progress-based quality control in the construction industry.
Furthermore, it was found that the prediction accuracy of the proposed method on test videos was lower than that on training and validation datasets. One of the likely reasons for this fact was that the training dataset and validation datasets were too similar to each other. Once the trained model had been tested on samples from different distributions, the accuracy might have decreased due to the issue of model generalization. This decline in prediction accuracy could be alleviated by selecting a considerable validation dataset, increasing the diversity of the training data, and employing advanced hyper-parameter optimization methods, which could be an important research direction in our future research.

6.2. Potential Applications

(1)
Process-based quality control
Conventionally, quality assurance personnel routinely conduct site inspections to ensure that the process and quality comply with regulatory requirements, which is time-consuming and labor-intensive. The proposed framework can be used to implement process-based control. In particular, for concealed work, the construction process can be automatically extracted and assessed from video data, as illustrated by the application in the plastering works in this study. Moreover, state information, including activity durations and transitions, can be further extracted to evaluate whether an operation follows its method statement [41]. For example, in compacting concrete, the cycle duration of insertion and withdrawal can be withdrawn from the model to assess whether the vibration is sufficient.
(2)
E-learning in construction training
Labor productivity highly depends on the levels of skills and abilities of the laborers. A labor training program has been developed with the scope of improving the productivity and performance of the labor force in construction, which consists of the following processes [52]:
(1)
Designing labor training exercises and training delivery methods;
(2)
Measuring the relative weights of labor training exercises;
(3)
Assessing labor competencies;
(4)
Developing a performance score system and grading scheme for laborers;
(5)
Training reinforcement.
The activity recognition and activity order assessment framework proposed in this paper can be a functional tool and is expected to facilitate the assessment of labor competencies by automatically detecting the competencies of planned construction sequences.

6.3. Limitations and Future Directions

Since the proposed approach is an early attempt at process-based quality control via ST-GCN in the construction industry, several limitations call for further investigation. Due to external factors, such as the distance and angle of the camera, the people in the surveillance video will be obscured, and sometimes occluded [53]. A solution to this limitation is the combination of vision-based techniques with sensor-based technologies. The application in a multi-person environment is another critical factor. Exploring an efficient recognition and tracking strategy for multiple people deserves further investigation.

6.4. Conclusions

This study introduced a novel approach for leveraging surveillance cameras to automatically enact process-based quality control of concealed construction works using extracted skeleton data and the ST-GCN algorithm to recognize different categories of activities. Experiments were conducted using field videos of plastering works to verify the effectiveness of the proposed method. An ST-GCN model was first trained to recognize the four primary activities in plastering works. The model achieved 99.48% accuracy on the validation set, which indicates that the model can attain relatively high accuracy in terms of recognizing activities in plastering works. Then, extra experiments were designed to illustrate that the order of activities can be assessed based on the activity-recognizing results. These results indicate that the proposed CV-based activity recognition framework could be a valuable tool in automatic quality control, and may contribute to more convenient and effective quality management of concealed work.

Author Contributions

Conceptualization, H.L.; methodology, L.X.; software, T.P.; validation, L.X. and X.Y.; formal analysis, R.G. and X.Y.; investigation, L.X. and T.P.; resources, R.G. and T.P.; data curation, L.X. and R.G.; writing—original draft preparation, L.X. and R.G.; writing—review and editing, R.G., T.P., and X.Y.; visualization, L.X. and T.P.; supervision, H.L.; project administration, H.L. All authors have read and agreed to the published version of the manuscript.

Funding

Our research was funded by Huawei International Co. Limited. It is also to be noted that the video dataset collection was supported by a research project (KJYFKT2022002) of the Second Construction Company Ltd. of China Construction Second Bureau.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the worker(s) to publish this paper.

Data Availability Statement

Some or all data, models, or code that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

Author T.P. was employed by the Second Construction Company Ltd. of China Construction Second Bureau. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Asadi, K.; Ramshankar, H.; Noghabaei, M.; Han, K. Real-time image localisation and registration with BIM using perspective alignment for indoor monitoring of construction. J. Comput. Civ. Eng. 2019, 33, 04019031. [Google Scholar] [CrossRef]
  2. Sherafat, B.; Ahn, C.R.; Akhavian, R.; Behzadan, A.H.; Golparvar-Fard, M.; Kim, H.; Lee, Y.-C.; Rashidi, A.; Azar, E.R. Automated methods for activity recognition of construction workers and equipment: State-of-the-art review. J. Constr. Eng. Manag. 2020, 146, 03120002. [Google Scholar] [CrossRef]
  3. Akhavian, R.; Behzadan, A.H. Smartphone-based construction workers’ activity recognition and classification. Autom. Constr. 2016, 71, 198–209. [Google Scholar] [CrossRef]
  4. Sherafat, B.; Rashidi, A.; Asgari, S. Sound-based multiple-equipment activity recognition using convolutional neural networks. Autom. Constr. 2022, 135, 104104. [Google Scholar] [CrossRef]
  5. Rashid, K.M.; Louis, J. Activity identification in modular construction using audio signals and machine learning. Autom. Constr. 2020, 119, 103361. [Google Scholar] [CrossRef]
  6. Zhang, S.; Wei, Z.; Nie, J.; Huang, L.; Wang, S.; Li, Z. A review on human activity recognition using vision-based method. J. Healthc. Eng. 2017, 2017, 3090343. [Google Scholar] [CrossRef] [PubMed]
  7. Han, S.; Lee, S.; Peña-Mora, F. Vision-based detection of unsafe actions of a construction worker: Case study of ladder climbing. J. Comput. Civ. Eng. 2013, 27, 635–644. [Google Scholar] [CrossRef]
  8. Ding, L.; Fang, W.; Luo, H.; Love, P.E.; Zhong, B.; Ouyang, X. A deep hybrid learning model to detect unsafe behavior: Integrating convolution neural networks and long short-term memory. Autom. Constr. 2018, 86, 118–124. [Google Scholar] [CrossRef]
  9. Kim, K.; Cho, Y.K. Effective inertial sensor quantity and locations on a body for deep learning-based worker’s motion recognition. Autom. Constr. 2020, 113, 103126. [Google Scholar] [CrossRef]
  10. Khan, N.; Saleem, M.R.; Lee, D.; Park, M.W.; Park, C. Utilising safety rule correlation for mobile scaffolds monitoring leveraging deep convolution neural networks. Comput. Ind. 2021, 129, 103448. [Google Scholar] [CrossRef]
  11. Xu, W.; Wang, T.-K. Dynamic safety prewarning mechanism of human–machine–environment using computer vision. Eng. Constr. Archit. Manag. 2020, 27, 1813–1833. [Google Scholar] [CrossRef]
  12. Zhu, D.C.; Wen, H.Y.; Deng, Y.C. Pro-active warning system for the crossroads at construction sites based on computer vision. Eng. Constr. Archit. Manag. 2020, 27, 1145–1168. [Google Scholar] [CrossRef]
  13. Piao, Y.; Xu, W.; Wang, T.-K.; Chen, J.-H. Dynamic fall risk assessment framework for construction workers based on dynamic Bayesian network and computer vision. J. Constr. Eng. Manag. 2021, 147, 04021171. [Google Scholar] [CrossRef]
  14. Rashid, K.M.; Louis, J. Automated activity identification for construction equipment using motion data from articulated members. Front. Built Environ. 2020, 5, 144. [Google Scholar] [CrossRef]
  15. Slaton, T.; Hernandez, C.; Akhavian, R. Construction activity recognition with convolutional recurrent networks. Autom. Constr. 2020, 113, 103138. [Google Scholar] [CrossRef]
  16. Deng, H.; Hong, H.; Luo, D.; Deng, Y.; Su, C. Automatic indoor construction process monitoring for tiles based on BIM and computer vision. J. Constr. Eng. Manag. 2020, 146, 04019095. [Google Scholar] [CrossRef]
  17. Luo, H.; Lin, L.; Chen, K.; Antwi-Afari, M.F.; Chen, L. Digital technology for quality management in construction: A review and future research directions. Dev. Built Environ. 2022, 12, 100087. [Google Scholar] [CrossRef]
  18. Yan, S.; Xiong, Y.; Lin, D. Spatial temporal graph convolutional networks for skeleton-based action recognition. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  19. Li, M.; Chen, S.; Chen, X.; Zhang, Y.; Wang, Y.; Tian, Q. Actional-structural graph convolutional networks for skeleton-based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  20. Gao, X.; Hu, W.; Tang, J.; Liu, J.; Guo, Z. Optimised skeleton-based action recognition via sparsified graph regression. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019. [Google Scholar]
  21. Shi, L.; Zhang, Y.; Cheng, J.; Lu, H. Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  22. Duan, W.; Wang, Y.; Li, J.; Zheng, Y.; Ning, C.; Duan, P. Real-time surveillance-video-based personalised thermal comfort recognition. Energy Build. 2021, 244, 110989. [Google Scholar] [CrossRef]
  23. Seo, J.; Han, S.; Lee, S.; Kim, H. Computer vision techniques for construction safety and health monitoring. Adv. Eng. Inform. 2015, 29, 239–251. [Google Scholar] [CrossRef]
  24. Xu, S.; Wang, J.; Shou, W.; Ngo, T.; Sadick, A.-M.; Wang, X. Computer vision techniques in construction: A critical review. Arch. Comput. Methods Eng. 2021, 28, 3383–3397. [Google Scholar] [CrossRef]
  25. Paneru, S.; Jeelani, I. Computer vision applications in construction: Current state, opportunities & challenges. Autom. Constr. 2021, 132, 103940. [Google Scholar]
  26. Zhang, M.; Shi, R.; Yang, Z. A critical review of vision-based occupational health and safety monitoring of construction site workers. Saf. Sci. 2020, 126, 104658. [Google Scholar] [CrossRef]
  27. Guo, B.H.; Zou, Y.; Fang, Y.; Goh, Y.M.; Zou, P.X. Computer vision technologies for safety science and management in construction: A critical review and future research directions. Saf. Sci. 2021, 135, 105130. [Google Scholar] [CrossRef]
  28. Fang, W.; Love, P.E.; Luo, H.; Ding, L. Computer vision for behaviour-based safety in construction: A review and future directions. Adv. Eng. Inform. 2020, 43, 100980. [Google Scholar] [CrossRef]
  29. Reja, V.K.; Varghese, K.; Ha, Q.P. Computer vision-based construction progress monitoring. Autom. Constr. 2022, 138, 104245. [Google Scholar] [CrossRef]
  30. Omar, T.; Nehdi, M.L. Data acquisition technologies for construction progress tracking. Autom. Constr. 2016, 70, 143–155. [Google Scholar] [CrossRef]
  31. Ma, Z.; Liu, S. A review of 3D reconstruction techniques in civil engineering and their applications. Adv. Eng. Inform. 2018, 37, 163–174. [Google Scholar] [CrossRef]
  32. Ekanayake, B.; Wong, J.K.W.; Fini, A.A.F.; Smith, P. Computer vision-based interior construction progress monitoring: A literature review and future research directions. Autom. Constr. 2021, 127, 103705. [Google Scholar] [CrossRef]
  33. Braun, A.; Tuttas, S.; Borrmann, A.; Stilla, U. Improving progress monitoring by fusing point clouds, semantic data and computer vision. Autom. Constr. 2020, 116, 103210. [Google Scholar] [CrossRef]
  34. Kim, J.; Chi, S. Multi-camera vision-based productivity monitoring of earthmoving operations. Autom. Constr. 2020, 112, 103121. [Google Scholar] [CrossRef]
  35. Kim, H.; Bang, S.; Jeong, H.; Ham, Y.; Kim, H. Analysing context and productivity of tunnel earthmoving processes using imaging and simulation. Autom. Constr. 2018, 92, 188–198. [Google Scholar] [CrossRef]
  36. Chen, C.; Zhu, Z.; Hammad, A. Automated excavators activity recognition and productivity analysis from construction site surveillance videos. Autom. Constr. 2020, 110, 103045. [Google Scholar] [CrossRef]
  37. Maalek, R.; Lichti, D.D.; Ruwanpura, J.Y. Automatic Recognition of Common Structural Elements from Point Clouds for Automated Progress Monitoring and Dimensional Quality Control in Reinforced Concrete Construction. Remote Sens. 2019, 11, 1102. [Google Scholar] [CrossRef]
  38. Mechtcherine, V.; van Tittelboom, K.; Kazemian, A.; Kreiger, E.; Nematollahi, B.; Nerella, V.N.; Santhanam, M.; de Schutter, G.; Van Zijl, G.; Lowke, D.; et al. A roadmap for quality control of hardening and hardened printed concrete. Cem. Concr. Res. 2022, 157, 106800. [Google Scholar] [CrossRef]
  39. Iglesias, C.; Martínez, J.; Taboada, J. Automated vision system for quality inspection of slate slabs. Comput. Ind. 2018, 99, 119–129. [Google Scholar] [CrossRef]
  40. Chai, Y.T.; Wang, T.-K. Evaluation and decision-making framework for concrete surface quality based on computer vision and ontology. Eng. Constr. Archit. Manag. 2022. ahead-of-print. [Google Scholar] [CrossRef]
  41. Luo, X.; Li, H.; Yang, X.; Yu, Y.; Cao, D. Capturing and understanding workers’ activities in far-field surveillance videos with deep action recognition and Bayesian nonparametric learning. Comput.-Aided Civ. Infrastruct. Eng. 2019, 34, 333–351. [Google Scholar] [CrossRef]
  42. Chin, S.; Kim, K.; Kim, Y.-S. A process-based quality management information system. Autom. Constr. 2004, 13, 241–259. [Google Scholar] [CrossRef]
  43. Zhong, B.; Ding, L.; Luo, H.; Zhou, Y.; Hu, Y.; Hu, H. Ontology-based semantic modeling of regulation constraint for automated construction quality compliance checking. Autom. Constr. 2012, 28, 58–70. [Google Scholar] [CrossRef]
  44. Plizzari, C.; Cannici, M.; Matteucci, M. Skeleton-based action recognition via spatial and temporal transformer networks. Comput. Vis. Image Underst. 2021, 208, 103219. [Google Scholar] [CrossRef]
  45. Du, Y.; Wang, W.; Wang, L. Hierarchical recurrent neural network for skeleton based action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  46. Simonyan, K.; Zisserman, A. Two-stream convolutional networks for action recognition in videos. Adv. Neural Inf. Process. Syst. 2014, 27, 1–11. [Google Scholar]
  47. Han, S.; Lee, S. A vision-based motion capture and recognition framework for behavior-based safety management. Autom. Constr. 2013, 35, 131–141. [Google Scholar] [CrossRef]
  48. Ray, S.J.; Teizer, J. Real-time construction worker posture analysis for ergonomics training. Adv. Eng. Inform. 2012, 26, 439–455. [Google Scholar] [CrossRef]
  49. Guo, H.; Yu, Y.; Ding, Q.; Skitmore, M. Image-and-Skeleton-Based Parameterized Approach to Real-Time Identification of Construction Workers’ Unsafe Behaviors. J. Constr. Eng. Manag. 2018, 144, 04018042. [Google Scholar] [CrossRef]
  50. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  51. Luo, H.; Wang, M.; Wong, P.K.-Y.; Cheng, J.C. Full body pose estimation of construction equipment using computer vision and deep learning techniques. Autom. Constr. 2020, 110, 103016. [Google Scholar] [CrossRef]
  52. Manoharan, K.; Dissanayake, P.; Pathirana, C.; Deegahawature, D.; Silva, R. A competency-based training guide model for labourers in construction. Int. J. Constr. Manag. 2021, 23, 1323–1333. [Google Scholar] [CrossRef]
  53. Li, Z.; Li, D. Action recognition of construction workers under occlusion. J. Build. Eng. 2022, 45, 103352. [Google Scholar] [CrossRef]
Figure 1. Pipeline of the presented framework.
Figure 1. Pipeline of the presented framework.
Sensors 24 01220 g001
Figure 2. Skeleton joints provided by OpenPose.
Figure 2. Skeleton joints provided by OpenPose.
Sensors 24 01220 g002
Figure 3. Example of skeletons obtained for an action sequence.
Figure 3. Example of skeletons obtained for an action sequence.
Sensors 24 01220 g003
Figure 4. A spatial–temporal graph of a skeleton sequence with intra-skeleton and intra-frame connections, which comprise the spatial and temporal features of the skeleton sequence, respectively.
Figure 4. A spatial–temporal graph of a skeleton sequence with intra-skeleton and intra-frame connections, which comprise the spatial and temporal features of the skeleton sequence, respectively.
Sensors 24 01220 g004
Figure 5. Comparisons between convolution on 2D images and skeleton graphs: (a) CNN, (b) GCN, and (c) TCN. In the case of GCN, the receptive fields of a filter with D = 1 are drawn on the skeleton graph.
Figure 5. Comparisons between convolution on 2D images and skeleton graphs: (a) CNN, (b) GCN, and (c) TCN. In the case of GCN, the receptive fields of a filter with D = 1 are drawn on the skeleton graph.
Sensors 24 01220 g005
Figure 6. Neighbor set partition strategies: (a) uni-labeling, (b) distance partitioning, and (c) spatial configuration partitioning.
Figure 6. Neighbor set partition strategies: (a) uni-labeling, (b) distance partitioning, and (c) spatial configuration partitioning.
Sensors 24 01220 g006
Figure 7. Typical operations of each task in plastering work.
Figure 7. Typical operations of each task in plastering work.
Sensors 24 01220 g007
Figure 8. Statistical distribution of confidence scores for 18 human joints.
Figure 8. Statistical distribution of confidence scores for 18 human joints.
Sensors 24 01220 g008
Figure 9. Summary of clips for each activity.
Figure 9. Summary of clips for each activity.
Sensors 24 01220 g009
Figure 10. Architecture of spatial–temporal convolutional networks (ST-GCN).
Figure 10. Architecture of spatial–temporal convolutional networks (ST-GCN).
Sensors 24 01220 g010
Figure 11. Performance metrics over model training epochs: (a) accuracy, (b) precision, (c) recall, and (d) F1-score.
Figure 11. Performance metrics over model training epochs: (a) accuracy, (b) precision, (c) recall, and (d) F1-score.
Sensors 24 01220 g011
Figure 12. (a) Confusion matrix and (b) comparison of the prediction and ground truth throughout the frames for the independent testing video clips, representing a process with four activities in the right order. The labels ‘1’ to ‘4’ in the confusion matrix represent ‘Preparing’, ‘Covering mesh’,’ Screeding’, and ‘Applying coat’, respectively.
Figure 12. (a) Confusion matrix and (b) comparison of the prediction and ground truth throughout the frames for the independent testing video clips, representing a process with four activities in the right order. The labels ‘1’ to ‘4’ in the confusion matrix represent ‘Preparing’, ‘Covering mesh’,’ Screeding’, and ‘Applying coat’, respectively.
Sensors 24 01220 g012
Figure 13. (a) Confusion matrix and (b) comparison of the prediction and ground truth for the testing video clips, representing a process without covering mesh. The labels ‘1’ to ‘4’ in the confusion matrix represent ‘Preparing’, ‘Covering mesh’,’ Screeding’, and ‘Applying coat’, respectively.
Figure 13. (a) Confusion matrix and (b) comparison of the prediction and ground truth for the testing video clips, representing a process without covering mesh. The labels ‘1’ to ‘4’ in the confusion matrix represent ‘Preparing’, ‘Covering mesh’,’ Screeding’, and ‘Applying coat’, respectively.
Sensors 24 01220 g013
Figure 14. (a) Confusion matrix and (b) comparison of the prediction and ground truth for the testing video clips, representing a process in which the order of covering mesh and screeding was wrong. The labels ‘1’ to ‘4’ in the confusion matrix represent ‘Preparing’, ‘Covering mesh’,’ Screeding’, and ‘Applying coat’, respectively.
Figure 14. (a) Confusion matrix and (b) comparison of the prediction and ground truth for the testing video clips, representing a process in which the order of covering mesh and screeding was wrong. The labels ‘1’ to ‘4’ in the confusion matrix represent ‘Preparing’, ‘Covering mesh’,’ Screeding’, and ‘Applying coat’, respectively.
Sensors 24 01220 g014
Table 1. Summary of collected videos for ST-GCN model construction.
Table 1. Summary of collected videos for ST-GCN model construction.
ActivityNumber of
Videos
Duration in
Total (s)
Maximum
Duration (s)
Minimum
Duration (s)
Surface preparation2011891227
Fiberglass mesh covering 12115126914
Screeding297251827
Coat applying 2831854429
Table 2. Performance metrics of the ST-GCN model on the training and validation sets.
Table 2. Performance metrics of the ST-GCN model on the training and validation sets.
IndexAccuracyPrecisionRecallF1-Score
Training0.99470.98900.99300.9909
Validation0.99480.98900.99310.9909
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xiao, L.; Yang, X.; Peng, T.; Li, H.; Guo, R. Skeleton-Based Activity Recognition for Process-Based Quality Control of Concealed Work via Spatial–Temporal Graph Convolutional Networks. Sensors 2024, 24, 1220. https://doi.org/10.3390/s24041220

AMA Style

Xiao L, Yang X, Peng T, Li H, Guo R. Skeleton-Based Activity Recognition for Process-Based Quality Control of Concealed Work via Spatial–Temporal Graph Convolutional Networks. Sensors. 2024; 24(4):1220. https://doi.org/10.3390/s24041220

Chicago/Turabian Style

Xiao, Lei, Xincong Yang, Tian Peng, Heng Li, and Runhao Guo. 2024. "Skeleton-Based Activity Recognition for Process-Based Quality Control of Concealed Work via Spatial–Temporal Graph Convolutional Networks" Sensors 24, no. 4: 1220. https://doi.org/10.3390/s24041220

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop