Next Article in Journal
Real-Time Implementation of the Prescribed Performance Tracking Control for Magnetic Levitation Systems
Next Article in Special Issue
Using Deep Learning to Recognize Therapeutic Effects of Music Based on Emotions
Previous Article in Journal
Statistical Analysis of Silicon Photomultiplier Output Signals
Previous Article in Special Issue
Small UAS Online Audio DOA Estimation and Real-Time Identification Using Machine Learning
 
 
Correction published on 14 November 2023, see Sensors 2023, 23(22), 9160.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Augmented Hearing of Auditory Safety Cues for Construction Workers: A Systematic Literature Review

1
Department of Informatics, Ying Wu College of Computing, New Jersey Institute of Technology, University Heights, Newark, NJ 07102, USA
2
Glenn Department of Civil Engineering, College of Engineering, Computing and Applied Sciences, Clemson University, Lowry Hall, Clemson, SC 29634, USA
3
Department of Civil, Construction and Environmental Engineering, College of Engineering, North Dakota State University, Fargo, ND 58102, USA
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(23), 9135; https://doi.org/10.3390/s22239135
Submission received: 24 October 2022 / Revised: 15 November 2022 / Accepted: 22 November 2022 / Published: 24 November 2022 / Corrected: 14 November 2023
(This article belongs to the Special Issue Machine Learning and Signal Processing Based Acoustic Sensors)

Abstract

:
Safety-critical sounds at job sites play an essential role in construction safety, but hearing capability is often declined due to the use of hearing protection and the complicated nature of construction noise. Thus, preserving or augmenting the auditory situational awareness of construction workers has become a critical need. To enable further advances in this area, it is necessary to synthesize the state-of-the-art auditory signal processing techniques and their implications for auditory situational awareness (ASA) and to identify future research needs. This paper presents a critical review of recent publications on acoustic signal processing techniques and suggests research gaps that merit further research for fully embracing construction workers’ ASA of hazardous situations in construction. The results from the content analysis show that research on ASA in the context of construction safety is still in its early stage, with inadequate AI-based sound sensing methods available. Little research has been undertaken to augment individual construction workers in recognizing important signals that may be blocked or mixed with complex ambient noise. Further research on auditory situational awareness technology is needed to support detecting and separating important acoustic safety cues from complex ambient sounds. More work is also needed to incorporate context information into sound-based hazard detection and to investigate human factors affecting the collaboration between workers and AI assistants in sensing the safety cues of hazards.

1. Introduction

The recognition of auditory safety cues at job sites plays a vital role in preventing injuries and deaths for construction workers [1]. Research has found a significant correlation between a lack of auditory signal awareness and an increase in unsafe actions leading to fatalities due to the failure to stay vigilant in hazardous situations [2,3]. Inadequate auditory situation awareness is often caused by declining hearing capability due to the use of hearing protection and the complex nature of construction noise [3]. Therefore, the development of wearable devices capable of automated detection of acoustic safety signals has received increasing attention from the research community.
Despite the critical need for augmenting the hearing of safety cues for construction workers, research in this area is still lagging. Previous studies on wearable safety devices have been focused on employing real-time data analytics to continuously measure a wide variety of safety performance metrics [4] other than auditory signals. Typical functions of existing wearable safety devices include physiological monitoring, environmental sensing, proximity detection, and location tracking of construction hazards [5]. Adopting computer vision algorithms to detect hazards is among the most popular approaches for extracting information from images or videos [6,7,8]. This approach still involves various challenges, such as limited field of view, illuminations, and occlusions of digital cameras that could harm the widespread use of this method at complicated construction sites. Others employed kinematic sensors, such as gyroscopes and accelerometers, to record the kinematic signals of the equipment and to detect its activities [9,10]. However, using kinematic sensors fails to detect hazards in such situations when sensors cannot be directly attached to the body of the machine that vibrates during operation, such as jackhammers, concrete pumps, and concrete truck mixers. It is also expensive and complex to deploy sensing devices on every piece of construction equipment.
Audio-based hazard detection has emerged as a supplement technique that can augment workers in effectively monitoring the work environment. Processing audio signals for automated detection has meaningful implications because hazardous situations usually cause strong acoustic emissions. A variety of hazard detection models employing advanced machine learning techniques for signal processing have been developed in many other workplace environments other than construction sites [11,12,13,14,15]. They possess great potential to enable workers to quickly detect auditory safety cues that are important for their safety. These computing advances have been implemented in many different environments, including indoor and public environments [16,17,18,19,20,21,22,23,24,25,26], medical and health care systems [27,28], and working environments [29,30,31,32,33]. However, current research does not sufficiently address automated auditory surveillance for construction safety.
The past review papers on auditory surveillance in construction were only focused on analyzing existing methods to support the monitoring of work tasks [34] and understanding workers’ use of wearable technology to prevent hearing loss [4]. This paper is the first attempt aiming to systematically review advanced Artificial Intelligence (AI) techniques in auditory signal processing and to provide insights into their potential capabilities for enhancing construction safety. The paper examines how hazardous situations can be detected by leveraging recent advances in audio signal processing and to what extent its efficiency and effectiveness can be improved. At the same time, it analyzes research and application trends and then identifies research gaps that require further research. The scope of this review includes academic articles on the significance of auditory situational awareness for construction safety and the applications of auditory surveillance in detecting hazardous situations. Moreover, a critical review of computational advances is summarized for researchers and practitioners interested in auditory signal processing.

2. Research Background

Many construction hazards produce strong acoustic emissions; thus, preserving or even improving the auditory situational awareness of construction workers could reduce the risk of fatal accidents on construction sites. This paper argues that auditory situational awareness is critical to the safety of construction workers because hearing safety-critical acoustic signals can allow timely responses to hazards. In this section, various types of construction accidents and their association with safety signals are first given. Then, the role of auditory situational awareness for construction safety is presented, followed by a critical analysis of the current state-of-the-art work on auditory surveillance in workplaces, including construction sites.

2.1. Construction Hazards

Figure 1 presents common causes of construction accidents along with the number of cases in a one-year period combined in three countries, including the US, Israel, and the UK, recorded in 2008, 2011, and 2009, respectively [35]. As shown in the figure, it is apparent that “Struck by moving/falling object” is the most common cause associated with more than 24,000 incidents, followed by “Falling from a high level” with nearly 17,000 accidents. Furthermore, there were approximately 14,500 accidents due to “Slips, trips, or falls on the same level” and “Injured while handling, lifting, or carrying”. Other types that also cause a significant number of accidents are “Strike against something fixed”, “Exposure to a harmful substance”, and “Struck by a moving vehicle”.
Another report by [36] revealed the number of fatal cases from the Occupational Safety and Health Administration (OSHA) between 1990 and 2007 for different types of construction hazards, including (a) getting struck by equipment or vehicle, (b) getting struck by part of equipment/material, (c) equipment rollover, (d) equipment in water, and (e) electrocution in contact with the equipment. Accordingly, struck by equipment/vehicle hazard was the cause of 659 cases, accounting for 87.7% of the total fatalities. In addition to struck by moving equipment, workers were also involved in accidents by the hit of equipment buckets, material being dropped or lowered during transferring, electrocutions when workers were touching equipment that contacted power lines, rollovers when the equipment was operated on steep slopes, and drowning when equipment rolled into deep water.
As noted above, equipment is one primary source of construction hazards. Table 1, adopted from Ref. [36], ranks 16 types of construction equipment by their frequencies of job-site fatal accidents. Overall, dump trucks had the highest frequency of involvement in fatality cases with 173 incidents, almost two times higher than the combined fatalities cases by backhoe excavators (second rank) and private vehicles (third rank) with 50 and 42 incidents, respectively. Following the top three vehicles, dozers, graders, front-end loaders, and forklifts also had a relatively high quantity of incidents, ranging between 30 and 40 each. Another point that could be derived from the table is that most of the accidents occurred when the equipment was moving backward. For example, 91.1% of the fatal accidents caused by dump trucks were associated with reserve movement. This percentage for many other types of equipment, including graders, loaders, and water trucks, was reported to be over 90% and it was at least 50% for excavators, dozers, front-end loaders, forklifts, tractor-trailers, compactors, and scrapers. Those statistics of accidents indicate the great hazardousness of construction equipment in the reverse direction.
Furthermore, it is worth mentioning that only a small portion of those accidents due to the reverse direction occurred when the backup alarm was not operated. For instance, the alarm was not functional in only 12.1% of the total cases for dump trucks and even 4% for backhoes. Noticeably, 26.7% of scrapers’ cases were attributed to reverse alarm failures. This is probably due to large blind areas on all sides of the equipment. According to Schultheis’s [37] report, the blind area range for scrapers is high compared with other equipment types. A scraper operator typically cannot easily see what is behind, which decreases the visibility in a reverse direction. These data suggest that a reserve alarm was largely inefficient in preventing fatal accidents. This may be due to the habituation of warnings or the impact of other ambient noise. Thus, augmenting the worker’s auditory sensing would be an urgent need.

2.2. Significant Role of Auditory Situational Awareness for Construction Safety

Of those construction hazards discussed above, many are associated with acoustical signals. Early recognition of these safety-critical sounds could be beneficial in preventing construction accidents. Figure 2 classifies construction hazards that can be detected by their associated sounds. These sound-producing hazards are: contact collision with moving machinery/vehicles, objects, collapsing/overturning, exposure to fire, and exposure to an explosion. Moving machinery/vehicles, which can cause contact hazards, are mentioned in Table 1. For the type of accident in close contact with moving objects, the hazardous scenarios include missing footboards on a scaffold, moving a crane with a load where workers are present, working with loose materials (blocks) at a height, working with façade elements on a scaffold at a height, and working with unsecured hand tools at a height. Regarding being trapped by something collapsing/overturning, hazards come from improperly secured slab formwork and improperly supported wall formwork. The awareness of these signals would help construction workers evaluate the situations and make proper responses to any potentially hazardous situations
Auditory situational awareness is understood as a cognitive process of acoustic signal recognition. According to the cognitive model of construction workers’ unsafe behaviors (CM-CWUB) developed by [2], the process of responding to a potential hazard includes five stages: (a) obtaining information, (b) understanding information, (c) perceiving responses, and (d) selecting responses and taking action. The model is useful as it can allow us to understand the mechanism of human errors as well as how unsafe behaviors are made at various cognitive stages. In the context of auditory situational awareness in construction safety, workers need to receive the signal and recognize its original source and direction (e.g., engine sound from a heavy vehicle moving towards the workers). Based on their experience, judgment will be made regarding the occurrence of a certain potential hazard and the corresponding safety actions. It is worth noticing that the first stage (obtaining information) and the fourth stage (selecting responses) are the two critical sources of unsafe behaviors of construction workers. Unfortunately, the loudness level of sound produced in the construction site due to its complicated nature is loosely defined and regulated, making it hard for workers to identify hazards. In addition, making high-stakes decisions to deal with dangerous situations is prone to errors. Due to the above barriers, cognitive assistance is necessary to allow construction workers to timely detect possible hazardous auditory events. Augmented hearing devices would be particularly needed to assist workers in evaluating situations and providing decision-making recommendations.
The ASA of construction workers is affected by various factors that can be categorized into job-site-related and worker-related factors, as summarized in Table 2. Regarding the job-site-related factors, the most significant cause of the decrease in ASA is the complexity of ambient sounds. Due to loud and complicated noises, the audible sensing of workers to hazards becomes limited. This limitation prevents construction workers from perceiving their surroundings and recognizing alerts of impending dangers. It is reported that the performance in signal recognition of normal-hearing listeners decreases as the number of concurrent sound sources increases [38]. Using hearing protection to prevent hearing loss would make further difficulty in maintaining workers’ ASA. As a result, essential sounds can be easily ignored or misidentified [39].
Additionally, worker-related factors such as gender and occupational hearing loss are reported to significantly impact the worker’s hearing capability. Research suggests that men have a higher degree of hearing difficulty than women [38]. Hearing loss in men is more prevalent because they are typically tasked to work in places exposed to loud noise [40]. Another research on the causes of hearing loss, investigated by Rigters et al., [41], found that men tend to have more hearing loss due to high frequency sounds than women. This may attribute to the severe hearing issues among male workers in construction as construction equipment noises are typically sounds with a high frequency (above 2000 Hz) [42]. In addition, workers with normal hearing still encounter a high level of hearing difficulties if exposed to hazardous noise [3]. The study in [3] showed profound challenges relating to job-related conservation for noise-exposed and hearing-impaired workers. The awareness of these issues is crucial to the future development of appropriate protective measures for noise-exposed and hearing-impaired workers. The next section will introduce the methodology adopted to develop this qualitative literature review on using auditory surveillance for hazard detection in the construction industry.

3. Methodology

This paper involves a literature review of academic articles that focus on the capabilities of advanced AI for auditory signal processing and sound surveillance in supporting construction safety. This research employed the review method proposed by [43], consisting of the following three main steps (see Figure 3): (1) defining the research scope based on identified keywords; (2) collecting the articles within the scope of this review by conducting a keyword search in the scientific databases; (3) content analysis to identify major research approaches, research gaps, and recommendations for future research. The details of each step are provided in the following sections.

3.1. Step 1: Define the Research Scope

The work scope of automated auditory surveillance is broad and diverse. There are various sorts of acoustic events present in our everyday environment. For example, the most typical acoustic events in the office come from people chatting, doors opening and closing, and computers and other devices. At the same time, an automated auditory surveillance system could have many different purposes, including automated management of safety, security, and monitoring applications. To define the research scope, three main topics closely related to automated auditory surveillance for construction safety, referring to “hearing”, “construction”, and ”audio signal processing” (see Figure 4), were identified. For this study, articles with topics or abstracts containing either “construction safety” and “hazard” combined with any of the sentences, such as “sound activities”, “workers hearing”, “audio signal processing”, and “noise”, are studied. The research used for this study is not limited to studies conducted in the United States. Based on these initial topics, we found 17 related keywords previously used in past research, including four for hearing-related, five for construction-related, and eight for audio signal processing topics. After collecting relevant keywords, we revised the initial topics into more narrowed and specialized ones. It is worth noting that a keyword can relate to one or more revised topics.

3.2. Step 2: Literature Search

In the literature search, all keywords shown in Figure 4 were used to search for related publications through the publicly available scientific databases, including Google Scholar, Web of Science, Elsevier, WILEY, ASCE, Springer, IEEE Xplore, and ACM Digital Library, published from 1998 to 2022. This process resulted in a total of 178 articles. Only the papers published in high-impact journals and conference proceedings were considered. After identifying the initial set of related publications, their title and abstracts were reviewed by filtering out the articles outside of our research scope through a content analysis thereby removing irrelevant documents. As a result of this process, 85 articles were included, which were then critically reviewed by the authors. The statistics of reviewed articles are shown in Figure 5. A total of 35 articles were found regarding the sound classification, while 22 were found for the detection of abnormal situations such as unusual sound produced by a piece of equipment. For sound recognition of workers with hearing and detecting device of surrounding situations 10 articles were found. Less than ten articles present the application of sound-based surveillance in the construction field. Only four articles study the detection of hazardous situations in the construction field and six articles were found for sound localization. Therefore, studies on the localization of sound sources to prevent hazards on the construction site are still in the early stages and are yet to be fully exploited.
To highlight the research trend on automated auditory surveillance, we summarized targeted articles every two years, as illustrated in Figure 6. The number of articles throughout the last two decades is increasing, indicating that more and more researchers are interested in sound-based surveillance. Especially within the last six years, the total number of articles from 2014 to 2019 reaches more than 40. This number proves that studies on the topic of automated auditory surveillance are crucial in today’s world.

3.3. Step 3: Content Analysis

In this step, the 85 identified articles found relevant to automated auditory surveillance for construction safety were rigorously examined. The articles were reviewed in the following aspects: (1) the role of auditory situational awareness in construction safety, (2) advanced methodologies in auditory signal processing, (3) applications of signal processing in safety surveillance, and (4) benefits and challenges. This step also involved critically analyzing the leading research gaps in previous studies. The findings of these steps are presented in the following sections.

4. Results

4.1. Audio Signal Processing Applications and Methods

4.1.1. Application Contexts of Auditory Surveillance

Auditory surveillance has been applied to detect abnormal events in various contexts. As shown in Table 3, the literature shows the implementations of sound-based technology in the following areas: home security [23], public environments [16,17,18,19,20,21,22,24,25,26], office [29], medical and health care facilities [27,28], and in industrial plants [31,32].
Public environments have received the most attention from the research community. Despite complicated noises in public, many studies achieved promising results for the surveillance of abnormal sounds such as glass breaking, screams, gunshots, and explosions. For example, Ref. [24] developed a technique for detecting shouting events in a real-life railway environment. Not only are audio events automatically detected, but the positions of the acoustic sources are also localized [17]. Audio events of gunshots could be detected based on a novelty detection approach, which offers a solution to detect abnormality in continuous audio recordings in public places [16]. Another similar architecture for acoustic surveillance of abnormal situations under different acoustic backgrounds was built to detect vocal reactions (screams, expressions of pain) and non-vocal atypical events associated with hazardous situations (gunshots and explosions) [19]. Models for the detection of abnormal acoustic events from normal background sound were also developed by several authors [20]. A few efforts have been made focusing on the detection of crimes in elevators [25] and the detection of human emotions based on verbal sounds in hazardous situations [18,26]. Ref. [29] built a system for acoustic surveillance to detect abnormal sounds from people talking, opening and closing doors, and using computers and other devices in an office.
In addition, many studies have implemented signal processing for the surveillance of healthcare facilities. For example, a system to extract sound features for medical telesurvey was developed by [27]. It can classify detected sounds into normal and abnormal types. The system’s purpose is to detect severe accidents such as falls or faintings at any place in the living area, which is useful for the surveillance of the elderly, the convalescent, or pregnant women. Furthermore, since older adults living alone potentially get into trouble when they fall and are even unable to call for assistance, a framework that detects falls by using acoustic signals by analyzing environmental sounds was proposed [28].
In industrial sectors, current research on applications of sound-based detection has focused on detecting abnormal behaviors of the machine or the equipment. Acoustic-based fault diagnosis techniques of a three-phase induction motor are presented to see if the motor is in bad or good condition [31]. Many rotating electric motors can be diagnosed using acoustic signals; this can prevent unexpected failure and can improve the maintenance of electric motors. The advantages of the proposed acoustic-based fault diagnosis technique are its non-invasive technique, low cost, and instant measurement of acoustic signals. A novel optimization technique for the unsupervised anomaly detection of sound signals using an autoencoder (AE) is proposed [32]. The goal is to detect unknown sounds without training data to identify abnormality or failure in the operation of the stereolithography 3D printer, the air blower pump, and the water pump.

4.1.2. Principles of Audio-Based Hazard Detection

The detection of hazards using acoustic data is typically based on the extraction of the following four types of information: type of sound, location of the sound, the direction of moving sound sources, and the abnormality of sound (see Figure 7). These auditory event characteristics are used as input for detecting hazardous situations. The type of sound is the most common indicator used for detecting hazardous events and differentiating abnormal sounds from normal sound events. The occurrence of abnormal sounds, such as a gunshot or an explosion, is an important indicator of a dangerous situation requiring a quick safety response. Additionally, sound localization cues, such as the location and the direction of a moving sound source, are also essential for detecting potential abnormal events. They could inform receivers whether or not they are at an unsafe distance from the hazard. Moreover, measuring the abnormality of ambient sound has been used for evaluating hazardous situations, such as a scenario where a machine is operated and any abnormal sound that occurs in the parts or in the assembly process is often regarded as an abnormal sound and will require the attention of maintenance officers. Given the fact that we may not be able to classify all the unknown abnormal sound events that occur during equipment operational noise inspection, the anomaly would be useful in cases where we are developing the abnormal noise inspection device to automate the process.
Automatic detection of auditory cues requires computational advances in processing the auditory signals. As discussed earlier, hazardous situations are detected based on auditory features, including the type of sound, sound localization, and sound abnormality. Figure 8 provides a typical procedure of sound-based hazard detection found in various literature. The process entails three steps: (1) acoustic feature extraction, (2) auditory event recognition, and (3) assessment of hazardous situations. Of those steps, recognizing auditory events in Step 2 is the most critical step in the signal-processing procedure. This paper is therefore focused on the critical review of recent computational advances in sound classification, sound localization, and the detection of sound abnormality, which are required to recognize auditory events. Common types of acoustic features to be extracted from Step 1 will also be discussed.

4.1.3. AI-Based Sound Classification

This part focuses on the classification of sounds using AI. The main processes of AI-based sound classification are depicted in Figure 9. The first and foremost step is to extract the acoustic features from audio signals. Depending on the type of sound, a specific extraction function is used to extract the features. After all the necessary acoustic features are extracted, machine learning techniques, including traditional techniques and deep learning techniques, are employed to train models for sound classification. Common methods of feature extraction employing machine learning techniques are provided below.

Feature Extraction

Feature extraction is an inevitable step because time-domain signals contain irrelevant data to directly use for sound classification. It is crucial to use appropriate features for successful classification as lousy features contain less discriminating power, which could hardly be classified. The summary of acoustic features used in sound classification is summarized in Table 4. Acoustic features can be grouped into cepstral features, spectral features, energy features, and temporal features.
Cepstral features are commonly used to define auditory signals. A cepstrum is computed by taking the Inverse Discrete Fourier Transform (IDFT) of the logarithm of the Fourier transform of a given signal. As shown in the table, there are various variations of cepstral features. Mel-Frequency Cepstral Coefficients (MFCCs) are the most popular cepstral features, which are specifically based on the power spectrum of sounds. An example of the MFCC’s representation extracted from the sound of a moving grader is shown in Figure 10a. While the positive values of MFCCs represent a sonorant sound with low-frequency signals, negative values represent a fricative sound with high-frequency signals. MFCCs have been used to extract the spectral envelope of signals in many audio and speech recognition applications [11,12,13,15,17,24,27,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67]. Many researchers used only MFCC parameters [11,12,13,45,46,47,50,51,54,55,56,57,58,59,64,67], while a few others used MFCCs in conjunction with other spectral features including Zero Crossing Rate (ZCR), spectral roll-off, and spectral centroid [27].
In speech processing, Linear Prediction Coefficients (LPC) have been widely used for the cepstral analysis of sounds since they well approximate the characteristics of vocal sounds and give a good performance in distinguishing normal talk and excited shouting [68]. The advantage of LPC is the simplicity of the algorithm, because past signals are modulated as the weighted sum of previous values to reduce an error function [61]. LPC has a variation called the Linear Predictive Cepstral Coefficient (LPCC). In an investigation [63], LPCC is obtained using a direct recursion from the LPC. Compared with LPC, LPCC can better handle sounds with sudden changes or that involve complex noises [68]. The Log Frequency Cepstral Coefficient (LFCC) is another commonly used type of cepstral feature. It is typically extracted by calculating the frequency domain of sound signals using the logarithmic filter bank. It has been shown to outperform LPC in demarcating between the vocal and the non-vocal events [68]. To address the limitations of MFCC for signals having strong temporal domain signatures, such as non-speech audio signals at low frequencies, Ref. [61] proposed another biologically feature, namely the Gammatone Frequency Cepstral Coefficient (GFCC). The uniqueness of GFCC is the use of a Gammatone filter bank, which is inspired by the human auditory filter response. A few attempts that use concurrently different cepstral and perceptual features have been undertaken when developing models for sound classification [62]. Rouas, Louradour, and Ambellouis [24] examined this approach by integrating MFCC, LPC, and PLP. The most effective feature sets from their study are the combination of MFCC and PLP. Janjua et al. [61] used the combination of MFCC, LPC, and GFCC to detect a rare event and to achieve precision and recall measures above 90%. Atrey, Maddage, and Kankanhalli [68] optimized the parameters for LPC, LPCC, and LFCC to detect various kinds of acoustic events.
Sounds can also be characterized using spectral features that are based on the frequency domain. Spectral features are the most generic features that can be applied to a wide range of signals. One example of a spectral feature is the spectral centroid, which indicates the location of the “center of mass” [27,44,62,63,65,66]. Spectral roll-off provides the same information as spectral centroid, but it is used to measure the frequency delimiting 95% of the power spectrum [17,27,44,63,65,66]. Figure 10b,c visualize an auditory signal’s spectral centroid and roll-off values. Spectral flux is another spectral measure that computes how quickly the power spectrum of a signal changes over time [44,63,65,66]. Mel spectrum represents the approximately logarithmic frequency sensitivity of human hearing. It also reduces the dimensionality of the spectrum for automatic sound classification in many applications [48,49,52,53,60]. Several other spectral features have also been used in previous studies, including spectral moments [17], spectral slope [17], spectral decrease [17], modulation spectral features [69], and spectrogram [70].
Table 4. Summary of acoustic features used in sound classification.
Table 4. Summary of acoustic features used in sound classification.
Acoustic FeaturesFeatures SetReferences
Cepstral featuresMel-Frequency Cepstral Coefficients (MFCC)[11,12,13,15,17,24,27,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67]
Linear Prediction Coefficients (LPC)[24,61,63,68]
Linear Predictive Cepstral Coefficients (LPCC)[63,68]
Log Frequency Cepstral Coefficients (LFCC)[68]
Gammatone Frequency Cepstral Coefficients (GFCC)[61]
Perceptual Linear Prediction Coefficients (PLP)[24]
Spectral featuresSpectral centroid[27,44,62,63,65,66]
Spectral roll-off[17,27,44,63,65,66]
Spectral flux[44,63,65,66]
Mel spectrum[48,49,52,53,60]
Spectral moments[17]
Spectral slope[17]
Spectral decrease[17]
Modulation spectral features[70]
Spectrogram[71]
Energy featuresShort Time Energy (STE)[44,62,63,65,66]
Discrete Wavelet Transform Coefficients (DWTC)[27]
Temporal featureZero Crossing Rate (ZCR)[15,17,27,44,62,63,65,66,68]
Other types of features used for signal processing include energy features and temporal features. Short Time Energy, which represents the energy of the short speech segment, is the most popular type of energy feature used for signal processing [44,62,63,65,66]. Short Time Energy can be represented using a wave plot, as illustrated in Figure 10d. The plot shows the loudness of the sounds along with a given time range. Another commonly used energy feature is the Wavelet-Based Coefficient (DWTC). It is computed through the wavelet transform function. Models with DWTC could diminish processing time, due to the use of fewer parameters, but showed lower performance than models with MFCC [27]. Zero Crossing Rate (ZCR) is the most popular temporal feature in speech recognition and sound retrieval [15,17,27,44,62,63,65,66,68]. It indicates the number of times when the signal amplitude crosses the zero value [27]. It is more popular for use in discriminating periodic signals from signals corrupted by noises. Figure 10e depicts the zoomed-in wave plot of 200 array columns extracted from the audio of a moving grader with two zero crossings.
Energy, temporal, and spectral features can be used simultaneously for sound classification. Valenzise et al. [17] used many acoustic features that are not too sensitive to SNR conditions, including ZCR, STE, spectral features, perceptual features, and periodicity descriptors. Moreover, they also used new features based on the auto-correlation function. Correlation features are comparable to spectral features, though they are computed starting from the auto-correlation function of each frame. The purpose of these features is to describe energy distribution over different time lags. In fact, much of the energy of impulsive noises, such as gunshots, is concentrated in the first time lag, whereas the energy of harmonic sounds, such as screams, is spread over a broader range of time lags.

Training of Sound Classification Models

Sound classification is the process of analyzing acoustic signals to recognize and classify auditory events. Machine learning is the primary approach to classifying sounds found in most of the past literature. This section will summarize all the machine learning techniques used in the previous research for sound classification. As shown in Figure 11, machine learning techniques are grouped into conventional machine learning and deep learning. The traditional machine learning techniques include Gaussian Mixture Models (GMM), Support Vector Machine (SVM), K-Means Algorithms, Hidden Markov Models (HMM), and K-Nearest Neighbor Algorithms (KNN). Of those past techniques used in classifying sound, GMM was the most widely used method, with 11 reviewed articles. The deep learning approach comprises Convolutional Neural Network (CNN), Recurrent Neural Network (RNN), and Auto Encoder (AE). CNN was found as the most commonly used technique, with eight reviewed articles.
Five different types of traditional machine-learning techniques for sound classification are listed in Table 5. GMM is a probabilistic model that fits the data points in a mixture of various finite Gaussian distributions with unknown parameters. There has been a large amount of research on the sound classification process using GMM due to its adaptability and simplicity. The sound classification GMM model developed by [16] for discriminating emotional contents and normal events showed that this method was significantly impacted by the noise in the training data. To reduce the false rates in the classification, Ref. [23] proposed adjusting GMM with an exponential criterion and a weighted least square solution. The classification performance also improved when GMM was modified using the Figueiredo and Jain algorithm [17].
Moreover, integrating a Bayesian criterion decision space with Gaussian distributions could well model the feature distributions of all the classes [24]. The use of two [25] or four [27] Gaussian components should be considered for building a classification model; although, higher numbers of components showed no improvement in the preliminary experiments conducted [27]. It is worth noting that adopting a more sophisticated multi-level classification approach will work better than single-level approaches [68]. Therefore, many studies considered GMM for classifying the hierarchical structure of sound events [19,20,71]. Finally, GMM could occasionally be used as a clustering method to generate sound data from functional representations by being trained on randomly sampled data from the training set [62].
While GMM captures the feature distributions of classes, an SVM technique aims to model a discriminative function to separate the classes [24]. It determines an optimal hyperplane that can separate the classes with minimized errors. In other words, it maximizes the margin between the decision boundary and the training samples to avoid overfitting on small training sets [21]. SVM works well for high-dimensional data and relatively small sets of labeled data points [74]. In addition, it is often applied in binary classification [28,57]. Using SVM, the selection of an appropriate kernel has a strong influence on the prediction performance. Several commonly used types of kernels are Radial Basis Function (RBF) kernel [11,30,57,72,74], linear kernel [44,74], and Gaussian kernel [73].
The K-Means algorithm, which searches for cluster centroids that minimize the Euclidean distance to the data points, was also used for auditory signal processing. This method does not require labeled samples to perform the clustering [21]; this is a great advantage over other methods for reducing the human labor necessary to train the event detection system on a new set of sounds. Thus, the K-Means algorithm is often used to perform clustering on auditory data [29,46,51]. Spherical K-Means is a modified form of the K-Means algorithm that is achieved by modifying the iterative update procedure for the K-Means algorithm [48].
Other popular traditional machine learning techniques are HMM and KNN. HMM relies on the statistical Markov model that trains a signal classification model by assuming the data to follow the Markov process. There are various applications employing HMM, such as the detection and classification system for sound surveillance [22], construction activity identification and task performance analyses [75], audio-based context recognition system [63], acoustic environment classification [45], and automatic detection of different types of acoustic events [47]. Many researchers proved it as a reliable classifier for sound classification [75]. The simplest machine-learning algorithm implemented for sound classification is probably KNN [57]. This method could properly classify linearly separable and non-linearly separable feature vectors. It can classify feature vectors using just a small number of training examples. Therefore, KNN is often used for signal processing and data classification of multi-class problems with a high recognition rate [31]. However, one noticeable drawback of KNN is time complexity, since, with a large dataset, the cost of calculating the distance between a new point and each existing point is very high, which in turn degrades the performance of the algorithm.
Recently, a Deep Neural Network (DNN) has been proven to outperform other traditional machine learning techniques in a variety of tasks for sound classification [58]. DNN is a complex feed-forward network comprising several hidden layers. DNN can produce a 54% higher performance than GMM in classifying audio concepts [59]. In another study that used DNN for detecting the typical sounds of everyday life, the classification performance was as high as 88% [58]. The use of DNN for recognizing acoustic events, such as footsteps, baby crying, motorcycle, and rain, achieved a classification accuracy of 60.3% compared with 54.8% for the conventional HMM classifier [60]. Recently, the following more complex architectures of DNN have been used for audio classification: CNN, RNN, and AE. Several studies have used Convolutional Neural Networks (CNN) with large input fields for sound classification of large audio datasets [12,15,55,62]. In a CNN network, each convolution layer generates a successively higher-level abstraction of the input and preserves its essential unique information. For sound processing applications, CNN is able to learn filters that are shifted in both time and frequency so that it can cover a large number of input fields [62]. The performance of CNN networks for signal classification is highly influenced by the optimum number of convolutional layers, which varies across different studies. Ref. [15] reported that CNN performed well with two convolutional layers, while three convolutional layers were suggested by [55]. Researchers also trained CNN models with a back-propagation mechanism consisting of pattern creation and testing processes [31]. CNN has shown impressive performance in auditory signal processing. Ref. [62] proposed to convert audio into vectors which are then trained using CNN. This approach achieved an average F1 score of 0.839. Table 6 sumarizes popular deep learning techniques for sound classification.
A few efforts have been made that use RNNs for sound classification. With RNNs, the information from previous time steps can circulate continuously inside the network through the directed cycles, where the hidden layers also act as memory [62]. Hence, RNN can capture better long temporal context information. A derivative of RNN is the Bi-Directional Long Short-Term Memory (BLSTM), which trains a special RNN by adopting a very long-ranged symmetric sequence context with a combination of nonlinear processing elements and linear feedback loops to store a long-range context. Although BLSTM has been used in a few applications to detect acoustic events with remarkable performance, its ability to convey contextual information in a long audio sequence was not really advantageous compared with CNN classifiers [76].
AE is a novel approach to acoustic novelty detection. When some of the input acoustic features are correlated, AE can learn those correlations and reconstruct the data from a compressed representation [76]. AE was reported to significantly exceed the performance of other traditional machine learning methods. The results from a study demonstrated that the AE approach outperforms GMM and HMM classifiers [77]. Furthermore, the use of an AE to minimize the reconstruction error of normal sounds could efficiently decrease the false positive rate [32].

4.1.4. Sound Localization

Sound localization refers to the ability to identify acoustic sources in terms of direction and distance. It is one of the essential acoustic parameters that enables the ability to recognize and locate hazardous events. When dangerous objects or equipment are within an unsafe distance of a construction worker, detecting the location of acoustic sources can enable them to make appropriate preventive responses. Methods to localize sound events are mainly based on calculating the difference in the arrival times of the signal. Then, the similarity measure of the signal at different times is examined in either time or frequency domain for sound localization. For the time domain, the acoustic impulses reach the microphones at varying Times of Arrival (TOA) when they are spatially distant from one another. The signal’s Direction of Arrival (DOA) is determined from the recorded time delays using the given array geometry. Each pair of microphones in the array has a projected time delay. Then, using time delays and geometry, the best estimate of the DOA is determined, while the frequency domain is the difference between the time sound pressure reaching the array geometry and is mostly used to localize higher frequency sounds. Table 7 lists the methods for sound localization found in this review, of which the details are provided below.
Calculating the Time Difference of Arrivals (TDOA) of the signal is one approach to the detection of sound location. Valenzise et al. [17] adopted the Maximum-Likelihood Generalized Cross Correlation (GCC) method and linear-correction least-square localization algorithm for estimating the TDOA of the signal. Ref. [78] utilized the Euclidean Distance (EUD) to measure the similarity in the time domain. Other alternatives include measuring the similarity between signals at different arrivals based on the frequency domain. Compared with the time domain, the frequency domain showed significant improvements in the precision for sound localization as demonstrated by Satoh et al.’s [79] study, where they computed the cross-correlation in the frequency domain based on the Normalized Cross Correlation (NCC) method. Tarzia et al. [80] used Fast Fourier Transform (FFT) to measure the frequency domain similarity. Lastly, Wirz, Roggen, and Tröster [81] proposed an innovative method called the fingerprinting algorithm to measure the similarity based on the naive Bayes algorithm. Their results showed that the estimation of the quantitative distance in meters between a device DA and another device DB reaches an accuracy of approximately 80% when using ambient sound as a relevant source for obtaining proximity information.
Recently, a Deep Neural Network (DNN)-based approach has been proposed over the parametric approach (TDOA and IID) for sound source localization. Adavanne et al. [82] employed a convolutional RNN for sound localization of multiple overlapping sound events in three-dimensional (3D) space. The approach network takes a sequence of consecutive spectrogram timeframes as a multichannel input and maps it to a multi-output regression. The study shows that the proposed method is generic and applicable to any array structures and is robust to unseen DOA values, reverberation, and low SNR scenarios.

4.1.5. Sound Abnormality Detection

Measuring sound abnormality is another challenging issue, given that preparing all labeled sound data related to hazardous situations is unrealistic. If a surveillance system is trained using the data of specific sounds, such as explosions or gunshots, it cannot be applied to detect other auditory events. A method for detecting unknown abnormal sound events is required to develop an automated auditory surveillance system that is useful in more general cases. To address this issue, Ref. [71] developed a method that models the abnormality of sounds without using any samples of labeled sounds. The technique can detect those sounds that rarely occur in a normal situation. It first processes sound in the usual situation and trains a statistical model of the normal sounds. After training the model, the system continues to process sounds and calculates the likelihood of the sound. If the likelihood value goes beyond a predefined threshold, that sound is considered abnormal. Lu et al. [83] used another approach to detect abnormal sounds using the case-based identification algorithm. In this method, it is necessary to first convert the sound data into feature representation vectors and then to apply an establishment distribution of a supervised learning model. This supervised learning requires a small training dataset of sample elements of abnormality.
In the construction industry, sound equipment abnormality will affect the model performance since mobile equipment sound varies between different types of equipment. This could be because the acoustic characteristic of one sound is more difficult to train than that of another. Other equipment characteristics that may affect sound abnormality measurement are the models, brands, age, and maintenance programs. Obtaining different model metadata for the audio dataset would help us understand how they affect the detection capability. Collecting such data requires a significant effort in terms of time, cost, and human resources.

4.2. Audio-Based Surveillance in Construction

4.2.1. Feasibility of Implementing Audio-Based Hazard Detection in Construction

The potential auditory indicators of hazards in construction are provided in Table 8. As shown, one of the most important types of sounds is the sound source of moving heavy equipment or machinery, which can cause collision hazards [30,33,36]. Furthermore, the detection of screams, shouts, or cry sounds is also supportive for acoustic surveillance and monitoring of negative situations, since human emotions are somehow delivered in sound events [17,18,19,20,23,24,25,26,27,28,61,62,68,76,77]. Other sounds released by people can also help detect abnormal situations. Detecting ground ambient sound (i.e., a group of people) allows workers to be aware of violent events, natural disasters, riots, or chaotic acts in crowds [26]. Some types of safety-critical sounds which lead to dangerous situations, such as alarms of fire, earthquakes, explosions, and gunshots, could improve the situational awareness of hazards [23,25,76,77]. For example, the detection of an explosion allows workers to stay away from the hazardous source [16] and gunshot detection allows workers to stay away from a gun attack [16,17,19,20,25,61,62].
Other than the classification of sound, sound source location [78,79,80,81], the direction of a moving sound source [78,79,80,81], and sound abnormality [71] are useful for detecting hazardous situations in the construction field. For example, the information on the location and the direction of moving construction equipment from the engine sound could help alert if a heavy construction vehicle is in proximity to a worker. Additionally, suppose the construction equipment breaks down by falling, collapsing, colliding, or by failures in the engine, the information of the abnormality in the sound could give a cue to carry out proper assistance, e.g., the noise produced when excessive vibration of equipment occurs that is not usually expected to vibrate.
Existing frameworks for detecting safety cues solely rely upon a single type of information, for example, high-frequency filtering [84], sound identify classification [85,86], or direction of arrival of sound [87]. Using the information individually is insufficient for evaluating hazardousness in construction that requires a simultaneous consideration of many factors, including the size of the equipment in contact with, the breaking speed of a machine, the average reaction time of a worker, and the speed of the worker. Table 9 presents a list of hazardous situations in construction sites along with required auditory indicators. Specifically, hazardous situations include heavy equipment/machines being at an unsafe distance from workers when detected using the sound of moving equipment. The detection of equipment approaching or operating in an abnormal condition will be alerted if the direction of the moving sound source is toward the worker or if the sound source is abnormal. Other situations, such as someone crying/shouting, a crowd approaching, alert alarms, an explosion, or a gunshot, can indicate a hazardous situation. Defining hazardous situations that require quick and effective responses from construction workers is the priority of automated auditory surveillance in the construction field and could contribute to construction workers’ safety.

4.2.2. State-of-the-Art Research in Auditory Signal Processing for Construction

There have been emerging studies on audio-based activity detection for improving construction management and productivity due to its advantages in terms of cost and applicability [88]. Other efforts have aimed to develop new methods for assisting workers in hearing critical sounds, which is a crucial need given the typical heterogeneity of sounds generated from diverse construction work activities, including static equipment and hand tools [89,90]. The examination of various OSHA accident reports by Hinze et al., 2011 [36] revealed that the heterogeneous nature of concurrent construction sounds (e.g., equipment sounds and alarms) indeed decreased workers’ safety awareness, since alarm signals may be drowned out or not audible enough for workers. They also reported that there were cases where multiple alarm signals issued warnings at the same time, influencing workers’ judgment and making the alert signals less effective or ineffective. However, due to the technical challenges of processing complex soundscapes, this problem has gained little attention from the academic community in the past decade.
Only a few papers were found related to this technology in construction management. In general, previous studies were focused on the tracking of activities of construction equipment, identification of working and operation activities, proximity detection and alert systems, and embedded sensory systems. Of those areas, a majority number of studies aimed at monitoring the activities of heavy construction equipment to reduce operating costs [30,72,74,91]. This is probably because a large portion of the expenses in a construction project is allocated towards the operating costs of heavy equipment. For example, Ref. [92] applied the Hidden Markov Model (HMM) and a frequency-domain technique on spectrogram data to accurately classify types of construction sounds and to identify patterns from each type of construction task. The classified sound signals’ strength and location are visualized with a Building Information Modeling (BIM) platform. The acoustic signals from construction activities were used to calculate working periods to allow field managers to track work progress and productivity and to provide a means to efficiently enhance project schedule management. Ref. [93] worked on a sound monitoring system for the prevention of underground pipeline damage. To develop a dataset similar to what is found on the construction site, they collected working equipment sound data of typical construction threats, including excavators, hammers, road cutters, and electric hammers. They also collected the background noise of a typical construction environment, such as pedestrians, traffic, and wind sound. Two random forest-based classifiers were developed to detect suspicious sounds and to help prevent pipeline damages caused by construction activities. The endpoint was to create an alarm system that uploads a report if the duration of a construction threat sound exceeds the threshold value specified.
Another study proposed a hybrid system for recognizing multiple construction equipment activities [30]. The study trained a machinery task classifier on integrated data of both audio and kinematics using Support Vector Machines (SVM). The proposed system results indicate that a hybrid system is capable of providing up to 20% more accurate results compared with cases using individual sources of data such as images [6,7,8], sensors [9,10], and audio [75,91,94]. The system allows the construction managers to monitor and track productivity, equipment downtime/idle time detection, equipment cycle time estimation, and equipment fuel use control. Wei et al. [95] also developed a noise hazard prediction method that combines a wearable audio sensor with Building Information Modeling (BIM) data to predict and visualize spatial noise distribution on BIM models.
There has also been a widespread usage of machine learning algorithms to train sound data for construction activity monitoring. Ref. [75] implemented a supervised machine learning-based sound identification approach to enhance the monitoring of construction site activities. Ref. [89] also developed a risk assessment framework using a machine learning algorithm. This method used the activity classification information from auditory signals to estimate safety risks based on the occupational injury and illness information contained in historical construction accident data. All these audio-based frameworks for detecting construction activities are effective, especially for night-time tasks, since activities on construction sites can be detected regardless of visibility levels that are not suitable for image-based approaches.
Another line of effort on auditory surveillance in construction is focused on the identification of construction collision hazards. The advanced computational techniques in auditory signal processing for collision hazard detection in construction are motivated by strong acoustic emissions from equipment operation, since construction equipment often produces unique sound patterns while performing certain activities (e.g., moving vs. idling) [74], which can be used as an indicator of safety-critical cues or warning signals. For example, the research done by Lee and Yang [84] used a high-frequency sound (18 kHz, inaudible to construction workers) to analyze the doppler effect change caused by a single subject’s movements to prevent struck-by hazards. They installed a speaker on equipment that plays a predefined high-frequency sound. A smartphone carried by an on-foot worker was used to capture the sound produced by the speaker. The input signal was processed to extract the position of the equipment relative to the on-foot worker. The proposed technology was able to classify the movement direction and speed with 97.4% accuracy. Although the study proves its potential for detecting collision threats from equipment, the study still has some limitations. Firstly, they only tested a struck-by hazard situation involving a single type of moving construction equipment. Since the construction site is a complex environment in which multiple pieces of equipment work simultaneously leading to signal overlap, the mixture of similar sounds would prevent the recognition of movement of individual pieces of equipment. Another limitation was that the sound source was a speaker attached to the equipment, not the sound produced by the mobile equipment. This means the deployment will require expensive installation of sound speakers on every piece of equipment present on the job site.
Recent studies by Refs. [85,86] developed sound classification models that can distinguish between mobile equipment and stationary equipment to support collision hazard detection. These studies collected and synthesized the sound of construction equipment at different signal-to-noise ratios and used the dataset to develop a machine learning model using a CNN for automated detection of mobile equipment occurrence. The efficacy of this model was tested on a real construction site and the result accuracy of the model was 99% in detecting sounds related to collision hazards when the signals were not buried in background noises. Compared with earlier efforts, Refs. [85,86] offer superior advancement, as their models are able to deal with complex soundscapes with overlapping sound sources, including mobile and stationary equipment and background noise (e.g., workers communicating, materials’ movement, and street noises). Another study that utilized equipment sounds for collision hazard assessment was performed by Ref. [87], which aimed at localizing the sound sources using the Direction of Arrival (DOA) signal processing techniques. The determination of sound location can supplement the sound classifiers developed by Refs. [85,86] to enable a more comprehensive assessment of the hazardousness-based situations. This information is vital for construction workers and safety engineers to precisely reduce false alarms by only notifying workers if they are in a danger zone based on the distance calculated using the direction of the hazard. For example, a mobile piece of equipment moving further away from an on-foot worker does not impose a hazard.
Quantitative benchmarks between existing frameworks for preventing struck-by hazards are greatly difficult as they all were tested with nonidentical testing conditions (e.g., software, hardware, job-site characteristics, and assumptions). Therefore, some performance metrics, such as recall, cost, computational power, and data usage, can still be used for a fair comparison. In terms of performance comparison, they all yielded a competitive recall of 99% [85], 98% [96], 99% [85,86], and 100% [97]. Cost comparison was another metric to measure past success in preventing struck-by hazards with mobile equipment. Audio-based collision hazard detection by Ref. [85] requires little financial investment as the model can be quickly deployed on workers’ smartphones, while a relatively high deployment cost is needed for many sensor devices [97] or high-quality cameras [96,98].

5. Discussions on Future Research

Research on automated auditory surveillance has shown a rapid development momentum in the past decade. AI-based sound processing techniques show great potential for providing construction workers with powerful cognitive assistance to improve their auditory situational awareness of construction hazards. The most promising sound sensing frameworks, particularly for sound-based detection of construction hazards including those sound classifications, are those developed by Refs. [30,72,74,85,86,91]. Still, research on leveraging AI innovation to improve construction workers’ auditory situational awareness in construction safety involves various technological and practical gaps that require further research.
First, construction sites are a dynamic environment where many activities are performed concurrently, resulting in important signals being buried in complicated occurrences of overlapping soundscapes, including unwanted noises such as passing vehicles on highway projects. However, most previous studies used single sound sources in their training data [74,75,91,99]. It would be helpful to implement advanced sound separation models to separate noises from important signals prior to feeding them to the classification models. Another approach is to include overlapping sound samples in the training set. Ref. [85] was among very few efforts adopting this approach by training AI models on a large set of acoustic collision hazard signals mixed with noises at various sound-to-noise ratios. The study considered only stationary equipment sounds as background noise. Future research is needed to expand training sound mixture datasets to cover other background noises such as winds, traffic noise, and unwanted backup alarms. Additionally, the performance of sound classification models is required to be tested in diverse types of construction sites in terms of location (urban, suburban, and rural) and work type (highway, building, and offshore project). Construction sites are highly dynamic; machine learning models built on old data may be inapplicable for supporting the detection of new anomalies. Thus, continuously updating the model with new training data is extremely important for the success of these technologies.
Another crucial gap in the literature is the lack of frameworks for detecting emergency sounds made by human subjects. Our review of recently published OSHA accident narrative revealed many types of audio cues requiring field personnel’s attention. Examples of verbal safety-critical signals mentioned in recent accident reports are “yelling”, “help”, “watch out”, “stop”, and “ahh”. The quote below is one great example illustrating the importance of preserving the audibility of such verbal signals.
At 2:00 p.m. on 13 February 2016, Employee #1 and Employee #2 were in an excavation installing sanitary sewer lines in a 13.5 foot trench near a gas line. A foreman saw that the excavation was caving-in and yelled at Employee #1 and Employee #2 to get out of the trench. Employee #2 got out of the trench without injury, but Employee #1 was not able to make it out and was killed at the scene
These sounds can be classified into pre-accident (i.e., “watch out”) and post-accident warning alerts (i.e., yelling for help). Both kinds of signals are crucial to enable co-workers to be aware of their emergency situations. Due to the exceeding noise level at construction job sites, it is extremely difficult for workers to recognize such signals. Unfortunately, there are no technologies available that are capable of improving workers’ awareness of such urgent sounds in noisy construction sites. Future research should explore the feasibility of AI in recognizing urgent speeches buried in unimportant noises, including normal speeches.
Sound classification alone is insufficient, given that the meaning of sounds in construction often depends on the context. The same type of sound may mean something different to different workers. For example, reverse alarms are an unwanted noise to those standing outside of the equipment backup trajectory. However, missing safety cues even in a short time can lead to deadly consequences. Thus, the preciseness in sound labeling (wanted vs. unwanted sound) is extremely crucial for construction safety. Unfortunately, none of the previous studies address this complexity of sounds in construction. Researchers would need to develop robust risk assessment models considering the job-site layout and workers’ spatial location and integrate them into sound classification models. Such integration will help reduce false alarms and enhance trust in the automation. The development of such AI models requires significant efforts to collect sound data representing a diverse range of real-life situations with regard to equipment types, equipment maintenance condition, and equipment movement trajectories relative to on-foot workers. It may be unrealistic to obtain field sound data for every type of construction project and activity, thus generating synthetic sound data using simulation platforms such as Pyroomacoustics may be necessary to supplement the data needs for deep learning models.
Moreover, future efforts should be made to enable successful deployment of sound-processing AI models on wearable devices. There have been many wearable technologies developed for construction safety. However, most existing technologies are to support hazard sensing based on data that are different from audio signals. For example, Cho and Park [100] developed a tactile-based communication system to support equipment proximity detection. A similar tactile-based wearable device was recently developed and tested by Sakhakarmi et al., [101]. These experiments involved several shortcomings that need further validation. For example, human subjects were disallowed to perform their regular work activities during the test. This setting apparently did not reflect the nature of construction sites. Since workers are expected to continue working without distraction, future wearable devices must be comfortable and easy to use while performing other tasks. In addition, previous wearable safety devices in construction were built using a predefined type of feedback such as either text messages or visual alerts on a built-in screen. No studies have been found that investigate effective feedback mechanisms considering the complex nature of construction workflows. Research is needed to understand approaches suitable for effectively conveying the hazard detection outputs. Lessons learned from other fields, including hearing aids for hard-of-hearing people, would be a great point of departure. Refs. [102,103] are two examples of most recent work that compared different feedback methods, e.g., haptics, texts, and visuals, on various wearable devices including smart watches and brackets.
To ensure a high adoption of hearing aid devices for construction workers, future research also needs to examine the practical challenges and potential strategies to deal with technology adoption resistance. Factors such as those in Table 10 need to be considered to ensure successful adoption. Specifically, to increase the percentage of using wearable augmented hearing devices, effective interventions based on the protection motivation theory that aim to improve perceptions of severity, vulnerability, perceived cost, self-efficacy, and perceived response efficacy should be considered [104]. Once an augmented hearing model is integrated with hearing protection devices, the development of such technology should also take into account other factors, as identified by Ref. [105]. One of the factors is that the use of wearable devices among construction workers may be affected by social influence. Interventions considering personal and environmental factors using ecological and ergonomic models are required to promote adoption among construction workers. In addition, hearing-related problems such as noise-induced hearing loss or left and right ear differences need to be addressed [106]. Furthermore, the perception of noise and the perceived value of hearing assistance are essential factors that affect the use of augmented hearing devices. Research shows that most construction workers still refrain from the frequent use of hearing protection devices because they do not see many benefits gained from these devices. This means that the demonstration of the benefit of the new technology would be helpful to enhance their understanding of how it can provide valuable information, such as safety acoustic cues.
Lastly, human–computer interaction is worth investigating for future hearing assistance technologies for construction workers. Hazard detection results made by an audio processing technology would be useless if the involved parties did not trust it. Since construction sites, especially roadway work zones, are considerably harsh working environments due to complex noises generated from traffic and construction activities, errors such as false alarms or missing true alerts would be unavoidable even with the most advanced deep learning networks. Thus, human involved in hazard assessment is crucial to provide real-time feedback to enhance the decision makings. For example, advanced sound filtering that can suppress unwanted background noises to an appropriate signal/noise ratio before passing the environmental sound to the workers through hearing protection equipment would allow them to process the sound themselves, resulting in human computing teaming. To facilitate such collaboration, human-AI systems for audio signal processing should be appropriately designed considering workers’ cognition capabilities and other barriers regarding limited vision and hearing while performing construction tasks, often in the nighttime.

6. Conclusions

Sound-based hazard detection for construction safety is a new and emerging area. This paper enhances the body knowledge by analyzing the current auditory signal processing using AI and its implication for construction safety, summarizing research trends and identifying promising future research directions. Previous research on sound sensing in construction has focused on the classification of equipment activities. There are still inadequate research efforts in areas such as sound-based measurement of hazardousness, detection of abnormal situations, and construction workers’ recognition of hazards at the construction site, shown by a steady trend of growing publications. AI-based signal processing for hazard assessment has great potential for improving the detection of hazards on the construction site. Deep learning methods such as CNN or AE using MFCC features have been shown to achieve impressive accuracy.
Further research that employs AI to enhance the worker’s auditory situational awareness still faces various technical and non-technical challenges. Technical challenges include hazard assessment and real-time data processing detection of abnormal situations without trained data. The current body of knowledge merits further research to precisely assess the hazardousness degree as well as the location of hazard events. Additionally, more studies, involving the survey of future users, are needed to understand the construction workers’ recognition and localization of hazards at the construction site, unsafe behaviors to auditory events, and important ergonomic and ecological factors to their auditory situational awareness. These new understandings will enable a successful deployment of a human-AI system for the auditory sensing of construction hazards. Lastly, more studies on the important factors of the implementation of AI-based hearing protection devices and sound-sensing technology are required to justify the value of AI-based sound surveillance towards construction safety and to promote the user’s adoption.

Author Contributions

Conceptualization, K.D. and T.L.; investigation, K.D. (reviewed 78 articles), K.E. (reviewed 7 articles) and T.L.; methodology, K.D.; visualization, K.D.; supervision, T.L.; writing—original draft, K.D. (Section 1, Section 2, Section 3, Section 4.1, Section 4.2.1 and Section 6) and K.E. (Section 4.2.2 and Section 5); writing—review and editing, K.D., K.E., T.L. and C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by CPWR, The Center for Construction Research and Training through cooperative Agreement No. U60-OH009762 from the National Institute of Occupational Safety and Health (NIOSH). Its contents are solely the responsibility of the authors and do not necessarily represent the official views of CPWR or NIOSH.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

This publication was made possible by CPWR, The Center for Construction Research and Training through cooperative agreement number U60- OH009762 from the National Institute of Occupational Safety and Health (NIOSH). Its contents are solely the responsibility of the authors and do not necessarily represent the official views of CPWR or NIOSH.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Deshaies, P.; Martin, R.; Belzile, D.; Fortier, P.; Laroche, C.; Leroux, T.; Nélisse, H.; Girard, S.-A.; Arcand, R.; Poulin, M.; et al. Noise as an explanatory factor in work-related fatality reports. Noise Health 2015, 17, 294–299. [Google Scholar] [CrossRef] [PubMed]
  2. Fang, D.; Zhao, C.; Zhang, M. A Cognitive Model of Construction Workers’ Unsafe Behaviors. J. Constr. Eng. Manag. 2016, 142, 4016039. [Google Scholar] [CrossRef]
  3. Morata, T.C.; Themann, C.L.; Randolph, R.; Verbsky, B.L.; Byrne, D.C.; Reeves, E.R. Working in Noise with a Hearing Loss: Perceptions from Workers, Supervisors, and Hearing Conservation Program Managers. Ear Hear. 2005, 26, 529–545. [Google Scholar] [CrossRef] [PubMed]
  4. Awolusi, I.; Marks, E.; Hallowell, M. Wearable technology for personalized construction safety monitoring and trending: Review of applicable devices. Autom. Constr. 2018, 85, 96–106. [Google Scholar] [CrossRef]
  5. Awolusi, I.; Nnaji, C.; Marks, E.; Hallowell, M. Enhancing Construction Safety Monitoring through the Application of Internet of Things and Wearable Sensing Devices: A Review. In Proceedings of the Computing in Civil Engineering, Atlanta, Georgia, USA, 17–19 June 2019; American Society of Civil Engineers: Reston, VA, USA; pp. 530–538. [Google Scholar]
  6. Golparvar-Fard, M.; Heydarian, A.; Niebles, J.C. Vision-based action recognition of earthmoving equipment using spatio-temporal features and support vector machine classifiers. Adv. Eng. Inform. 2013, 27, 652–663. [Google Scholar] [CrossRef]
  7. Gong, J.; Caldas, C.H. Computer Vision-Based Video Interpretation Model for Automated Productivity Analysis of Construction Operations. J. Comput. Civ. Eng. 2010, 24, 252–263. [Google Scholar] [CrossRef]
  8. Gong, J.; Caldas, C.H.; Gordon, C. Learning and classifying actions of construction workers and equipment using Bag-of-Video-Feature-Words and Bayesian network models. Adv. Eng. Inform. 2011, 25, 771–782. [Google Scholar] [CrossRef]
  9. Akhavian, R.; Behzadan, A.H. Construction equipment activity recognition for simulation input modeling using mobile sensors and machine learning classifiers. Adv. Eng. Inform. 2015, 29, 867–877. [Google Scholar] [CrossRef]
  10. Ahn, C.R.; Lee, S.; Peña-Mora, F. Application of Low-Cost Accelerometers for Measuring the Operational Efficiency of a Construction Equipment Fleet. J. Comput. Civ. Eng. 2015, 29, 4014042. [Google Scholar] [CrossRef]
  11. Salamon, J.; Jacoby, C.; Bello, J.P. A Dataset and Taxonomy for Urban Sound Research. In Proceedings of the 2014 ACM Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014; Association for Computing Machinery: New York, NY, USA, 2014; pp. 1041–1044. [Google Scholar]
  12. Salamon, J.; Bello, J.P. Deep Convolutional Neural Networks and Data Augmentation for Environmental Sound Classification. IEEE Signal Process. Lett. 2017, 24, 279–283. [Google Scholar] [CrossRef]
  13. Cramer, J.; Wu, H.-H.; Salamon, J.; Bello, J.P. Look, Listen, and Learn More: Design Choices for Deep Audio Embeddings. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 3852–3856. [Google Scholar]
  14. Gemmeke, J.F.; Ellis, D.P.W.; Freedman, D.; Jansen, A.; Lawrence, W.; Moore, R.C.; Plakal, M.; Ritter, M. Audio Set: An Ontology and Human-Labeled Dataset for Audio Events. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 776–780. [Google Scholar]
  15. Piczak, K.J. Environmental Sound Classification with Convolutional Neural Networks. In Proceedings of the 2015 IEEE 25th International Workshop on Machine Learning for Signal Processing (MLSP), Boston, MA, USA, 17–20 September 2015; pp. 1–6. [Google Scholar]
  16. Clavel, C.; Ehrette, T.; Richard, G. Events Detection for an Audio-Based Surveillance System. In Proceedings of the 2005 IEEE International Conference on Multimedia and Expo, Amsterdam, The Netherlands, 5–6 July 2005; pp. 1306–1309. [Google Scholar]
  17. Valenzise, G.; Gerosa, L.; Tagliasacchi, M.; Antonacci, F.; Sarti, A. Scream and Gunshot Detection and Localization for Audio-Surveillance Systems. In Proceedings of the 2007 IEEE Conference on Advanced Video and Signal Based Surveillance, London, UK, 5–7 September 2007; pp. 21–26. [Google Scholar]
  18. Vasilescu, I.; Devillers, L.; Clavel, C.; Ehrette, T. Fiction Database for Emotion Detection in Abnormal Situations. In Proceedings of the 31st International Symposium on Computer Architecture (ISCA 2004), Munich, Germany, 19–23 June 2004; pp. 2277–2280. [Google Scholar]
  19. Ntalampiras, S.; Potamitis, I.; Fakotakis, N. An Adaptive Framework for Acoustic Monitoring of Potential Hazards. EURASIP J. Audio Speech Music Process 2009, 2009, 594103. [Google Scholar] [CrossRef]
  20. Choi, W.; Rho, J.; Han, D.K.; Ko, H. Selective Background Adaptation Based Abnormal Acoustic Event Recognition for Audio Surveillance. In Proceedings of the 2012 IEEE Ninth International Conference on Advanced Video and Signal-Based Surveillance, Beijing, China, 18–21 September 2012; pp. 118–123. [Google Scholar]
  21. Carletti, V.; Foggia, P.; Percannella, G.; Saggese, A.; Strisciuglio, N.; Vento, M. Audio Surveillance Using a Bag of Aural Words Classifier. In Proceedings of the 2013 10th IEEE International Conference on Advanced Video and Signal Based Surveillance, Krakow, Poland, 27–30 August 2013; pp. 81–86. [Google Scholar]
  22. Chan, C.-F.; Yu, E.W.M. An Abnormal Sound Detection and Classification System for Surveillance Applications—IEEE Conference Publication. In Proceedings of the 2010 18th European Signal Processing Conference, Aalborg, Denmark, 23–27 August 2010; pp. 1851–1855. [Google Scholar]
  23. Lee, Y.; Han, D.; Ko, H. Acoustic Signal Based Abnormal Event Detection in Indoor Environment Using Multiclass Adaboost. IEEE Trans. Consum. Electron. 2013, 59, 615–622. [Google Scholar] [CrossRef]
  24. Rouas, J.-L.; Louradour, J.; Ambellouis, S. Audio Events Detection in Public Transport Vehicle. In Proceedings of the 2006 IEEE Intelligent Transportation Systems Conference, Toronto, ON, Canada, 17–20 September 2006; pp. 733–738. [Google Scholar]
  25. Radhakrishnan, R.; Divakaran, A. Systematic Acquisition of Audio Classes for Elevator Surveillance. In Proceedings of the Image and Video Communications and Processing 2005, San Jose, CA, USA, 12–20 January 2005; Said, A., Apostolopoulos, J.G., Eds.; SPIE: Bellingham, WA, USA, 2005; Volume 5685, pp. 64–71. [Google Scholar]
  26. Clavel, C.; Vasilescu, I.; Devillers, L.; Richard, G.; Ehrette, T. Fear-Type Emotion Recognition for Future Audio-Based Surveillance Systems. Speech Commun. 2008, 50, 487–503. [Google Scholar] [CrossRef]
  27. Vacher, M.; Istrate, D.; Besacier, L.; Serignat, J.-F.; Castelli, E. Sound Detection and Classification for Medical Telesurvey. In Proceedings of the 2nd Conference on Biomedical Engineering, Innsbruck, Austria, 16–18 February 2004. [Google Scholar]
  28. Adnan, S.M.; Irtaza, A.; Aziz, S.; Ullah, M.O.; Javed, A.; Mahmood, M.T. Fall Detection through Acoustic Local Ternary Patterns. Appl. Acoust. 2018, 140, 296–300. [Google Scholar] [CrossRef]
  29. Härmä, A.; Mckinney, M.F.; Skowronek, J. Automatic Surveillance of the Acoustic Activity in Our Living Environment. In Proceedings of the IEEE International Conference on Multimedia and Expo, ICME 2005, Amsterdam, The Netherlands, 6–8 July 2005; Volume 2005, pp. 634–637. [Google Scholar]
  30. Sherafat; Rashidi; Lee; Ahn A Hybrid Kinematic-Acoustic System for Automated Activity Detection of Construction Equipment. Sensors 2019, 19, 4286. [CrossRef]
  31. Glowacz, A. Acoustic Based Fault Diagnosis of Three-Phase Induction Motor. Appl. Acoust. 2018, 137, 82–89. [Google Scholar] [CrossRef]
  32. Koizumi, Y.; Saito, S.; Uematsu, H.; Kawachi, Y.; Harada, N. Unsupervised Detection of Anomalous Sound Based on Deep Learning and the Neyman–Pearson Lemma. IEEE/ACM Trans. Audio Speech Lang. Process 2019, 27, 212–224. [Google Scholar] [CrossRef]
  33. Marks, E.D.; Teizer, J. Method for Testing Proximity Detection and Alert Technology for Safe Construction Equipment Operation. Constr. Manag. Econ. 2013, 31, 636–646. [Google Scholar] [CrossRef]
  34. Sherafat, B.; Ahn, C.R.; Akhavian, R.; Behzadan, A.H.; Golparvar-Fard, M.; Kim, H.; Lee, Y.-C.; Rashidi, A.; Azar, E.R. Automated Methods for Activity Recognition of Construction Workers and Equipment: State-of-the-Art Review. J. Constr. Eng. Manag. 2020, 146, 3120002. [Google Scholar] [CrossRef]
  35. Perlman, A.; Sacks, R.; Barak, R. Hazard recognition and risk perception in construction. Saf. Sci. 2014, 64, 22–31. [Google Scholar] [CrossRef]
  36. Hinze, J.W.; Teizer, J. Visibility-related fatalities related to construction equipment. Saf. Sci. 2011, 49, 709–718. [Google Scholar] [CrossRef]
  37. Schultheis, U. Human Factors: Blind Spots in Heavy Construction Machinery. Available online: https://www.jdsupra.com (accessed on 23 November 2022).
  38. Brungart, D.S.; Cohen, J.; Cord, M.; Zion, D.; Kalluri, S. Assessment of auditory spatial awareness in complex listening environments. J. Acoust. Soc. Am. 2014, 136, 1808–1820. [Google Scholar] [CrossRef] [PubMed]
  39. Vinnik, E.; Itskov, P.M.; Balaban, E. Individual Differences in Sound-in-Noise Perception Are Related to the Strength of Short-Latency Neural Responses to Noise. PLoS ONE 2011, 6, e17266. [Google Scholar] [CrossRef] [PubMed]
  40. Engdahl, B.; Tambs, K. Occupation and the risk of hearing impairment—results from the Nord-Trøndelag study on hearing loss. Scand. J. Work. Environ. Health 2009, 36, 250–257. [Google Scholar] [CrossRef] [PubMed]
  41. Rigters, S.C.; Metselaar, M.; Wieringa, M.H.; De Jong, R.J.B.; Hofman, A.; Goedegebure, A. Contributing Determinants to Hearing Loss in Elderly Men and Women: Results from the Population-Based Rotterdam Study. Audiol. Neurotol. 2016, 21, 10–15. [Google Scholar] [CrossRef]
  42. Kwon, N.; Park, M.; Lee, H.-S.; Ahn, J.; Shin, M. Construction Noise Management Using Active Noise Control Techniques. J. Constr. Eng. Manag. 2016, 142, 4016014. [Google Scholar] [CrossRef]
  43. Gao, X.; Pishdad-Bozorgi, P. BIM-Enabled Facilities Operation and Maintenance: A Review. Adv. Eng. Inform. 2019, 39, 227–247. [Google Scholar] [CrossRef]
  44. Han, W.; Coutinho, E.; Ruan, H.; Li, H.; Schuller, B.; Yu, X.; Zhu, X. Semi-Supervised Active Learning for Sound Classification in Hybrid Learning Environments. PLoS ONE 2016, 11, e0162075. [Google Scholar] [CrossRef]
  45. Ma, L.; Milner, B.; Smith, D. Acoustic Environment Classification. ACM Trans. Speech Lang. Process 2006, 3, 1–22. [Google Scholar] [CrossRef]
  46. Logan, B.; Salomon, A. A Music Similarity Function Based on Signal Analysis. In Proceedings of the IEEE International Conference on Multimedia and Expo, Tokyo, Japan, 22–25 August 2001; pp. 745–748. [Google Scholar]
  47. Cotton, C.V.; Ellis, D.P.W. Spectral vs. Spectro-Temporal Features for Acoustic Event Detection. In Proceedings of the 2011 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), New York, NY, USA, 16–19 October 2011; pp. 69–72. [Google Scholar]
  48. Stowell, D.; Plumbley, M. Automatic Large-Scale Classification of Bird Sounds Is Strongly Improved by Unsupervised Feature Learning. PeerJ 2014, 2, e488. [Google Scholar] [CrossRef]
  49. Salamon, J.; Bello, J.P. Unsupervised Feature Learning for Urban Sound Classification. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane, Qld, Australia, 19–24 April 2015; pp. 171–175. [Google Scholar]
  50. McFee, B.; Humphrey, E.J.; Bello, J.P. A Software Framework for Musical Data Augmentation. In Proceedings of the ISMIR, Malaga, Spain, 26–30 October 2015; pp. 248–254. [Google Scholar]
  51. Chaudhuri, S.; Raj, B. Unsupervised Hierarchical Structure Induction for Deeper Semantic Analysis of Audio. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 833–837. [Google Scholar]
  52. Salamon, J.; Bello, J.P. Feature Learning with Deep Scattering for Urban Sound Analysis. In Proceedings of the 2015 23rd European Signal Processing Conference (EUSIPCO), Nice, France, 31 August–4 September 2015; pp. 724–728. [Google Scholar]
  53. Salamon, J.; Bello, J.P.; Farnsworth, A.; Robbins, M.; Keen, S.; Klinck, H.; Kelling, S. Towards the Automatic Classification of Avian Flight Calls for Bioacoustic Monitoring. PLoS ONE 2016, 11, e0166866. [Google Scholar] [CrossRef]
  54. Bello, J. Grouping Recorded Music by Structural Similarity. In Proceedings of the 10th International Society for Music Information Retrieval Conference, ISMIR 2009, Kobe, Japan, 26-30 October 2009; Kobe International Conference Center: Kobe, Japan; pp. 531–536. [Google Scholar]
  55. Salamon, J.; Bello, J.P.; Farnsworth, A.; Kelling, S. Fusing Shallow and Deep Learning for Bioacoustic Bird Species Classification. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, CA, USA, 5–9 March 2017; pp. 141–145. [Google Scholar]
  56. Bello, J.P.; Mydlarz, C.; Salamon, J. Sound Analysis in Smart Cities. In Computational Analysis of Sound Scenes and Events; Springer International Publishing: Cham, Switzerland, 2018; pp. 373–397. [Google Scholar]
  57. Alsouda, Y.; Pllana, S.; Kurti, A. A Machine Learning Driven IoT Solution for Noise Classification in Smart Cities. In Proceedings of the 21st Euromicro Conference on Digital System Design (DSD 2018), Workshop on Machine Learning Driven Technologies and Architectures for Intelligent Internet of Things (ML-IoT), Prague, Czech Republic, 29–31 August 2018. [Google Scholar]
  58. Asgari, M.; Shafran, I.; Bayestehtashk, A. Inferring Social Contexts from Audio Recordings Using Deep Neural Networks. In Proceedings of the 2014 IEEE International Workshop on Machine Learning for Signal Processing (MLSP), Reims, France, 21–24 September 2014; pp. 1–6. [Google Scholar]
  59. Ravanelli, M.; Elizalde, B.; Ni, K.; Friedland, G. Audio Concept Classification with Hierarchical Deep Neural Networks. In Proceedings of the 22nd European Signal Processing Conference, Lisbon, Portugal, 1–5 September 2014. [Google Scholar]
  60. Gencoglu, O.; Virtanen, T.; Huttunen, H. Recognition of Acoustic Events Using Deep Neural Networks. In Proceedings of the 22nd European Signal Processing Conference, Lisbon, Portugal, 1–5 September 2014; pp. 506–510. [Google Scholar]
  61. Janjua, Z.H.; Vecchio, M.; Antonini, M.; Antonelli, F. IRESE: An Intelligent Rare-Event Detection System Using Unsupervised Learning on the IoT Edge. Eng. Appl. Artif. Intell. 2019, 84, 41–50. [Google Scholar] [CrossRef]
  62. Salekin, A.; Ghaffarzadegan, S.; Feng, Z.; Stankovic, J. A Real-Time Audio Monitoring Framework with Limited Data for Constrained Devices. In Proceedings of the 2019 15th International Conference on Distributed Computing in Sensor Systems (DCOSS), Santorini Island, Greece, 29–31 May 2019; pp. 98–105. [Google Scholar]
  63. Eronen, A.J.; Peltonen, V.T.; Tuomi, J.T.; Klapuri, A.P.; Fagerlund, S.; Sorsa, T.; Lorho, G.; Huopaniemi, J. Audio-Based Context Recognition. IEEE Trans. Audio Speech Lang. Process 2006, 14, 321–329. [Google Scholar] [CrossRef]
  64. Kim, B.; Pardo, B. A Human-in-the-Loop System for Sound Event Detection and Annotation. ACM Trans. Interact. Intell. Syst. 2018, 8, 1–23. [Google Scholar] [CrossRef]
  65. Qian, K.; Zhang, Z.; Baird, A.; Schuller, B. Active learning for bird sound classification via a kernel-based extreme learning machine. J. Acoust. Soc. Am. 2017, 142, 1796–1804. [Google Scholar] [CrossRef]
  66. Qian, K.; Zhang, Z.; Baird, A.; Schuller, B. Active Learning for Bird Sounds Classification. Acta Acust. United Acust. 2017, 103, 361–364. [Google Scholar] [CrossRef]
  67. Shuyang, Z.; Heittola, T.; Virtanen, T. Active Learning for Sound Event Classification by Clustering Unlabeled Data. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 751–755. [Google Scholar]
  68. Atrey, P.K.; Maddage, N.C.; Kankanhalli, M.S. Audio Based Event Detection for Multimedia Surveillance. In Proceedings of the 2006 IEEE International Conference on Acoustics Speech and Signal Processing, Toulouse, France, 14–19 May 2006. [Google Scholar]
  69. Sephus, N.H.; Lanterman, A.D.; Anderson, D.V. Modulation Spectral Features: In Pursuit of Invariant Representations of Music with Application to Unsupervised Source Identification. J. New Music Res. 2014, 44, 58–70. [Google Scholar] [CrossRef]
  70. Wang, Y.; Mendez, A.E.M.; Cartwright, M.; Bello, J.P. Active Learning for Efficient Audio Annotation and Classification with a Large Amount of Unlabeled Data. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 880–884. [Google Scholar]
  71. Ito, A.; Aiba, A.; Ito, M.; Makino, S. Detection of Abnormal Sound Using Multi-Stage GMM for Surveillance Microphone. In Proceedings of the 2009 Fifth International Conference on Information Assurance and Security, Xi’an, China, 18–20 August 2009; Volume 1, pp. 733–736. [Google Scholar]
  72. Cheng, C.F.; Rashidi, A.; Davenport, M.A.; Anderson, D. Audio Signal Processing for Activity Recognition of Construction Heavy Equipment. In Proceedings of the ISARC 2016—33rd International Symposium on Automation and Robotics in Construction, Auburn, AL, USA, 18–21 July 2016; pp. 642–650. [Google Scholar]
  73. Ellis, D.P.W.; Zeng, X.; McDermott, J.H. Classifying Soundtracks with Audio Texture Features. In Proceedings of the 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 5880–5883. [Google Scholar]
  74. Cheng, C.-F.; Rashidi, A.; Davenport, M.A.; Anderson, D.V. Activity analysis of construction equipment using audio signals and support vector machines. Autom. Constr. 2017, 81, 240–253. [Google Scholar] [CrossRef]
  75. Zhang, T.; Lee, Y.-C.; Scarpiniti, M.; Uncini, A. A supervised machine learning-based sound identification for construction activity monitoring and performance evaluation. In Proceedings of the Construction Research Congress, New Orleans, LA, USA, 2–4 April 2018; pp. 358–366. [Google Scholar]
  76. Marchi, E.; Vesperini, F.; Eyben, F.; Squartini, S.; Schuller, B. A Novel Approach for Automatic Acoustic Novelty Detection Using a Denoising Autoencoder with Bidirectional LSTM Neural Networks. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane, QLD, Australia, 19–24 April 2015; pp. 1996–2000. [Google Scholar]
  77. Principi, E.; Vesperini, F.; Squartini, S.; Piazza, F. Acoustic Novelty Detection with Adversarial Autoencoders. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 3324–3330. [Google Scholar]
  78. Azizyan, M.; Choudhury, R.R. SurroundSense: Mobile Phone Localization Using Ambient Sound and Light. SIGMOBILE Mob. Comput. Commun. Rev. 2009, 13, 69–72. [Google Scholar] [CrossRef]
  79. Satoh, H.; Suzuki, M.; Tahiro, Y.; Morikawa, H. Ambient Sound-Based Proximity Detection with Smartphones. In Proceedings of the 11th ACM Conference on Embedded Networked Sensor Systems; Association for Computing Machinery, Rome, Italy, 11–15 November 2013. [Google Scholar]
  80. Tarzia, S.P.; Dinda, P.A.; Dick, R.P.; Memik, G. Indoor Localization without Infrastructure Using the Acoustic Background Spectrum. In Proceedings of the MobiSys’11—Compilation Proceedings of the 9th International Conference on Mobile Systems, Applications and Services and Co-located Workshops, Washington, DC, USA, 28 June–1 July 2011. [Google Scholar]
  81. Wirz, M.; Roggen, D.; Tröster, G. A Wearable, Ambient Sound-Based Approach for Infrastructureless Fuzzy Proximity Estimation. In Proceedings of the International Symposium on Wearable Computers (ISWC) 2010, Seoul, Republic of Korea, 10–13 October 2010; pp. 1–4. [Google Scholar]
  82. Adavanne, S.; Politis, A.; Nikunen, J.; Virtanen, T. Sound Event Localization and Detection of Overlapping Sources Using Convolutional Recurrent Neural Networks. IEEE J. Sel. Top. Signal Process. 2018, 13, 34–48. [Google Scholar] [CrossRef]
  83. Lu, H.; Li, Y.; Mu, S.; Wang, D.; Kim, H.; Serikawa, S. Motor Anomaly Detection for Unmanned Aerial Vehicles Using Reinforcement Learning. IEEE Internet Things J. 2018, 5, 2315–2322. [Google Scholar] [CrossRef]
  84. Lee, J.; Yang, K. Mobile Device-Based Struck-By Hazard Recognition in Construction Using a High-Frequency Sound. Sensors 2022, 22, 3482. [Google Scholar] [CrossRef] [PubMed]
  85. Elelu, K.; Le, T.; Le, C. Collision Hazard Detection for Construction Worker Safety Using Audio Surveillance. J. Constr. Eng. Manag. 2023, 149, 4022159. [Google Scholar] [CrossRef]
  86. Elelu, K.; Le, T.; Le, C. Collision Hazards Detection for Construction Workers Safety Using Equipment Sound Data. In Proceedings of the 9th International Conference on Construction Engineering and Project Management, Las Vegas, NV, USA, 20–24 June 2022; pp. 736–743. [Google Scholar]
  87. Elelu, K.; Le, T.; Le, C. Direction of Arrival of Equipment Sound in the Construction Industry. In Proceedings of the 39th International Symposium on Automation and Robotics in Construction, Bogota, Colombia, 13–15 July 2022. [Google Scholar]
  88. Lee, Y.-C.; Shariatfar, M.; Rashidi, A.; Lee, H.W. Evidence-driven sound detection for prenotification and identification of construction safety hazards and accidents. Autom. Constr. 2020, 113, 103127. [Google Scholar] [CrossRef]
  89. Lee, Y.-C.; Scarpiniti, M.; Uncini, A. Advanced Sound Classifiers and Performance Analyses for Accurate Audio-Based Construction Project Monitoring. J. Comput. Civ. Eng. 2020, 34, 4020030. [Google Scholar] [CrossRef]
  90. Xie, Y.; Lee, Y.C.; Shariatfar, M.; Zhang, Z.D.; Rashidi, A.; Lee, H.W. Historical accident and injury database-driven audio-based autonomous construction safety surveillance. Computing in Civil Engineering 2019: Data, Sensing, and Analytics, Atlanta, GA, USA, 17–19 June 2019; pp. 105–113. [Google Scholar]
  91. Sabillon, C.A.; Rashidi, A.; Samanta, B.; Cheng, C.-F.; Davenport, M.A.; Anderson, D. V A Productivity Forecasting System for Construction Cyclic Operations Using Audio Signals and a Bayesian Approach. In Proceedings of the Construction Research Congress 2018, New Orleans, LA, USA, 2–4 April 2018; American Society of Civil Engineers: Reston, VA, USA, 2018; pp. 295–304. [Google Scholar]
  92. Cho, C.; Lee, Y.-C.; Zhang, T. Sound Recognition Techniques for Multi-Layered Construction Activities and Events. In Proceedings of the Computing in Civil Engineering 2017, Seattle, WA, USA, 25–27 June 2017; American Society of Civil Engineers: Reston, VA, USA; pp. 326–334. [Google Scholar]
  93. Liu, Z.; Li, S. A Sound Monitoring System for Prevention of Underground Pipeline Damage Caused by Construction. Autom. Constr. 2020, 113, 103125. [Google Scholar] [CrossRef]
  94. Sherafat, B.; Rashidi, A.; Lee, Y.; Ahn, C. Automated Activity Recognition of Construction Equipment Using a Data Fusion Approach. In Proceedings of the The 2019 ASCE International Conference on Computing in Civil EngineeringAt: Georgia Institute of Technology, Atlanta, GA, USA, 17–19 June 2019. [Google Scholar]
  95. Wei, W.; Wang, C.; Lee, Y. BIM-Based Construction Noise Hazard Prediction and Visualization for Occupational Safety and Health Awareness Improvement. In Proceedings of the ASCE International Workshop on Computing in Civil Engineering 2017, Seattle, WA, USA, 25–27 June 2017. [Google Scholar]
  96. Kim, D.; Liu, M.; Lee, S.H.; Kamat, V.R. Remote Proximity Monitoring between Mobile Construction Resources Using Camera-Mounted UAVs. Autom. Constr. 2019, 99, 168–182. [Google Scholar] [CrossRef]
  97. Son, H.; Seong, H.; Choi, H.; Kim, C. Real-Time Vision-Based Warning System for Prevention of Collisions between Workers and Heavy Equipment. J. Comput. Civ. Eng. 2019, 33, 4019029. [Google Scholar] [CrossRef]
  98. Huang, Y.; Hammad, A.; Zhu, Z. Providing Proximity Alerts to Workers on Construction Sites Using Bluetooth Low Energy RTLS. Autom. Constr. 2021, 132, 103928. [Google Scholar] [CrossRef]
  99. Scarpiniti, M.; Comminiello, D.; Uncini, A.; Lee, Y.-C. Deep Recurrent Neural Networks for Audio Classification in Construction Sites. In Proceedings of the 2020 28th European Signal Processing Conference (EUSIPCO) Virtual Conference, Amsterdam, The Netherlands, 18–22 January 2021; pp. 810–814. [Google Scholar]
  100. Cho, C.; Park, J. An Embedded Sensory System for Worker Safety: Prototype Development and Evaluation. Sensors 2018, 18, 1200. [Google Scholar] [CrossRef]
  101. Sakhakarmi, S.; Park, J.; Singh, A. Tactile-based wearable system for improved hazard perception of worker and equipment collision. Autom. Constr. 2021, 125, 103613. [Google Scholar] [CrossRef]
  102. Goodman, S.; Kirchner, S.; Guttman, R.; Jain, D.; Froehlich, J.; Findlater, L. Evaluating Smartwatch-Based Sound Feedback for Deaf and Hard-of-Hearing Users Across Contexts. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25–30 April 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 1–13. [Google Scholar]
  103. Ching, K.W.; Singh, M.M. Wearable Technology Devices Security and Privacy Vulnerability Analysis. Int. J. Netw. Secur. Its Appl. 2016, 8, 19–30. [Google Scholar] [CrossRef]
  104. Arabtali, B.; Solhi, M.; Shojaeezadeh, D.; Gohari, M. Related factors of using hearing protection device based on the protection motivation theory in Shoga factory workers. Iran Occup. Health 2015, 12, 12–19. [Google Scholar]
  105. McCullagh, M.; Lusk, S.L.; Ronis, D.L. Factors Influencing Use of Hearing Protection among Farmers: A Test of the Pender Health Promotion Model. Nurs. Res. 2002, 51, 33–39. [Google Scholar] [CrossRef]
  106. Nélisse, H.; Gaudreau, M.-A.; Boutin, J.; Voix, J.; Laville, F. Measurement of Hearing Protection Devices Performance in the Workplace during Full-Shift Working Operations. Ann. Occup. Hyg. 2011, 56, 221–232. [Google Scholar] [CrossRef]
  107. Reddy, R.K.; Welch, D.; Thorne, P.; Ameratunga, S. Hearing Protection Use in Manufacturing Workers: A Qualitative Study. Noise Health 2012, 14, 202–209. [Google Scholar] [CrossRef]
Figure 1. Total number of construction accidents by type in one year combined in the US, UK, and Israel [35].
Figure 1. Total number of construction accidents by type in one year combined in the US, UK, and Israel [35].
Sensors 22 09135 g001
Figure 2. Construction hazards detected by acoustic signals.
Figure 2. Construction hazards detected by acoustic signals.
Sensors 22 09135 g002
Figure 3. Research methods and procedure of this study.
Figure 3. Research methods and procedure of this study.
Sensors 22 09135 g003
Figure 4. Identify the scope of this review based on keywords.
Figure 4. Identify the scope of this review based on keywords.
Sensors 22 09135 g004
Figure 5. The number of reviewed articles by main contents.
Figure 5. The number of reviewed articles by main contents.
Sensors 22 09135 g005
Figure 6. The number of reviewed articles by years.
Figure 6. The number of reviewed articles by years.
Sensors 22 09135 g006
Figure 7. Indicators of hazardous events.
Figure 7. Indicators of hazardous events.
Sensors 22 09135 g007
Figure 8. Overall architecture of auditory signal processing to detect hazardous situations.
Figure 8. Overall architecture of auditory signal processing to detect hazardous situations.
Sensors 22 09135 g008
Figure 9. Main structure of AI-based sound classification.
Figure 9. Main structure of AI-based sound classification.
Sensors 22 09135 g009
Figure 10. Graphic representation of acoustic features extracted from the sound of a moving grader.
Figure 10. Graphic representation of acoustic features extracted from the sound of a moving grader.
Sensors 22 09135 g010
Figure 11. The number of reviewed articles by training models.
Figure 11. The number of reviewed articles by training models.
Sensors 22 09135 g011
Table 1. Summary of accidents caused by construction equipment between 1990 and 2007 in USA, adopted from [36].
Table 1. Summary of accidents caused by construction equipment between 1990 and 2007 in USA, adopted from [36].
No.Type of EquipmentNumber of Fatality CasesPercentage of Cases in Reverse DirectionPercentage of Cases without Reverse Alarms
1Dump truck17391.1%12.1%
2Excavator/backhoe5053.3%4%
3Private vehicle (car, pickup, van)4228.6%-
4Dozer3882.4%-
5Grader3791.9%13.5%
6Front-end loader3172.4%-
7Forklift3057.1%-
8Tractor-trailer2454.5%-
9Compactor1882.4%-
10Scraper1580.0%26.7%
11Skid Steer Loader1592.9%-
12Water truck1391.7%15.4%
13Paver8--
14Tractor (agricultural)6--
15Crane4--
16Sweeper4--
Note: empty cells indicate that the figures were not mentioned in the reference.
Table 2. Summary of factors causing the decline in auditory situational awareness of construction workers.
Table 2. Summary of factors causing the decline in auditory situational awareness of construction workers.
FactorsConclusionReferences
Jobsite-related1There are a significant number of sound sources in the acoustic event at construction sitesEven normal hearing listeners have a poor performance when complex auditory events happen[38]
2Construction sites are considered noisy workplacesImportant sounds can be easily ignored or misidentified when there is the presence of extraneous noise[39]
Worker-related3The majority of construction workers are maleMen have higher rates of difficulty in hearing than women[3]
4Construction workers may experience hearing loss (due to hazardous noise)Noise-exposed workers with hearing loss face many problems of job safety as the result of a reduction of hearing ability to warning signals[3]
Table 3. Applications of sound surveillance.
Table 3. Applications of sound surveillance.
ApplicationReferences
Home security[23]
Detection of critical situations in a railway system[24]
General surveillance in public areas[16,17,19,20,21,22]
Detection of crimes in elevators[25]
Detection of human emotions in public spaces[18,26]
Office surveillance[29]
Medical and health care facilities
Surveillance of the elderly, the convalescent, or pregnant women[27]
Automatic fall detection to improve the quality of life for independent older adults[28]
Industrial plants
Fault diagnosis of an induction motor[31]
Detection of abnormality or failure of equipment[32]
Table 5. Summary of traditional machine learning techniques for sound classification.
Table 5. Summary of traditional machine learning techniques for sound classification.
TechniquesReferences
1Gaussian Mixture Models (GMM)[16,17,19,20,23,24,25,27,62,68,71]
2Support Vector Machine (SVM)[11,21,24,28,30,44,57,72,73,74]
3K-Means Algorithms[21,29,46,48,51]
4Hidden Markov Models (HMM)[22,45,47,63,75]
5K-Nearest Neighbor Algorithms (KNN)[31,57]
Table 6. Summary of deep learning techniques for sound classification.
Table 6. Summary of deep learning techniques for sound classification.
TechniquesReferences
Deep Neural Network (DNN)[58,59,60]
Convolutional Neural Network (CNN)[9,12,28,46,70]
Recurrent Neural Network (RNN)/Bi-Directional Long Short-Term Memory (BLSTM)[62,76]
Auto Encoder (AE)[32,76,77]
Table 7. Summary of methods for sound localization.
Table 7. Summary of methods for sound localization.
MethodsReferences
1. Maximum-Likelihood Generalized Cross Correlation (GCC)[17]
2. Linear-correction least-square localization algorithm[17]
3. Similarity measure based on the Euclidean Distance (EUD) in the time domain[78]
4. Similarity measure based on Normalized Cross Correlation (NCC) in the frequency domain[79]
5. Similarity measure based on the Euclidean Distance (EUD) in the frequency domain[80]
6. Fast Fourier Transform (FFT)[80]
7. Fingerprinting algorithm[81]
Table 8. Summary of auditory event characteristics in the construction field.
Table 8. Summary of auditory event characteristics in the construction field.
CodeAuditory Event CharacteristicsReferences
AAbnormality of sound
Greater than the threshold?[71]
BDirection of moving sound source
Towards the worker?[78,79,80,81]
CLocation of sound
Less than a safe distance to the worker?[78,79,80,81]
DType of sound
D.1Alert alarm (fire/earthquake)[23,25,76,77]
D.2Explosion[19]
D.3Gunshot[16,17,19,20,25,61,62]
D.4Humana. Announcement[23]
b. Crowd ambient sound[26]
c. Scream/shout/cry[17,18,19,20,23,24,25,26,27,28,61,62,68,76,77]
D.5Moving heavy equipment/machine[30,33,36]
Table 9. List of hazardous situations and required auditory event characteristics.
Table 9. List of hazardous situations and required auditory event characteristics.
No.Hazardous SituationsCombination of Auditory Event Characteristics
1Heavy equipment/machine is at an unsafe distanceD.5 + C
2Heavy equipment/machine is approachingD.5 + C + B
3Heavy equipment/machine is operating in an abnormal conditionD.5 + A
4There is someone screaming/shouting/cryingD.4c
5There is a crowd approachingD.4b + B
6There is an abnormal crowd approachingD.4b + B + A
7There is an alert alarmD.1
8There is an explosionD.2
9There is a gunshotD.3
10There is an abnormal soundA
Table 10. Influencing factors of hearing protection devices.
Table 10. Influencing factors of hearing protection devices.
No.Influencing FactorsReferences
1Planning effective interventions[104]
2Workplace interpersonal interaction[105,107]
3Poor insertion and poor fitting of earplugs[106]
4Left and right ear differences[106]
5Perception of noise[107]
6Hearing protection use[107]
7Reluctance to use HPDs[107]
8Value of hearing influence hearing protection use[104,107]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dang, K.; Elelu, K.; Le, T.; Le, C. Augmented Hearing of Auditory Safety Cues for Construction Workers: A Systematic Literature Review. Sensors 2022, 22, 9135. https://doi.org/10.3390/s22239135

AMA Style

Dang K, Elelu K, Le T, Le C. Augmented Hearing of Auditory Safety Cues for Construction Workers: A Systematic Literature Review. Sensors. 2022; 22(23):9135. https://doi.org/10.3390/s22239135

Chicago/Turabian Style

Dang, Khang, Kehinde Elelu, Tuyen Le, and Chau Le. 2022. "Augmented Hearing of Auditory Safety Cues for Construction Workers: A Systematic Literature Review" Sensors 22, no. 23: 9135. https://doi.org/10.3390/s22239135

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop