Next Article in Journal
Embroidered Interdigitated Electrodes (IDTs) with Wireless Readout for Continuous Biomarker Monitoring
Previous Article in Journal
A Wideband High Gain Differential Patch Antenna Featuring In-Phase Radiating Apertures
Previous Article in Special Issue
Soundscape Characterization Using Autoencoders and Unsupervised Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Biases in Ecoacoustics Analysis: A Protocol to Equalize Audio Recorders

by
Andrea Potenza
1,*,
Valentina Zaffaroni-Caorsi
1,
Roberto Benocci
1,
Giorgia Guagliumi
1,
Jalal M. Fouani
2,
Alessandro Bisceglie
1 and
Giovanni Zambon
1
1
Department of Earth and Environmental Sciences, University of Milano-Bicocca, 20126 Milan, Italy
2
Independent Researcher, 40132 Bologna, Italy
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(14), 4642; https://doi.org/10.3390/s24144642
Submission received: 24 May 2024 / Revised: 28 June 2024 / Accepted: 2 July 2024 / Published: 17 July 2024
(This article belongs to the Special Issue Advances in Acoustic Sensors and Deep Audio Pattern Recognition)

Abstract

:
Eco-acoustic indices allow us to rapidly evaluate habitats and ecosystems and derive information about anthropophonic impacts. However, it is proven that indices’ values and trends are not comparable between studies. These incongruences may be caused by the availability on the market of recorders with different characteristics and costs. Thus, there is a need to reduce these biases and incongruences to ensure an accurate analysis and comparison between soundscape ecology studies and habitat assessments. In this study, we propose and validate an audio recording equalization protocol to reduce eco-acoustic indices’ biases, by testing three soundscape recorder models: Song Meter Micro, Soundscape Explorer Terrestrial and Audiomoth. The equalization process aligns the signal amplitude and frequency response of the soundscape recorders to those of a type 1 level meter. The adjustment was made in MATLAB R2023a using a filter curve generated comparing a reference signal (white noise); the measurements were performed in an anechoic chamber using 11 audio sensors and a type 1 sound level meter (able to produce a .WAV file). The statistical validation of the procedure was performed on recordings obtained in an urban and Regional Park (Italy) assessing a significant reduction in indices’ biases on the Song Meter Micro and Audiomoth.

1. Introduction

The production of pocket digital recorders in the past few decades allowed researchers and amateurs to record many hours of environmental noise and sounds [1]. These recorders along with mobile phones have been implemented by researchers into their work and nowadays are used in various acoustic research branches (e.g., noise mapping [2,3], soundscape assessments [4,5,6,7], bio-acoustics research and sound event detection through machine learning [8,9]).
This study focuses on an equalization procedure to reduce biases in the comparison of audio recorders which can be used in all the disciplines mentioned before. However, in this study, the evaluation of the procedure will be carried out focusing on soundscape ecology given the authors’ main field of research (soundscape ecology). Therefore, the term “soundscape” will be used in the terms of this discipline from here on.
Soundscape ecology studies witnessed remarkable growth in the past few decades [10] thanks to new technologies that allowed for the production of different autonomous passive recorders [11]. These developments boosted the investigation of soundscape environments (terrestrial and maritime) and the role of anthropogenic noise on habitat degradation and biodiversity loss.
The exploration of soundscapes involves the application of passive acoustic monitoring (PAM), a method that entails deploying recording devices in the studied area and retrieving them after days or weeks. PAM serves various purposes, from monitoring a particular taxon to recording sounds in a specific environment [11]. To achieve this, researchers employ terrestrial recorders with specific characteristics like sensitivity, frequency response, dynamic range and bandwidth limits. The analysis of recordings relies mainly on eco-acoustic indices [12,13,14] which allow for summarizing the audio information and classifying soundscapes by analyzing their pitch, saturation and amplitude. This process involves confronting time steps or frequency bins to discern patterns and variations [14]. Soundscape monitoring can be carried out in different spatial designs using a regular grid of devices or arrays of sensors operating in tandem [11,15]. While the majority of studies use one to three recorders, a minority deploy more than ten instruments in the field [15,16,17,18]. Given these monitoring spatial designs, the use of all available devices, regardless of the manufacturers, is an attractive strategy to expand the study area at reduced costs.
The main challenge of this wide array of available recording instruments lies in the potential introduction of biases since recorders of different brands will mostly produce audio files with diverse properties when exposed to the same soundscape (i.e., amplitude values over the spectrum). Moreover, these biases can also be present in devices of the same brand and model due to the inherent uncertainty in sensitivity and frequency response. For these reasons, comparing sound pressure levels (SPLs) or eco-acoustic indices’ values between recordings made with instruments located in different parts of a forest or different habitats can be problematic. Even when using a single-sensor model, these biases persist due to manufacturing variability, as assessed in [16] and declared by manufacturers in their user guides (i.e., sensitivity uncertainty range). The bias increases when different models are employed due to frequency response differences, wider sensitivity ranges, dynamic range and bandwidth limits. Furthermore, these biases may cause incongruities in indices’ values and trends if the comparison is performed between different areas using different recorders [19,20,21]. In fact, previous studies have evidenced that differences in the frequency response led to conflicting conclusions when using indices as proxies for biodiversity or soundscape characteristics [22]. Furthermore, it is common to observe a non-correspondence between eco-acoustic indices’ trends, like ACI values [19] in studies that compare eco-acoustic indices across different regions or of different habitats. In [21], Sethi and colleagues analyzed four diverse datasets (temperate forest in the USA, rainforest in Malaysia, agricultural tea landscape in Taiwan and grassland in India) investigating the correlation between acoustic features (i.e., convolutional neural network and 60 soundscape indices) and avian species richness. The calculated indices also included the acoustic complexity index (ACI), the acoustic diversity index (ADI), the bio-acoustic index (BI), the normalized difference soundscape index (NDSI) and the temporal and spectral entropy (Ht, Hf) [23]. As they reported, there was not a single eco-acoustic index correlated to avian richness across all datasets. The authors hypothesized that these incongruities may be due to comparing diverse habitats and ecosystems at different latitudes as well as using audio recorders of different brands or in different settings (i.e., spatial deployment) and conditions (i.e., rain protection bags).
To overcome these biases, it is crucial to equalize the audio files acquired from various recorders. Equalization involves processing these audio files using a curve derived from comparing two recordings of white noise: one obtained with the soundscape sensor and another with a type 1 sound level meter, which serves as reference. This process enables the generation of soundscape recordings with a “flat” frequency response (akin to that of the sound level meter), ensuring consistent pressure levels and eco-acoustic indices across soundscape devices.
The necessity of granting an accurate comparison between recorders has been previously recognized in other papers which proposed calibration methods or pipelines to study the spectral portions in which the recordings are comparable [24,25,26].
In [24], three procedures with various accuracy to calibrate audio files were proposed. Among these, the most effective approach involves calibrating the recording using a known frequency level as a reference value. Unfortunately, the different sensitivity curves of the instruments (which are not flat) do not allow for a perfect calibration, and thus, an equalization procedure is necessary.
In [25], the authors perform an equalization on one-third octave bands between 0.025 and 6.3 kHz. Their work is an initial step in facilitating accurate comparisons of soundscape recordings.
Finally, in [26], the authors proposed a pipeline to compare the eco-acoustic indices between Audiomoth and SM4 devices. In their article, Luna-Naranjo and colleagues addressed the inherent variability in these two devices mainly by selecting comparable frequency ranges between the two instruments (on which the eco-acoustic indices are calculated) and normalizing the signal amplitude.
In this study, we propose an equalization protocol aiming at mitigating biases inherent in soundscape recordings obtained from three different devices (Song Meter Micro, Audiomoth and Soundscape Explorer Terrestrial), by aligning these audio files to those captured by class 1 level meter (LD831-C), a standard reference device. This process entails adjusting the signal amplitude and frequency response of the soundscape recorders to match those of the level meter, ensuring consistency across the entire spectrum. Alongside the protocol proposal, its validation is carried out in this paper to show its efficacy. As a result, we aim to help researchers with further comparisons within and outside the study area with greater reliability and precision, enabling a more complete understanding of acoustic ecosystems and their dynamics.
This paper is structured around four key pillars, outlined as follows:
  • An explanation of the equalization procedure’s steps;
  • Recording white noise to calculate the equalization curves;
  • The application of the equalization curves to the soundscape recordings;
  • The validation of the procedure using three case studies.

2. Materials and Methods

In this section, we describe the instrumentation and steps followed for the development of the equalization process:
  • Obtaining white noise recordings to generate the equalization curves;
  • Defining the equalization procedure;
  • Validating the equalization process
    • On white noise measures (obtained in the lab);
    • In an in-field experiment placing the devices in a single measurement point;
  • Providing an example of a monitoring campaign carried out in a Regional Park using two different brands of devices.

2.1. Materials

For the creation of this protocol, we considered three models of different brands of soundscape recorders and a sound level meter (Figure 1):
  • Song Meter Micro (Wildlife Acoustics, Inc., Maynard, MA, USA);
  • Soundscape Explorer Terrestrial (Lunilettronik Coop S.p.a., Fivizzano, Italy);
  • Audiomoth (Open Acoustic Devices, Oxford, UK);
  • LD831-C (Larson-Davis, Depew, NY, USA).
These models were chosen given their diffusion in the soundscape ecology community, their relative economical cost and the different sensitivity curves. We used a class 1 level meter to obtain the best input result.
The Song Meter Micro (SMM hereafter) is a passive recorder with a maximum sampling rate of 96 kHz. The sensitivity of the whole signal transmission chain (i.e., microphone, gain and analog-to-digital converter) is 2 dBV ± 4 dBV relative to 1 Pa at 1 kHz Full-Scale, measured using a gain of +18 dB. The sensitivity curve is not linear with frequency (Figure 2). It generates output files in wave format, 16-bit and not compressed. The maximum recording length is 60 min. It works using three AA batteries. The model tested here is the first version of the Song Meter Micro series.
The Soundscape Explorer Terrestrial (SET hereafter) is a programmable recorder equipped with two microphones (with a sampling rate of 48 and 192 kHz, respectively) and environmental sensors (for humidity, temperature, light and atmospheric pressure). The microphone has a sensitivity of −28 dBV ± 3 dBV relative to 1 Pa at 1 kHz; its frequency response is almost flat up to 6 kHz (Figure 3). It generates output files in wave format, 16-bit and not compressed. It works using eight AA batteries.
The Audiomoth 1.2.0 (AM hereafter) is a programmable, low-cost recorder equipped with an analog MEMS (Micro Electrical–Mechanical System) microphone with a maximum sampling rate of 384 kHz. The sensitivity of the microphone is −38 dBV ± 6 dBV relative to 1 Pa at 1 kHz, and its frequency response is almost flat (without the waterproof case) (Figure 4). An analysis of the response with and without the waterproof case is carried out. Its output files are .wav, 16-bit and not compressed. It runs using three AA batteries.
Finally, we used as a reference device the sound level meter of class “A” LD-831C (831 hereafter) produced by Larson-Davis. It is equipped with a class A microphone, calibrated in a certified laboratory, with a sensitivity of −26.19 dBV relative to 1 Pa at 251.2 Hz. Moreover, it has a flat response over all frequencies in the interval 0–20 kHz. This level meter can produce an uncompressed .wav file, 16-bit.

2.2. Methods

2.2.1. Recording White Noise in the Anechoic Chamber

The first step of the equalization process involves acquiring white noise recordings using the soundscape devices and the level meter simultaneously.
These measurements were carried out in an anechoic chamber, 4 × 4 m and 2.8 m high, configured to operate in “full anechoic” mode. The chamber can also operate in the “semi-anechoic” configuration (walkable floor without pyramid absorbers), but the “full anechoic” mode ensures an optimal background noise level of 10 dBA for precise recordings. Finally, the cut-off frequencies are around 120 Hz due to the size of the chamber and of the sound-absorbing elements.
The reference signal (white noise) was built using Audacity 3.0.2 [27] with a sample rate of 48 kHz, RMS value of −18.74 dBFS and a duration of 20 s. To reproduce the signal in the anechoic chamber, we employed the loudspeaker TD508 Mk3 of the Eclipse series (Denso Ten, Kobe, Japan) (Figure 5). This speaker has a sensitivity of 82 dB/W/m, emitting between 52 Hz and 27 kHz with an angular coverage of +15/−10°.
Measurements were performed by placing one device at a time in the same position: 1 m from the loudspeaker, with the microphone facing the sound source and vertically centered in the middle of the woofer (Figure 5).
The appropriate functioning and frequency response of the loudspeaker were tested previous to the measurements. The results showed that the white noise emitted by the device was not flat (Figure 6). For this reason, the loudspeaker output was recorded with the level meter and equalized to generate a new output in the range 24–15 kHz that was perfectly flat when measured with the level meter. The equalization of the loudspeaker was performed using the white noise generated with Audacity and the signal recorded with the level meter when emitted by the loudspeaker.
The following settings were used to measure the white noise with the soundscape recorders (Table 1). To understand intra-model variability, we tested all the devices available in our laboratory and some kindly lent to us by Professor Guarnaccia of the Università degli Studi di Salerno; the devices were set as usually employed in the field by the research groups. All devices were set with parameter values used by our group during in-field monitoring [13,16,28,29]. The sampling frequency was set at 48 kHz to correctly sample all birds’ vocalizations. The amplitude gains were set for SMM at +18 dB since this is the default value suggested by Wildlife Acoustic, for SET, two values were analyzed due to in-field choices performed by the group during the years and the AM was set at “medium” (+15 dB) since it is the nearest value to the other device’s amplitude gains. The level meter gain was set to +0 dB since the device choice is between 0 or +20 dB. Finally, the level meter was employed without the windproof cap since the measurement was performed in a closed environment, while the AM was analyzed with and without the waterproof case to better understand its effect on the sensitivity curve.

2.2.2. Equalization Procedure

The equalization procedure consists of calculating an equalization curve using the white noise measures and then applying it to in-field recordings. The process is carried out in the MATLAB environment.

Calculating the Equalization Curve

The process to generate the equalization curves is summarized in Figure 7, and the MATLAB script is available in the Supplementary Materials Section. This procedure is an amended version of the one conducted in [30] for the aim of this study and to be able to run it through MATLAB R2023a [31].
The curve is used to generate a filter in the frequency domain which is applied to the in-field recordings.
The first step consists of calculating the power spectral density (PSD) of the level meter and soundscape white noise recordings, which describes the frequency distribution of the signal’s power [32].
These vectors are compared by dividing the PSD of the level meter by the soundscape device one.
The curve is finally calculated using the “fir2” function. This function returns an nth-order filter with frequency–magnitude characteristics by interpolating the desired frequency response onto a dense grid and then using the inverse Fourier transform and a Hamming window. The input parameters are the ratio between the PSDs, the number of FFT points and the frequency vector.
The equalization curve is calculated for each device since the frequency response is diverse even between recorders of the same brand.
Furthermore, considering that the soundscape devices have amplitude gain (Table 1), the white noise recorded by the 831 was amplified to +18 dB before computing the equalization curves. This gain was applied to avoid altering the gains set in the soundscape devices which are optimized to record distant sound events such as birds’ vocalizations. This operation was implemented using Audacity 3.0.2, and a rapid evaluation showed that the gain was linearly applied on all frequencies without any distortion on the recordings. To ensure an accurate comparison of the devices after the equalization, the calculation of the curves was performed on a single white noise 831 recording amplified at +18 dB.

Equalization of In-Field Recordings

After the implementation of the equalization, the curves were applied to the in-field recordings (using a MATLAB script available in the Supplementary Materials Section) following the scheme described in Figure 8.
The first step consists of the correction of the possible presence of the DC offset. This fixed voltage offset present in the audio chain of the device is visible as a shift of the waveform from the 0.0 horizontal center. It affects the calculation of the eco-acoustic indices since it alters the frequency domain (being a direct current, it manifests as an intensity peak in the first frequency bin of the signal). The removal of the DC offset is conducted by subtracting the offset to the signal; the offset value is determined utilizing the mean estimation method, wherein it corresponds to the mean value of the audio signal [33].
The second passage is the application of the equalization curve to the field recording. It is performed using a rational transfer function (“filter”) defined by the numerator and denominator coefficients; the numerator is set to 1, while the denominator is the result of the “fir2” function, in other words, the equalization curve.
Finally, since filtering a signal introduces a delay (i.e., the output signal is shifted in time), the delay is calculated using the “grpdelay” function and then corrected. This function calculates the delay of a narrow-band “group” of sinusoidal components; if the filter has a linear phase response, the group delay and phase delay are identical [34].
The equalized recording is finally saved in .wav format.

Parameters Set in the Equalization Script

In the calculation of the equalization curve script, the number of FFT points was given in input to calculate the power spectral density. The number of FFT points was also applied to implement the curve using the “fir2” function; in particular, it was set as the filter’s order value. After analyzing their frequency response vector and angular frequency vector (Figure 9), we validated the equalization procedure using values of 512, 1024 and 16,384, since they allowed for good accuracy without excessive phase changes. This choice allowed us to obtain three calculated curves for each device.
The delay induced by the “filter” function consists of milliseconds of no-audio track placed at the beginning of the recording, shifting the start of the audio without increasing the time duration of the in-field file and thus losing the last recorded milliseconds. Given the linearity of the phase responses (Figure 9), the group delay and phase delay are identical [34]. The delay correction cuts those first milliseconds inserted by the filter, causing a reduction in the length of the recording; this reduction depends on the delay which depends on the order value: it ranges from 0.0053 s using a value of 512 to 0.17 s using a value of 16,384.

2.2.3. Validation of Equalization Process and Practical Example

The validation process (Figure 10) consists of evaluating the effects on the soundscape ecology analysis introduced by the proposed equalization procedure. Focusing on the three soundscape recorder models, it aims at identifying the best filter order to be used.
This validation process was carried out as follows:
  • On white noise measures (anechoic chamber measurements);
  • In an in-field experiment placing the devices at a single measurement point (urban park).
A practical example was used to better understand the effects of the procedure on eco-acoustic indices’ time trends derived from a monitoring campaign carried out in a Regional Park using two different brands of devices (SMM and SET also used in the previous phases of the validation process).
These three environments were considered due to their different characteristics:
  • The anechoic chamber provides an ideal environment for recording identical signals across all devices, facilitating the calculation of equalization curves and enabling a precise comparison of a singular recording for each device.
  • The monitoring in the urban park permits the evaluation of the effects of the equalization process in a real case scenario on nine 1 min recordings taken simultaneously at a single measurement site. This site is rich in traffic noise (i.e., cars, buses, motorcycles), birds’ vocalizations, cicadas’ vocalizations, human voices and airplane overflights; thus, the signals recorded are extremely diverse with events spanning the entire spectrum.
  • The example of the Regional Park is proposed as a classic monitoring scheme example [15,16,35]. Its soundscape was assessed by placing nine devices on a regular grid (each point distant 200 m from the others), and the eco-acoustic indices were calculated by analyzing a 24 h time trend both before and after the equalization process. This real-world scenario serves as an excellent example for testing the efficacy of the proposed protocol.

Eco-Acoustic Indices

The analysis of recordings in soundscape ecology is carried out using, among other methods, the eco-acoustic indices [12,13,14]. The indices calculated to validate the equalization procedure are the following:
  • ACI (acoustic complexity index): This quantifies the vocalizations of avifauna through the study of sound intensity modulation, which varies rapidly over time in the case of biophony but is very constant for numerous anthropogenic noises [14,36]. The implementation is based on the amplitude difference between adjacent time samples within a frequency band, relative to the total amplitude of that band [14].
  • ADI (acoustic diversity index): This provides a measure of the diversity of intensity distribution in the spectrum by dividing it into time intervals and calculating the Shannon index [13,14]. Low values are due to extreme diversity in intensity distribution (i.e., nocturnal insects [14]) while high values to a distribution evenness (i.e., high levels of geophony and anthropophony [14] and bird species richness [13]).
  • AEI (Acoustic Evenness Index): This is based on the same logic of ADI but applies the Gini coefficient instead of the Shannon index and consequently measures the inequality of signals in bands [37].
  • BI (bio-acoustic index): This measures avian abundance by calculating the area under the mean frequency spectrum in the frequency range occupied by biophonies and characterized by a certain amplitude value (this threshold value, expressed in dB, is greater than the lowest value represented in the spectrum) [38].
  • NDSI (normalized difference soundscape index): This assesses the distribution of the soundscape between anthropophony and biophony to estimate the level of anthropogenic disturbance of a habitat [39]. It is calculated by dividing the difference between biophony and anthropophony by their sum; the estimation of biophony and anthropophony is carried out by calculating the power spectral density in the frequency ranges of these soundscape components [39]. Values range in the [−1, +1] interval, where +1 ideally indicates the total dominance of biophonies and −1 the total dominance of anthropophonies [39,40].
  • H (Acoustic Entropy): This provides an estimate of the total entropy, or heterogeneity, of the recording. It is calculated by computing the product of Shannon’s spectral entropy and temporal entropy [41]. Values range in [0, +1]; +1 indicates an even signal (i.e., silent recording or faint bird calls), while 0 indicates a pure tone (i.e., insects’ vocalizations) [14].
  • DSC (Dynamic Spectral Centroid): This returns the spectral centroid of a recording (expressed in Hz) providing information about the sound events of a recording. It is calculated by dividing the spectrum in time intervals and computing the gravity center of the spectrum [13].
  • ZCR (Zero-Crossing Rate): This measures the number of times per second that a signal crosses the instantaneous pressure of 0 and provides a measure of noisiness; high values are associated with noisy recordings and the presence of biophony, while low values are linked to tonal sounds [42,43].
These indices were calculated in the “R” environment (version 2023.03.0) [44] using the packages “seewave” and “soundecology”. For computing the DSC, a dedicated script was written [13]. The scripts of the ADI and AEI from the “soundecology” package were modified to add the minimum frequency values and thus mirroring the other eco-acoustic indices (ACI, BI; NDSI, DSC). The indices were calculated using an FFT value of 1024 which corresponds to a frequency resolution of FR = 46.875 Hz and a time resolution TR = 1/FR = 0.0213 s. The time duration was set as the minimum of the three FFT values (59.83 s).
For the white noise and pocket park recordings, the indices parameters were set as the following:
  • ACI and DSC: min_freq = 500 Hz, max_freq = 12,000 Hz.
  • ADI and AEI: min_freq = 500 Hz, max_freq = 12,000 Hz, freq_step = 10 Hz, dB_threshold = −50 dB.
  • BI: min_freq = 1700 Hz, max_freq = 12,000 Hz.
  • NDSI: min_anthro_freq = 500 Hz, max_anthro_freq = 1700 Hz, min_bio_freq = 1700 Hz, max_bio_freq = 12,000 Hz.
  • H and ZCR: the entire spectrum.
The minimum frequency was set to 500 Hz due to the devices’ low sensitivity to low frequencies. In fact, even if the equalization process levels out the audio recordings to the one of a level meter, their original sensitivity to low frequencies does not allow for the optimal recording of sounds, and therefore, it is not possible to fully reproduce the fidelity of a sound level meter. For this reason, the minimum frequency was set at 500 Hz.
For the Regional Park monitoring, the indices’ parameters were the same except for the ADI and AEI for which the minimum frequency was set to 0 Hz and the dB_threshold to −50 dB.

Statistical Test

Regarding the simultaneous recordings acquired in the urban park, a statistical test was performed to assess the equivalence in the eco-acoustic indices derived from both the sound level meter and the soundscape devices (distinguishing between pre-processed and post-processed using 1 k, 16 k and 32 k FFT points). Given the simultaneity of the recordings, a pairwise test was chosen. Prior to conducting the pairwise Student’s test, Shapiro’s test and Bartlett’s test were employed to verify the assumptions of normality and homoscedasticity [45]. The indices that did not meet the assumptions were analyzed using the pairwise Wilcoxon test. The null hypothesis H0 of the Student and Wilcoxon test is that the mean difference is equal to 0; it is refused when p-value < 0.05.

White Noise (Anechoic Chamber)

The validation was first performed on the white noise recorded in the anechoic chamber. The recordings are not affected by the DC offset, probably due to the lack of humidity in the environment which does not alter the conductivity of the electronic circuit.
Two parameters were computed:
  • The root-mean-square error (RMSE) of the amplitude (1)
    RMSE amp = 1 n f i = f m i n f m a x ( A m p s l m ( f i ) A m p r e c o r d e r ( f i ) ) 2
This equation compares the soundscape recording to the level meter one (which serves as reference). It was performed on the entire spectrum, subtracting the amplitude of the soundscape recorder (recorder) from the one of the level meter (“slm”). It was then normalized by the number of frequency bins and rooted.
  • Percentage error on the eco-acoustic indices (ACI, ADI, AEI, BI, NDSI, H, DSC, ZCR) using the sound level meter as reference (2)
    % error index = I n d e x s l m I n d e x r e c o r d e r I n d e x s l m

Field Recordings at Bicocca Urban Park

This in-field validation was performed on a dedicated recording campaign carried out in June 2023 in a pocket park belonging to the University Campus, placing the devices at a single measurement point (Figure 11).
To obtain simultaneous recordings, each sensor was hung on a steel bar with the microphones oriented in the same direction. They were set with a sampling rate of 48 kHz (same gains used in the anechoic chamber, Table 1). The Audiomoths were used with their waterproof case.
Nine simultaneous 1 min recordings were acquired for each device. One of the Audiomoths did not work, and thus, it was not considered in the analysis.
Since the recorders were placed at a certain distance from the level meter, the recordings were examined to grant the registrations’ simultaneity. This operation was carried out in Audacity by selecting noticeable sound events and cutting the recordings with an accuracy at the millisecond level.
Validation was performed by calculating the eco-acoustic indices and Equations (1) and (2) and by comparing the indices calculated on the level meter with the ones of each soundscape sensor (using the three order values) through a statistical test. Given the simultaneity of the recordings, the Student pairwise test was carried out; the indices that did not meet the Student test assumptions (verified through Shapiro’s test and Bartlett’s test) were analyzed using the pairwise Wilcoxon test.
Just like for the 831′s white noise recording, the amplitude gain in the sound level meter was adjusted in the post-process phase to +18 dB.

Field Recordings in a Regional Park

Using data acquired in the previous steps of this study, the best number of FFT points to use for each device was detected. These parameters were used for the equalization of a monitoring campaign (24 h) performed in April 2022 at the Ticino Regional Park, Lanca del Moriano (Bereguardo, Italy), using the same devices employed in the anechoic chamber (in particular, 7 SMM and 1 SET) placed as in Figure 12. The AMs were not employed since the monitoring campaign was organized only by our research group.

3. Results

3.1. Validation of White Noise Measures

In this section, the mean frequency spectrums of the four devices are compared, before and after the equalization procedure (Figure 13) distinguishing the three cases (512, 1024 and 16,384 order values). In addition to Figure 13, the amplitude RMSE and percentage error on eco-acoustic indices are shown in the form of barplots (Figure 14).
The mean frequency spectrums of the devices before the equalization (Figure 13a) show the evident difference between the level meter and the soundscape recorders. As observable in both Figure 2 and Figure 13, SMMs present a sensitivity peak at 6–7 kHz and a lower response at low frequencies than the level meter. On the other hand, the SETs are characterized by an almost flat response up to 6 kHz which then drops. Finally, the AMs without the waterproof case (AMf) present a nearly flat response up to 5 kHz, while the AMs equipped with the case (AMw) show an extremely oscillating response; differences in the AM frequency response depending on the case were also assessed by [46].
After the equalization (Figure 13b–d), the differences in mean frequency response between devices were reduced in terms of the overall dB difference (see y-axis) and sensitivity oscillation; the amplitude RMSE decreased from a mean value of 10.7 dB in the original recordings to 0.06 dB in the post-processed recordings using the 16,384-order filter (see Table 2). Thus, the best results are noticeable using a 1024- and 16,384-order filter for calculating the equalization curves (i.e., smaller fluctuations of the mean frequency spectrums and less overall amplitude deviation).
Below in Figure 14 are reported the amplitude RMSE and percentage error on the eco-acoustic indices distinguishing the devices and their status (original, post-process using 512, 1 k, 16 k order values). In the case of the AM, only the data about the waterproof-cased one are reported since it was used with this configuration in the field monitoring.
Examining Figure 14, the benefits produced are clear by looking at the reduction in the parameters from the original audios (yellow bar) to the equalized audios (green, blue and pink). The best improvements are seen in the AMs (with the case) and SMMs since they are the ones with the most variable frequency response (Figure 13). The amplitude RMSE (A) is reduced for each device. The percentage errors on the ADI, AEI, BI, NDSI, H and DSC are greatly reduced since they analyze the repartition of intensity on the spectrum (which depends on the frequency response). Table 2 shows that the total percentage error of the eight indices is reduced from over 1000% to almost 20%, with some indices (AEI, BI and ZCR) bearing the majority of the error (Figure 11).
Regarding the filter order, the AM equipped with the waterproof case presents less biases when a value of 512 is used, the SETs with the 16,384-order filter and the SMM with values of 512 or 1024 depending on the index (Figure 11, Table 2). These differences may be explained by the different sensitivity curves of devices in comparison with the level meter one.

3.2. Validation of In-Field Single Measurement Site (Urban Park)

This section mirrors the previous one, showing the mean frequency spectrums of the four devices, before and after the equalization procedure (Figure 15).
The original mean spectrums (Figure 15a) show the differences between devices. SMMs and the AM with the waterproof case (AMw) deviate more from the level meter (831) and respectively present a 6–7 kHz peak and an oscillating trend (like in the anechoic chamber). The SETs’ trends are more similar to the level meter. After the equalization (Figure 15b–d), the SMMs’ trends are very similar to the level meter, especially in the 512 and 1024 cases, while the SETs’ trends shift nearer to the reference curve. Finally, the AM is still affected by the oscillatory trend but is indeed nearer to the level meter curve. These benefits can also be seen in Table 3 where the amplitude RMSE decreases from a mean value of 11.6 dB in the original recordings to 4.5 dB in the post-processed recordings using the 512- and 1024-order filter and to 4.8 dB using the 16,384-order filter.
Mirroring the white noise validation section, Figure 16 shows the amplitude RMSE and percentage error on eco-acoustic indices calculated on the nine one-minute recordings carried out in the urban park.
The boxplot analysis on the amplitude RMSE and the percentage error on the eco-acoustic indices (Figure 16) reflects the barplot analysis of the white noise recordings (Figure 14). We can assess the following:
  • The amplitude deviation is reduced for all devices (graph A);
  • The percentage error on the BI, NDSI, DSC and ZCR is greatly diminished, in particular for the AM and SMM; in Table 3, it is possible to notice this decrease in the total percentage error: it is 2000% for AM and 1580% for SMM while only 30% for SET 18 dB and 2% for SET 20 dB;
  • The ACI is not affected by the process since it compares adjacent temporal and frequency bins;
  • The percentage error on the ADI, AEI and H is reduced in the SMM and AM, while it is increased in the SETs.
This comparison is summarized in Table 3; the total percentage error on the AM and SMM is reduced by 2000% and 1580%, respectively, and in this view, the optimal choice of parameters is using an order of 1024 for AM and 16,384 for SMM, while the error improvement on the SETs is limited from the outset and is not further reduced.
For a complete overview, Figure 17 shows the boxplot of the eco-acoustic indices calculated on the sound level meter and the soundscape recorders. The Student and Wilcox pairwise tests were performed on the eco-acoustic indices to better understand the benefits of the equalization procedure.
Examining Figure 17, we can affirm that the indices calculated on the in-field recordings behave as follows:
  • ACI: The soundscape devices’ values are similar to the level meter ones due to the index calculation method. The AM is an exception probably due to its oscillatory frequency response. Looking at one device at a time, the variation in values between the original recordings and the equalized ones is constant or minor.
  • ADI: The equalized values of AM and SMM are more similar to the level meter ones, especially in the 512–1024 case for AM and 16,384 for SMM. On the contrary, the SETs are similar to the reference in the beginning and then deviate. This different behavior between the soundscape devices may be explained by the linearity of their sensitivity curves and thus of the intensities on the spectrum: the “flatter” the curve, the smaller the deviation of ADI values from the level meter ones.
  • AEI: As for the ADI, the AM and SMM values are nearer to the reference when equalized using a filter order of 1024 and 16,384, respectively. On the other hand, the SETs deviate when equalized. The deviation of the SET when equalized may be due to applying the procedure on a device which already presents a “flat” curve; trying to further linearize it generates errors since it is already linear.
  • BI: All sensors’ original values do not match the reference ones; this can be explained by the frequency response of the soundscape devices that are not linear above 5 kHz. With the equalization, the AM values match the level meters’, and the others are nearer to the level meter ones.
  • NDSI: Since this index greatly depends on the linearity of the frequency response, AM and SMMs are the ones with the greatest bias. It is possible to notice an improvement for all devices after the equalization, especially for AM and SMMs using an order value of 512.
  • H: AM and SMMs benefit from the process but are nowhere near the reference values. The SETs’ original values are nearer to the level meter ones than the equalized values due to their already linear sensitivity curve. Moreover, the similar behavior is present in the ADI and AEI which also evaluate the heterogeneity of the recordings.
  • DSC: AM and SMMs present the greatest biases due to their frequency response, as for the NDSI; the equalization reduces them, allowing for comparisons. The SETs do not benefit from the process thanks to their more linear frequency response.
  • ZCR: Even in this case, AM and SMMs are affected by biases which are reduced with the equalization process. The SETs’ original values are very similar to the level meter ones probably due to their more linear sensitivity curve.
It is possible to observe a general trend in the SETs (i.e., ADI, AEI, H, DSC, ZCR) which is probably caused by their already almost “flat” sensitivity curve. When the SETs’ recordings are equalized, a deterioration in their similarity to the sound level meter is observed and caused by trying to further linearize their nearly linear frequency response.
To statistically validate the procedure, the eco-acoustic indices calculated on the level meter’s urban park recordings were confronted with the ones obtained with the soundscape recorders. In particular, each index series calculated on the nine one-minute-long level meter’s recordings were compared to the one calculated on each of the nine one-minute-long soundscape recordings before and after the equalization. For this reason, the tests used were the pairwise Student’s test (when both series were normally distributed and homoscedastic) or Wilcoxon’s test [45]. The tests’ null hypothesis H0 assumes that the mean difference in the distributions is equal to 0; it is refused when p-value < 0.05.
Looking at Table 4, the AM presents indices that confirm H0 (BI) when equalized. SETs present the same number of indices that confirm H0 when they are not equalized and equalized using the 512 and 1024 filter order. Finally, the SMMs benefit the most from the equalization when it is performed using an order of 1024 FFT.

3.3. In-Field Monitoring Campaign Example (Ticino Park)

In this section, time trends of the eco-acoustic indices are shown (Figure 18). Given the results of the previous sections, the SMMs were equalized using a filter order of 1024, while the SET was not equalized (the AM was not used in this monitoring campaign).
In Figure 18, the indices’ time trends before (left column) and after (right column) the equalization are reported, and it is possible to affirm the following:
  • ACI: the time trends remain the same since the index is not affected by the process.
  • ADI: in the original time trends graph (C), the presence of the DC offset is noticeable which afflicts all the SMMs except Site 6 (blue curve); after the equalization, the SMMs present a similar trend.
  • AEI: The DC offset is also visible in these graphs. The effect of the equalization is noticeable since the time trends are more similar in the post-equalization graph (F).
  • BI: The SET (Site 9, black) is distant from the SMMs’ trends before the process (G) due to its flatter frequency response. After the equalization, the SMMs have a more similar trend to the SET.
  • NDSI: The procedure reduces the SMMs’ overestimated biophonic contribution to the soundscape and the underestimation of the anthropophonies thanks to the equalization of the frequency response; in fact, values change from being almost constant at +1 for Sites 1–6 (SMMs) to a more oscillatory trend, while the anthrophonic disturbance generated by the highway becomes more evident in Sites 7–8–9 (SMM, SMM and SET).
  • H: The DC offset is evident for this index just like the ADI and AEI (since it is not possible to define a low-frequency limit in these indices’ implementation in R). However, in the post-equalization graph (L), the SMMs’ trends are adjusted; given the results in the pocket park, these values may be overestimated.
  • DSC: In the pre-equalization graph (M), the bias affecting the SMMs due to their peaked frequency response is extremely evident. After the equalization (N), time trends are corrected since Sites 7–8 (SMMs) are very similar to Site 9 (SET) indicating the influence of the highway (higher intensities at low frequencies); moreover, the general DSC values are reduced from a mean value of 4 kHz to 2.5 kHz, underlining the bias which afflicts this index if not corrected.
  • ZCR: The DC offset is evident for this index. After the equalization (P), the values are corrected, and the difference between Sites 7–8–9 (nearer the highway) from the others is noticeable.

4. Discussion

The incongruence of eco-acoustic indices in analyzing the soundscape has been stressed by different studies, highlighting different trends [19,22] and limitations in predicting biodiversity parameters [21,47,48].
In particular, Alcocer and colleagues [47] showed the disparity in the number of studies published between 2007 and 2019 that found a correlation between eco-acoustic indices and biological parameters (i.e., avian richness) in favor of a lack of correlation between these two parameters. The absence of correlation was also confirmed by [21] through a meta-analysis involving four datasets from around the world; nevertheless, as the authors acknowledge in their paper, their findings may be explained with a non-harmonized data collection method in terms of the sampling design, bird survey method and recording devices [48].
Our study fits into this context, proposing an equalization protocol to mitigate the biases inherent in soundscape recordings and thus reducing the incongruence of eco-acoustic indices between studies. Nonetheless, the need for an accurate soundscape analysis is not recent, and other studies have covered this issue [24,25,26].
In [24], a calibration methodology is proposed, and, even if it is a different procedure than equalization, it highlights the need for more accurate soundscape analysis. Unfortunately, the sensitivity curves of the soundscape recorders are not linear; therefore, calibrating different devices will not obtain an equal output in terms of sound pressure levels and eco-acoustic indices’ values.
In [25], an equalization is performed on one-third octave bands between 25 Hz and 6.3 kHz using 1 s Leq values acquired with a type 1 sound level meter. This work lays a foundation for comparable recordings, but it is applied to a narrow frequency range (25–6300 Hz) using one-third octave bands, thus not covering the entire diversity of soundscape events (i.e., biophonies, anthropophonies and geophonies), and a single correction factor is used for a wide range of frequencies.
In our study, we performed an equalization procedure by taking as a reference the whole spectrum range using frequency ranges smaller than one-third octave bands thanks to the FFT analysis. Regarding this last point, and considering the validation stages conducted, we recommend using an order value of 1024 for equalizing the Song Meter Micro and 512 for the Audiomoth, while the Soundscape Explorer Terrestrial does not need to be corrected.
Finally, in [26], a pipeline to obtain similar eco-acoustic indices’ values between Audiomoth and Song Meter 4 (SM4) is proposed. The keystone of the process is the identification of comparable frequency ranges between the soundscape instruments (AM and SM4) on which the eco-acoustic indices are calculated. The approach led to a significant reduction of up to 70% in the variability observed in the calculation of eco-acoustic indices.
The selection of continuous comparable ranges suggested in [26] may not always be applicable since some devices may present oscillating sensitivity curves that do not allow for the selection of a continuous range and thus the implementation of eco-acoustic indices. Therefore, an equalization procedure is necessary to adjust the intensities throughout the entire spectrum allowing for the study of soundscapes in their integrity using virtually any recording device available on the market.

5. Conclusions

In this paper, we propose and validate an equalization procedure which we suggest implementing when studying soundscapes. In particular, it is extremely important when employing recorders characterized by a nonlinear sensitivity curve because it helps to reduce the biases that affect some eco-acoustic indices. Moreover, it reduces the errors when multiple brands of recorders are employed and when soundscapes from different habitats are compared.
Regarding the application of the procedure, we suggest applying it on devices that present a nonlinear frequency response. In our specific case study, we have proven that the equalization should be calibrated on device brands; moreover, each device should have its own equalization curve since there is a certain variance between devices of the same brand due to manufacturing processes. To calculate the equalization curve, the measurements of the reference signal should be carried out in a controlled environment, free-of-noise sources and with low reflectance (an anechoic chamber is the best choice). The MATLAB script of the procedure is available in the Supplementary Materials Section.
Future steps will involve its examination in different environments and the comparison between areas. Moreover, it will be applied to other soundscape sensors, and a new trial will be carried out using the add-on produced by Wildlife Acoustic to flatten the 6 kHz peak in the Song Meter Micro.
We hope this procedure will be adopted by the scientific community to enhance the understanding of the impacts of anthropic noise and the shaping of soundscapes, to help define guidelines and limits to effectively protect ecosystems and biodiversity.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/s24144642/s1.

Author Contributions

Conceptualization, A.P., R.B. and G.Z.; methodology, A.P., A.B., V.Z.-C., R.B. and G.Z.; software, A.P., J.M.F. and V.Z.-C.; validation, A.B., V.Z.-C., R.B. and G.Z.; formal analysis, A.P., V.Z.-C. and R.B.; investigation, A.P., A.B., V.Z.-C. and G.G.; resources, G.Z.; data curation, A.P.; writing—original draft preparation, A.P.; writing—review and editing, A.P., V.Z.-C., R.B., J.M.F., G.G. and G.Z.; visualization, A.P. and V.Z.-C.; supervision, G.Z.; project administration, G.Z.; funding acquisition, G.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

We would like to thank Professor Mazzarella Livio and Engineer Fumagalli Roberto (Politecnico di Milano) for letting us use the anechoic chamber of the Politecnico di Milano; it constituted a fundamental step in the production of this study. We would also like to thank Professor Guarnaccia Claudio (Università degli Studi di Salerno) for lending us his Soundscape Explorer Terrestrial and Audiomoths, allowing us to expand the devices under study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Erbe, C.; Thomas, J.A. History of Sound Recording and Analysis Equipment. In Exploring Animal Behavior Through Sound, 1st ed.; Springer Nature: Cham, Switzerland, 2022; Volume 1, pp. 1–36. [Google Scholar]
  2. Can, A.; Audubert, P.; Aumond, P.; Geisler, E.; Guiu, C.; Lorino, T.; Rossa, E. Framework for Urban Sound Assessment at the City Scale Based on Citizen Action, with the Smartphone Application NoiseCapture as a Lever for Participation. Noise Mapp. 2023, 10, 20220166. [Google Scholar] [CrossRef]
  3. Ventura, R.; Mallet, V.; Issarny, V. Assimilation of Mobile Phone Measurements for Noise Mapping of a Neighborhood. J. Acoust. Soc. Am. 2018, 144, 1279–1292. [Google Scholar] [CrossRef]
  4. Alsina-Pagès, R.M.; Ginovart-Panisello, G.J.; Freixes, M.; Radicchi, A. A Soundwalk in the Heart of Poblenou Superblock in Barcelona: Preliminary Study of the Acoustic Events. Noise Mapp. 2021, 8, 207–216. [Google Scholar] [CrossRef]
  5. Bonet-Solà, D.; Vidaña-Vila, E.; Alsina-Pagès, R.M. Prediction of the Acoustic Comfort of a Dwelling Based on Automatic Sound Event Detection. Noise Mapp. 2023, 10, 20220177. [Google Scholar] [CrossRef]
  6. Lee, H.P.; Lim, K.M.; Garg, S. A Case Study of Recording Soundwalk of Miyajima and Itsukushima Shrine Using Smartphone. Acoust. Aust. 2018, 46, 349–361. [Google Scholar] [CrossRef]
  7. Aletta, F.; Brambilla, G.; Maffei, L.; Masullo, M. Urban Soundscapes: Characterization of a Pedestrian Tourist Route in Sorrento (Italy). Urban Sci. 2017, 1, 4. [Google Scholar] [CrossRef]
  8. Bonet-Solà, D.; Vidaña-Vila, E.; Alsina-Pagès, R.M. Analysis and Acoustic Event Classification of Environmental Data Collected in a Citizen Science Project. Int. J. Environ. Res. Public Health 2023, 20, 3683. [Google Scholar] [CrossRef] [PubMed]
  9. Vidaña-Vila, E.; Navarro, J.; Stowell, D.; Alsina-Pagès, R.M. Multilabel Acoustic Event Classification Using Real-World Urban Data and Physical Redundancy of Sensors. Sensors 2021, 21, 7470. [Google Scholar] [CrossRef]
  10. Sugai, L.S.M.; Silva, T.S.F.; Ribeiro, J.W.; Llusia, D. Terrestrial Passive Acoustic Monitoring: Review and Perspectives. Bioscience 2019, 69, 5–11. [Google Scholar] [CrossRef]
  11. Erbe, C.; Thomas, J.A. (Eds.) Choosing Equipment for Animal Bioacoustic Research. In Exploring Animal Behavior through Sound, 1st ed.; Springer Nature: Cham, Switzerland, 2022; Volume 1, pp. 37–86. [Google Scholar]
  12. Gibb, R.; Browning, E.; Glover-Kapfer, P.; Jones, K.E. Emerging Opportunities and Challenges for Passive Acoustics in Ecological Assessment and Monitoring. Methods Ecol. Evol. 2019, 10, 169–185. [Google Scholar] [CrossRef]
  13. Benocci, R.; Brambilla, G.; Bisceglie, A.; Zambon, G. Eco-Acoustic Indices to Evaluate Soundscape Degradation Due to Human Intrusion. Sustainability 2020, 12, 10455. [Google Scholar] [CrossRef]
  14. Bradfer-Lawrence, T.; Gardner, N.; Bunnefeld, L.; Bunnefeld, N.; Willis, S.G.; Dent, D.H. Guidelines for the Use of Acoustic Indices in Environmental Research. Methods Ecol. Evol. 2019, 10, 1796–1807. [Google Scholar] [CrossRef]
  15. Sugai, L.S.M.; Desjonquères, C.; Silva, T.S.F.; Llusia, D. A Roadmap for Survey Designs in Terrestrial Acoustic Monitoring. Remote Sens. Ecol. Conserv. 2020, 6, 220–235. [Google Scholar] [CrossRef]
  16. Benocci, R.; Potenza, A.; Bisceglie, A.; Roman, H.E.; Zambon, G. Mapping of the Acoustic Environment at an Urban Park in the City Area of Milan, Italy, Using Very Low-Cost Sensors. Sensors 2022, 22, 3528. [Google Scholar] [CrossRef]
  17. Beason, R.D.; Riesch, R.; Koricheva, J. Investigating the Effects of Tree Species Diversity and Relative Density on Bird Species Richness with Acoustic Indices. Ecol. Indic. 2023, 154, 110652. [Google Scholar] [CrossRef]
  18. Retamosa Izaguirre, M.; Barrantes Madrigal, J. Soundscape Structure in Forests Surrounded by Protected and Productive Areas in Central Costa Rica. Front. Remote Sens. 2023, 4, 1051555. [Google Scholar] [CrossRef]
  19. Dröge, S.; Martin, D.A.; Andriafanomezantsoa, R.; Burivalova, Z.; Fulgence, T.R.; Osen, K.; Rakotomalala, E.; Schwab, D.; Wurz, A.; Richter, T.; et al. Listening to a Changing Landscape: Acoustic Indices Reflect Bird Species Richness and Plot-Scale Vegetation Structure across Different Land-Use Types in North-Eastern Madagascar. Ecol. Indic. 2021, 120, 106929. [Google Scholar] [CrossRef]
  20. Chen, Y.F.; Luo, Y.; Mammides, C.; Cao, K.F.; Zhu, S.; Goodale, E. The Relationship between Acoustic Indices, Elevation, and Vegetation, in a Forest Plot Network of Southern China. Ecol. Indic. 2021, 129, 107942. [Google Scholar] [CrossRef]
  21. Sethi, S.S.; Bick, A.; Ewers, R.M.; Klinck, H.; Ramesh, V.; Tuanmu, M.-N.; Coomes, D.A. Limits to the Accurate and Generalizable Use of Soundscapes to Monitor Biodiversity. Nat. Ecol. Evol. 2023, 7, 1373–1378. [Google Scholar] [CrossRef]
  22. Bradfer-Lawrence, T.; Bunnefeld, N.; Gardner, N.; Willis, S.G.; Dent, D.H. Rapid Assessment of Avian Species Richness and Abundance Using Acoustic Indices. Ecol. Indic. 2020, 115, 106400. [Google Scholar] [CrossRef]
  23. Ulloa, J.S.; Haupert, S.; Latorre, J.F.; Aubin, T.; Sueur, J. Scikit-Maad: An Open-Source and Modular Toolbox for Quantitative Soundscape Analysis in Python. Methods Ecol. Evol. 2021, 12, 2334–2340. [Google Scholar] [CrossRef]
  24. Merchant, N.D.; Fristrup, K.M.; Johnson, M.P.; Tyack, P.L.; Witt, M.J.; Blondel, P.; Parks, S.E. Measuring Acoustic Habitats. Methods Ecol. Evol. 2015, 6, 257–265. [Google Scholar] [CrossRef]
  25. Mennitt, D.J.; Fristrup, K.M. Obtaining Calibrated Sound Pressure Levels from Consumer Digital Audio Recorders. Appl. Acoust. 2012, 73, 1138–1145. [Google Scholar] [CrossRef]
  26. Luna-Naranjo, D.; Martínez, J.D.; Sánchez-Giraldo, C.; Daza, J.M.; López, J.D. Quantifying and Mitigating Recorder-Induced Variability in Ecological Acoustic Indices. bioRxiv 2023. [Google Scholar] [CrossRef]
  27. Audacity Team Audacity, Free Audio Editor and Recorder 2021. Available online: https://www.audacityteam.org/ (accessed on 28 June 2024).
  28. Zambon, G.; Potenza, A.; Confalonieri, C.; Bisceglie, A.; Canedoli, C.; Schioppa, E.P.; Benocci, R. Acoustic Monitoring to Evaluate the Effect of Anthropogenic Noise within a Park. In Proceedings of the Internoise 2022—51st International Congress and Exposition on Noise Control Engineering, Glasgow, UK, 21–24 August 2022. [Google Scholar]
  29. Benocci, R.; Potenza, A.; Bisceglie, A.; Confalonieri, C.; Canedoli, C.; Schioppa, E.P.; Zambon, G.; Roman, H.E. Towards an Environmental Sound Map at Parco Nord of Milan, Italy. In Proceedings of the Internoise 2022—51st International Congress and Exposition on Noise Control Engineering, Glasgow, UK, 21–24 August 2022. [Google Scholar]
  30. Cocroft, R.B.; Hamel, J.; Su, Q.; Gibson, J. Vibrational Playback Experiments: Challenges and Solutions. In Studying Vibrational Communication; Springer: Berlin/Heidelberg, Germany, 2014; pp. 249–274. [Google Scholar] [CrossRef]
  31. The MathWorks Inc. MATLAB Version: 9.14.0.2206163 (R2023a); The MathWorks Inc.: Natick, MA, USA, 2022; Available online: https://www.mathworks.com (accessed on 28 June 2024).
  32. Youngworth, R.N.; Gallagher, B.B.; Stamper, B.L. An Overview of Power Spectral Density (PSD) Calculations. In Proceedings of the Optical Manufacturing and Testing VI, San Diego, CA, USA, 18 August 2005; Volume 5869, pp. 206–216. [Google Scholar]
  33. Arva, M.-C.; Bizon, N.; Stanica, M.; Diaconescu, E. A Review of Different Estimation Methods of DC Offset Voltage For Periodic-Discrete Signals. In Proceedings of the 2019 11th International Conference on Electronics, Computers and Artificial Intelligence (ECAI), Pitesti, Romania, 27–29 June 2019; pp. 1–5. [Google Scholar]
  34. Smith, J.O. (Ed.) Frequency response analysis. In Introduction to Digital Filters: With Audio Applications; W3K Publishing: Stanford, CA, USA, 2007; pp. 149–174. [Google Scholar]
  35. Hao, Z.; Wang, C.; Sun, Z.; van den Bosch, C.K.; Zhao, D.; Sun, B.; Xu, X.; Bian, Q.; Bai, Z.; Wei, K.; et al. Soundscape Mapping for Spatial-Temporal Estimate on Bird Activities in Urban Forests. Urban For. Urban Green. 2021, 57, 126822. [Google Scholar] [CrossRef]
  36. Pieretti, N.; Farina, A.; Morri, D. A New Methodology to Infer the Singing Activity of an Avian Community: The Acoustic Complexity Index (ACI). Ecol. Indic. 2011, 11, 868–873. [Google Scholar] [CrossRef]
  37. Villanueva-Rivera, L.J.; Pijanowski, B.C.; Doucette, J.; Pekin, B. A Primer of Acoustic Analysis for Landscape Ecologists. Landsc. Ecol. 2011, 26, 1233–1246. [Google Scholar] [CrossRef]
  38. Boelman, N.T.; Asner, G.P.; Hart, P.J.; Martin, R.E. Multi-trophic invasion resistance in Ha-waii: Bioacoustics, field surveys, and airborne remote sensing. Ecol. Appl. 2007, 8, 2137–2144. [Google Scholar] [CrossRef]
  39. Kasten, E.P.; Gage, S.H.; Fox, J.; Joo, W. The Remote Environmental Assessment Laboratory’s Acoustic Library: An Archive for Studying Soundscape Ecology. Ecol. Inf. 2012, 12, 50–67. [Google Scholar] [CrossRef]
  40. Fairbrass, A.J.; Rennett, P.; Williams, C.; Titheridge, H.; Jones, K.E. Biases of Acoustic Indices Measuring Biodiversity in Urban Areas. Ecol. Indic. 2017, 83, 169–177. [Google Scholar] [CrossRef]
  41. Sueur, J.; Pavoine, S.; Hamerlynck, O.; Duvail, S. Rapid acoustic survey for biodiversity appraisal. PLoS ONE 2008, 3, e4065. [Google Scholar] [CrossRef]
  42. Quinn, C.A.; Burns, P.; Hakkenberg, C.R.; Salas, L.; Pasch, B.; Goetz, S.J.; Clark, M.L. Soundscape Components Inform Acoustic Index Patterns and Refine Estimates of Bird Species Richness. Front. Remote Sens. 2023, 4, 1156837. [Google Scholar] [CrossRef]
  43. Eldridge, A.; Guyot, P.; Moscoso, P.; Johnston, A.; Eyre-Walker, Y.; Peck, M. Sounding out Ecoacoustic Metrics: Avian Species Richness Is Predicted by Acoustic Indices in Temperate but Not Tropical Habitats. Ecol. Indic. 2018, 95, 939–952. [Google Scholar] [CrossRef]
  44. Brambilla, G.; Benocci, R.; Potenza, A.; Zambon, G. Stabilization Time of Running Equivalent Level LAeq for Urban Road Traffic Noise. Appl. Sci. 2023, 13, 207. [Google Scholar] [CrossRef]
  45. Kim Tae Kyun T Test as a Parametric Statistic. Korean J. Anesth. 2015, 68, 540–546. [CrossRef]
  46. Osborne, P.E.; Alvares-Sanches, T.; White, P.R. To Bag or Not to Bag? How AudioMoth-Based Passive Acoustic Monitoring Is Impacted by Protective Coverings. Sensors 2023, 23, 7287. [Google Scholar] [CrossRef]
  47. Alcocer, I.; Lima, H.; Sugai, L.S.M.; Llusia, D. Acoustic Indices as Proxies for Biodiversity: A Meta-Analysis. Biol. Rev. 2022, 97, 2209–2236. [Google Scholar] [CrossRef]
  48. Llusia, D. The Limits of Acoustic Indices. Nat. Ecol. Evol. 2024, 8, 606–607. [Google Scholar] [CrossRef]
Figure 1. Devices employed in this study and relative dimensions (from left to right: Song Meter Micro, Soundscape Explorer Terrestrial, Audiomoth and LD831-C).
Figure 1. Devices employed in this study and relative dimensions (from left to right: Song Meter Micro, Soundscape Explorer Terrestrial, Audiomoth and LD831-C).
Sensors 24 04642 g001
Figure 2. Sensitivity curve for Song Meter Micro. The sensitivity curve is referred to the whole signal transmission chain (provided by Wildlife Acoustic Support Team).
Figure 2. Sensitivity curve for Song Meter Micro. The sensitivity curve is referred to the whole signal transmission chain (provided by Wildlife Acoustic Support Team).
Sensors 24 04642 g002
Figure 3. Microphone sensitivity curve for Soundscape Explorer Terrestrial (Adapted from the Microphone technical data by PRIMO Co., Ltd., Tokyo, Japan, installed on the SET).
Figure 3. Microphone sensitivity curve for Soundscape Explorer Terrestrial (Adapted from the Microphone technical data by PRIMO Co., Ltd., Tokyo, Japan, installed on the SET).
Sensors 24 04642 g003
Figure 4. Microphone sensitivity curve for Audiomoth 1.2.0 (Adapted from RevSpace).
Figure 4. Microphone sensitivity curve for Audiomoth 1.2.0 (Adapted from RevSpace).
Sensors 24 04642 g004
Figure 5. White noise measurements in the anechoic chamber with a loudspeaker and LD831-C.
Figure 5. White noise measurements in the anechoic chamber with a loudspeaker and LD831-C.
Sensors 24 04642 g005
Figure 6. White noise mean frequency spectrum showing the original frequency response of the loudspeaker (blue) and equalized (orange) to obtain a white noise signal measured by LD831-C.
Figure 6. White noise mean frequency spectrum showing the original frequency response of the loudspeaker (blue) and equalized (orange) to obtain a white noise signal measured by LD831-C.
Sensors 24 04642 g006
Figure 7. A scheme illustrating the steps to calculate the equalization curves.
Figure 7. A scheme illustrating the steps to calculate the equalization curves.
Sensors 24 04642 g007
Figure 8. A scheme illustrating the equalization of an in-field recording.
Figure 8. A scheme illustrating the equalization of an in-field recording.
Sensors 24 04642 g008
Figure 9. The frequency response vector and angular frequency vector of the equalization curves.
Figure 9. The frequency response vector and angular frequency vector of the equalization curves.
Sensors 24 04642 g009
Figure 10. A scheme of the process explained in the following sub-sections.
Figure 10. A scheme of the process explained in the following sub-sections.
Sensors 24 04642 g010
Figure 11. Devices placed in the urban park at a single measurement point.
Figure 11. Devices placed in the urban park at a single measurement point.
Sensors 24 04642 g011
Figure 12. Monitoring scheme at the Ticino Regional Park (Bereguardo, PV, Italy). The SMMs are indicated as yellow circles while the SET as a violet triangle. Number 5 is missing due to malfunctioning. The main anthropic noise source (highway) is highlighted in red.
Figure 12. Monitoring scheme at the Ticino Regional Park (Bereguardo, PV, Italy). The SMMs are indicated as yellow circles while the SET as a violet triangle. Number 5 is missing due to malfunctioning. The main anthropic noise source (highway) is highlighted in red.
Sensors 24 04642 g012
Figure 13. The white noise mean frequency spectrums of the four devices, before and after the equalization procedure. (a) The mean frequency spectrum of the original white noise recordings; (b) the mean frequency spectrum of the equalized white noise recordings using an order of 512; (c) the mean frequency spectrum of the equalized white noise recordings using an order of 1024; (d) the mean frequency spectrum of the equalized white noise recordings using an order of 16,384.
Figure 13. The white noise mean frequency spectrums of the four devices, before and after the equalization procedure. (a) The mean frequency spectrum of the original white noise recordings; (b) the mean frequency spectrum of the equalized white noise recordings using an order of 512; (c) the mean frequency spectrum of the equalized white noise recordings using an order of 1024; (d) the mean frequency spectrum of the equalized white noise recordings using an order of 16,384.
Sensors 24 04642 g013
Figure 14. Barplots showing the effect of the equalization process on the amplitude RMSE and the percentage error of the eco-acoustic indices. Each graph represents a parameter ((A) median RSME, (B) median ACI percentage error, (C) median ADI percentage error, (D) median AEI percentage error, (E) median BI percentage error, (F) median NDSI percentage error, (G) median H percentage error, (H) median DSC percentage error, (I) median ZCR percentage error).The four devices (x-axis) and the three equalization options (colors) are reported.
Figure 14. Barplots showing the effect of the equalization process on the amplitude RMSE and the percentage error of the eco-acoustic indices. Each graph represents a parameter ((A) median RSME, (B) median ACI percentage error, (C) median ADI percentage error, (D) median AEI percentage error, (E) median BI percentage error, (F) median NDSI percentage error, (G) median H percentage error, (H) median DSC percentage error, (I) median ZCR percentage error).The four devices (x-axis) and the three equalization options (colors) are reported.
Sensors 24 04642 g014
Figure 15. Urban park mean frequency spectrums of the four devices, before and after the equalization procedure. One of the nine recordings is displayed. (a) The mean frequency spectrum of the original in-field recording; (b) the mean frequency spectrum of the equalized in-field recording using an order of 512; (c) the mean frequency spectrum of the equalized in-field recording using an order of 1024; (d) the mean frequency spectrum of the equalized in-field recording using an order of 16,384.
Figure 15. Urban park mean frequency spectrums of the four devices, before and after the equalization procedure. One of the nine recordings is displayed. (a) The mean frequency spectrum of the original in-field recording; (b) the mean frequency spectrum of the equalized in-field recording using an order of 512; (c) the mean frequency spectrum of the equalized in-field recording using an order of 1024; (d) the mean frequency spectrum of the equalized in-field recording using an order of 16,384.
Sensors 24 04642 g015
Figure 16. Boxplots of the amplitude RMSE (A) and the percentage errors on eco-acoustic indices (BI). Each graph represents a parameter and the four measurement settings (original—equalized using orders of 512, 1024 and 16,384) are reported.
Figure 16. Boxplots of the amplitude RMSE (A) and the percentage errors on eco-acoustic indices (BI). Each graph represents a parameter and the four measurement settings (original—equalized using orders of 512, 1024 and 16,384) are reported.
Sensors 24 04642 g016
Figure 17. Boxplots showing the eco-acoustic indices. (AH) Each graph represents an index and the four measurement settings (reference—original—equalized using orders of 512, 1024 and 16,384) are reported.
Figure 17. Boxplots showing the eco-acoustic indices. (AH) Each graph represents an index and the four measurement settings (reference—original—equalized using orders of 512, 1024 and 16,384) are reported.
Sensors 24 04642 g017
Figure 18. Time trend graphs of the eco-acoustic indices. Each row represents an index, on the left before the equalization and on the right after the equalization of the SMMs using a filter order of value 1024. Subfigures show: (A,B) time trend of ACI before-after the equalization, (C,D) time trend of ADI before-after the equalization, (E,F) time trend of AEI before-after the equalization, (G,H) time trend of BI before-after the equalization, (I,J) time trend of NDSI before-after the equalization, (K,L) time trend of H before-after the equalization, (M,N) time trend of DSC before-after the equalization, (O,P) time trend of ZCR before-after the equalization.
Figure 18. Time trend graphs of the eco-acoustic indices. Each row represents an index, on the left before the equalization and on the right after the equalization of the SMMs using a filter order of value 1024. Subfigures show: (A,B) time trend of ACI before-after the equalization, (C,D) time trend of ADI before-after the equalization, (E,F) time trend of AEI before-after the equalization, (G,H) time trend of BI before-after the equalization, (I,J) time trend of NDSI before-after the equalization, (K,L) time trend of H before-after the equalization, (M,N) time trend of DSC before-after the equalization, (O,P) time trend of ZCR before-after the equalization.
Sensors 24 04642 g018
Table 1. Parameters set for each device in the anechoic chamber.
Table 1. Parameters set for each device in the anechoic chamber.
Parameters831SMMSET 18 dBSET 20 dBAM
Sampling rate48 kHz48 kHz48 kHz48 kHz48 kHz
Amplitude gain+0 dB+18 dB+18 dB+20 dB+15 dB
Number of devices17212
Other characteristicsNo windproof cap With and without the waterproof case by Audiomoth
Table 2. The mean amplitude RMSE and total percentage error on the eco-acoustic indices calculated on white noise, specified for each device and status.
Table 2. The mean amplitude RMSE and total percentage error on the eco-acoustic indices calculated on white noise, specified for each device and status.
DeviceNoiseTypeMean Amplitude RMSE (dB)Total Percentage Error 1
AMWhiteOriginal11.393777.26
Eq n5122.30893.51
Eq n1k1.40596.30
Eq n16k0.1152.75
SET 18dBWhiteOriginal12.651567.79
Eq n5120.29104.25
Eq n1k0.1491.57
Eq n16k0.0221.81
SET 20 dBWhiteOriginal10.271540.98
Eq n5120.2378.12
Eq n1k0.1170.15
Eq n16k0.0142.86
SMMWhiteOriginal8.781445.55
Eq n5120.82158.96
Eq n1k0.4970.60
Eq n16k0.08777.18
1 The total percentage error is calculated for each device at each status by summing the errors of the eight indices and averaging over the type of device.
Table 3. The mean amplitude RMSE and total percentage error on the eco-acoustic indices, specified for each device and status, in the urban park monitoring.
Table 3. The mean amplitude RMSE and total percentage error on the eco-acoustic indices, specified for each device and status, in the urban park monitoring.
DeviceNoiseTypeMean Amplitude RMSE (dB)Total Percentage Error 1
AMUrban parkOriginal10.423676.39
Eq n5124.99500.41
Eq n1k5.30440.31
Eq n16k5.81729.19
SET 18 dBUrban parkOriginal13.89165.49
Eq n5123.94145.85
Eq n1k3.98170.94
Eq n16k4.00169.44
SET 20 dBUrban parkOriginal12.16146.55
Eq n5124.12144.50
Eq n1k4.14158.51
Eq n16k4.15173.74
SMMUrban parkOriginal9.951876.25
Eq n5124.81454.86
Eq n1k4.90374.68
Eq n16k5.07288.19
1 The total percentage error is calculated for each device at each status by summing the errors of the eight indices and averaging over the type of device.
Table 4. p-values derived from Student’s and Wilcoxon’s tests performed between the eco-acoustic indices’ values of the level meter and the ones of each soundscape sensor. When H0 is confirmed (p-value > 0.05), the cell is highlighted in green. Here, we report the most interesting results from four devices.
Table 4. p-values derived from Student’s and Wilcoxon’s tests performed between the eco-acoustic indices’ values of the level meter and the ones of each soundscape sensor. When H0 is confirmed (p-value > 0.05), the cell is highlighted in green. Here, we report the most interesting results from four devices.
Reference Device (Original)Soundscape DeviceSoundscape Device StatusACIADIAEIBINDSIDSCHZCR
831AM1Original3.91 × 10−33.91 × 10−33.91 × 10−31.93 × 10−81.74 × 10−93.91 × 10−33.91 × 10−33.91 × 10−3
Eq n5123.91 × 10−34.14 × 10−61.16 × 10−52.03 × 10−13.21 × 10−51.45 × 10−51.18 × 10−103.91 × 10−3
Eq n1k3.91 × 10−33.39 × 10−63.91 × 10−33.01 × 10−11.50 × 10−59.50 × 10−62.10 × 10−103.91 × 10−3
Eq n16k3.91 × 10−32.59 × 10−63.57 × 10−74.96 × 10−11.39 × 10−62.22 × 10−63.91 × 10−33.91 × 10−3
831SET1 *Original8.41 × 10−23.51 × 10−41.97 × 10−32.28 × 10−44.32 × 10−11.95 × 10−22.72 × 10−73.91 × 10−3
Eq n5123.10 × 10−11.30 × 10−53.91 × 10−34.75 × 10−51.83 × 10−29.10 × 10−18.16 × 10−98.90 × 10−7
Eq n1k2.26 × 10−11.52 × 10−73.91 × 10−34.67 × 10−54.73 × 10−25.70 × 10−11.98 × 10−113.91 × 10−3
Eq n16k3.75 × 10−23.91 × 10−34.64 × 10−64.41 × 10−54.31 × 10−24.26 × 10−14.81 × 10−133.91 × 10−3
831SET3 *Original6.47 × 10−17.38 × 10−56.26 × 10−42.27 × 10−61.65 × 10−33.76 × 10−31.03 × 10−63.91 × 10−3
Eq n5126.76 × 10−14.78 × 10−53.30 × 10−64.29 × 10−72.26 × 10−56.76 × 10−21.44 × 10−85.06 × 10−7
Eq n1k7.57 × 10−16.04 × 10−71.12 × 10−54.34 × 10−73.88 × 10−41.36 × 10−14.59 × 10−111.35 × 10−6
Eq n16k1.02 × 10−13.91 × 10−33.56 × 10−65.31 × 10−79.60 × 10−41.71 × 10−11.02 × 10−111.72 × 10−6
831SMM1393Original1.07 × 10−13.91 × 10−33.99 × 10−79.44 × 10−31.95 × 10−23.91 × 10−33.91 × 10−33.91 × 10−3
Eq n5129.68 × 10−27.38 × 10−51.59 × 10−41.38 × 10−53.74 × 10−11.74 × 10−14.91 × 10−113.91 × 10−3
Eq n1k8.98 × 10−21.37 × 10−45.89 × 10−41.87 × 10−54.68 × 10−31.79 × 10−19.57 × 10−113.91 × 10−3
Eq n16k5.58 × 10−21.84 × 10−39.25 × 10−31.78 × 10−57.68 × 10−53.54 × 10−56.53 × 10−103.91 × 10−3
831SMM1435Original5.65 × 10−21.55 × 10−79.64 × 10−73.91 × 10−33.91 × 10−33.91 × 10−33.91 × 10−33.91 × 10−3
Eq n5127.42 × 10−21.46 × 10−52.96 × 10−51.49 × 10−54.54 × 10−28.23 × 10−33.85 × 10−113.91 × 10−3
Eq n1k7.42 × 10−22.91 × 10−57.41 × 10−52.84 × 10−56.74 × 10−19.39 × 10−17.00 × 10−113.91 × 10−3
Eq n16k5.47 × 10−24.00 × 10−41.15 × 10−34.97 × 10−51.58 × 10−91.69 × 10−72.23 × 10−93.91 × 10−3
* SET 1 belongs to the SET 18 dB series; meanwhile, SET3 is the SET 20 dB.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Potenza, A.; Zaffaroni-Caorsi, V.; Benocci, R.; Guagliumi, G.; Fouani, J.M.; Bisceglie, A.; Zambon, G. Biases in Ecoacoustics Analysis: A Protocol to Equalize Audio Recorders. Sensors 2024, 24, 4642. https://doi.org/10.3390/s24144642

AMA Style

Potenza A, Zaffaroni-Caorsi V, Benocci R, Guagliumi G, Fouani JM, Bisceglie A, Zambon G. Biases in Ecoacoustics Analysis: A Protocol to Equalize Audio Recorders. Sensors. 2024; 24(14):4642. https://doi.org/10.3390/s24144642

Chicago/Turabian Style

Potenza, Andrea, Valentina Zaffaroni-Caorsi, Roberto Benocci, Giorgia Guagliumi, Jalal M. Fouani, Alessandro Bisceglie, and Giovanni Zambon. 2024. "Biases in Ecoacoustics Analysis: A Protocol to Equalize Audio Recorders" Sensors 24, no. 14: 4642. https://doi.org/10.3390/s24144642

APA Style

Potenza, A., Zaffaroni-Caorsi, V., Benocci, R., Guagliumi, G., Fouani, J. M., Bisceglie, A., & Zambon, G. (2024). Biases in Ecoacoustics Analysis: A Protocol to Equalize Audio Recorders. Sensors, 24(14), 4642. https://doi.org/10.3390/s24144642

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop