Next Article in Journal
3MSF: A Multi-Modal Adaptation of the 6TiSCH Minimal Scheduling Function for the Industrial IoT
Previous Article in Journal
Layout of Detection Array Based on Multi-Strategy Fusion Improved Adaptive Mayfly Algorithm in Bearing-Only Sensor Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Missing Value Imputation of Wireless Sensor Data for Environmental Monitoring

1
Department of Mathematics, University of Antwerp-imec, Middelheimlaan 1, 2000 Antwerp, Belgium
2
IDLab, Department of Computer Science, University of Antwerp-imec, Sint-Pietersvliet 7, 2000 Antwerp, Belgium
3
Plants and Ecosystems, Department of Biology, University of Antwerp, Universiteitsplein 1, 2610 Antwerp, Belgium
4
Geobiology, Department of Biology, University of Antwerp, Universiteitsplein 1, 2610 Antwerp, Belgium
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(8), 2416; https://doi.org/10.3390/s24082416
Submission received: 5 March 2024 / Revised: 4 April 2024 / Accepted: 8 April 2024 / Published: 10 April 2024

Abstract

:
Over the past few years, the scale of sensor networks has greatly expanded. This generates extended spatiotemporal datasets, which form a crucial information resource in numerous fields, ranging from sports and healthcare to environmental science and surveillance. Unfortunately, these datasets often contain missing values due to systematic or inadvertent sensor misoperation. This incompleteness hampers the subsequent data analysis, yet addressing these missing observations forms a challenging problem. This is especially the case when both the temporal correlation of timestamps within a single sensor and the spatial correlation between sensors are important. Here, we apply and evaluate 12 imputation methods to complete the missing values in a dataset originating from large-scale environmental monitoring. As part of a large citizen science project, IoT-based microclimate sensors were deployed for six months in 4400 gardens across the region of Flanders, generating 15-min recordings of temperature and soil moisture. Methods based on spatial recovery as well as time-based imputation were evaluated, including Spline Interpolation, MissForest, MICE, MCMC, M-RNN, BRITS, and others. The performance of these imputation methods was evaluated for different proportions of missing data (ranging from 10% to 50%), as well as a realistic missing value scenario. Techniques leveraging the spatial features of the data tend to outperform the time-based methods, with matrix completion techniques providing the best performance. Our results therefore provide a tool to maximize the benefit from costly, large-scale environmental monitoring efforts.

1. Introduction

Over the last decade, major advances in wireless communication technology, microelectronics, and (big) data analytics have caused a significant increase in the application of Wireless Sensor Networks (WSNs) [1,2]. A WSN comprises a network of many spatially distributed sensors that monitor certain parameters of a physical system and engage in wireless data communication. The WSN is made up of sensor nodes, sometimes also called sensor motes, which are essentially microcomputers with the ability to collect data, process these data internally, and finally transmit these data to a centralized location. WSNs have numerous applications in different fields, including environmental monitoring, health monitoring, logistics, and smart cities [3,4]. With the increasing use of WSNs, there is a growing demand for performant data analysis techniques capable of handling the vast volumes of collected data.
An important challenge within WSN research concerns missing value imputation for the extensive spatiotemporal datasets that are generated. Unavoidably, networks tend to lose readings from sensors for reasons that are difficult or impossible to anticipate, such as sensor failure due to power depletion, network outages, and communication errors, but also destruction due to storms or vandalism [5]. These missing readings can have important consequences for real-time monitoring, for example, in an emergency setting. Likewise, environmental monitoring applications relying on WSN data can suffer from missing data, which might lead to delayed or incorrect responses to environmental changes. Additionally, missing values can weaken the reliability of sensor data and increase the difficulty of sensor calibration. Finally, incomplete data can also compromise the performance of subsequent modeling and statistical analysis, which may result in biased conclusions or inaccurate predictions. A concrete example can be found in environmental research, where a WSN is commonly leveraged to measure variables such as temperature, humidity, atmosphere pressure, and sunlight, among others. Despite the wealth of data collected by sensor nodes, they often exist in raw form. Analytical tools commonly employed in such fields, such as support vector machines, principal component analysis, and singular value decomposition, face limitations when confronted with datasets containing missing data. Consequently, addressing the issue of missing data in these datasets presents a significant hurdle, impacting the efficacy of analyses and hindering the ability to draw meaningful conclusions [6].
The objective of this study was to evaluate the performance of missing value imputation techniques on a dataset generated by a WSN for environmental monitoring. To this end, we employ a unique dataset that originated from one of the largest citizen science projects to date involving Internet of Things (IoT) monitoring. Throughout the summers of 2021 and 2022, 4400 citizens within the region of Flanders (Belgium) installed IoT sensors in their gardens to measure the temperature and soil moisture at a high temporal frequency (every 15 min). The goal of this citizen science project, called Curieuze-Neuzen in de Tuin; Nosy Parkers in the Garden (CNidT), was to gain insight into how garden ecosystems can provide cooling for climate adaptation and mitigate the impacts of extreme weather events like heat waves. In projects like CNidT, missing values in the sensor time series are undesirable, both from a scientific and from a citizen perspective. From a scientific perspective, the data incompleteness reduces the power of the ensuing statistical analysis, which here aimed to uncover the factors that drive local garden cooling during extreme weather events. Likewise, data incompleteness was also highly unwanted from the citizen perspective: participating citizens were updated daily through personal dashboards, while society at large was informed through real-time maps on the website of a national newspaper. However, missing values were common in the recorded time series due to a combination of random sensor failure (e.g., battery problems), failed data transfers (e.g., due to network outages), and errors made by the citizens (e.g., destruction or damaging of the sensor). For these reasons, the dataset from the CNidT project was especially suitable as a case study for missing value imputation in WSN data. The CNidT dataset is an integral component of the SoilTemp project, which is a publicly available database outlined in Lembrechts et al. (2020) [7]. This extensive database comprises data from 7538 temperature sensors spanning 51 countries and encompassing diverse biomes. The primary objective of the SoilTemp project is to enhance the global comprehension of microclimates and to address discrepancies between existing climate data and the finer spatiotemporal resolutions pertinent to organisms and ecosystem dynamics [7].
Given that missing data within WSNs pose a fundamental challenge, the development of methods capable of imputing these missing values represents an active area of research. Within our study, several imputation approaches were evaluated to analyze their performance. An overview of all considered approaches and their imputation strategies is given in Table 1. A first approach involves techniques that take advantage of the temporal correlation between data, thus imputing missing values for a given sensor using the available data of that same sensor at different time steps. Evaluated methods for this approach include mean and linear spline imputation [8]. A second class of techniques utilizes spatial correlation to impute values, focusing on data from other sensors in the network at the same time step to impute the missing values of one sensor. Evaluated methods for this approach include k Nearest Neighbors (KNN) imputation [9], Multiple Imputation (MI) techniques such as Multiple Imputation using Chained Equations (MICE) and Markov Chain Monte Carlo (MCMC) [10,11,12], and Random Forests (RFs) to replace missing data (MissForest) [13]. The last strategy combines both the spatial and temporal aspects, taking full advantage of the patterns and intricacies present within the data. For this, specific methods for WSNs have been developed, such as Data Estimation using Statistical Models (DESM) and Applying k-Nearest Neighbor Estimation (AKE). Matrix Completion (MC) methods can also be exploited here as they use correlations within one sensor and across multiple sensors but assume that the data is static, i.e., they ignore the temporal component of the data [14,15]. Other methods in this class tend to leverage deep learning to impute missing values, for example Multiple Imputation using Denoising Autoencoders (MIDA) [16] or Recurrent Neural Network (RNN)-based approaches such as Bidirectional Recurrent Imputation for Time Series (BRITS) and Multi-directional Recurrent Neural Network (M-RNN) [15,17]. For a detailed explanation of all imputation methods evaluated in this study, we refer to Section 2.3.
Previous studies have conducted various comparative analyses, assessing different datasets, classes of algorithms, setups, and types or scenarios of missingness. However, in most studies, the focus is more on multivariate time series rather than on specific WSN data. Jadhav et al. [18] compared seven imputation methods across five publicly available datasets, concluding that KNN imputation exhibited the highest performance. Similarly, Jäger et al. [19] evaluated six imputation techniques on 69 datasets, noting that random forest-based solutions generally outperformed others. Notably, their study also evaluated performance in downstream Machine Learning (ML) tasks, finding that the imputation rendered a 10–20% performance increase. Khayati et al. [20] focused on sensor time series imputation, comparing 16 recovery algorithms on six public and two synthetic datasets, including block missings, which are more reflective of WSN data characteristics. Their findings suggested that the optimal recovery method often depends on dataset-specific characteristics. Yozgatligil et al. [21] assessed six imputation techniques using Turkish State Meteorological Service data, introducing the correlation dimension technique to account for spatiotemporal dependencies in imputation evaluation. Their study indicated that the MCMC approach yielded the most favorable results.
Table 1. The imputation techniques that were considered in this study, together with their respective imputation strategy.
Table 1. The imputation techniques that were considered in this study, together with their respective imputation strategy.
MethodImputation Strategy
AKE [6]WSN-specific
BRITS [17]Deep learning
DESM [22]WSN-specific
KNN [9]Spatial correlations
MC [14]Temporal and spatial correlations (static)
MCMC [12]Spatial correlations
MICE [11]Spatial correlations
MIDA [16]Deep learning
MRNN [15]Deep learning
Mean imputationTemporal correlations
MissForest [13]Spatial correlations
Spline [8]Temporal correlations
In this study, we evaluated 12 imputation techniques for different artificial missing scenarios (by inducing 10%, 20%, 30%, 40%, and 50% data removal), as well as a more realistic scenario defined as “masked” missings. In this scenario, we replicated the missing patterns observed in sensors with incomplete data onto sensors with complete information, simulating a real-world missings scenario. In this way, we created a standardized scenario through which we can evaluate how effective every method is in compensating for the missing patterns. Comparisons are made based on the Root-Mean-Square Error (RMSE) and Mean Absolute Error (MAE) to assess the accuracy of the imputed values. Our study advances the existing literature by conducting a comprehensive comparison of various missing value imputation methods, employing different strategies and model types. Moreover, we analyze a genuine WSN dataset featuring a substantial sensor count (1500) and expand the assessment of these techniques from random missing values to masked missing values, offering a more realistic evaluation scenario for practical deployment.
The remainder of this paper is structured as follows: Section 2.1 introduces the CNidT project, while Section 2.2 describes the dataset collected in the project, as well as the preprocessing steps that were used. In Section 2.3, the imputation methods evaluated in this study are described, and the evaluation criteria are detailed in Section 2.4. Section 3 presents the results and discusses the implications of these results. In Section 4, we summarize our findings, list the most important insights and conclusions, and provide possible directions for future research.

2. Materials and Methods

2.1. The Curieuze-Neuzen Citizen Science Project

The dataset analyzed in this study originated from the citizen science project “Curieuze-Neuzen in de Tuin” (CNidT), which translates as “Nosy Parkers in the Garden” [23,24]. The project engaged 4400 citizen participants across the strongly urbanized and densely populated region of Flanders (Belgium, Northwestern Europe) to monitor the microclimate in their garden. The scientific objective was to quantitatively assess the impact of gardens on the local microclimate and their potential role in the mitigation of extreme weather events [25]. Initially, 50,578 citizens registered as candidates to participate in the project. From this pool of registrations, 4400 sampling locations were selected using an environmental sampling algorithm, to obtain a representative subsample that covered the range of gardens in terms of size and composition but also ensured a suitable geographical distribution across the measurement domain [26]. To this end, metadata were collected for each garden, including variables related to urbanity, garden characteristics, garden management, topography, and geography. These metadata were obtained from participants through questionnaires combined with available remote sensing data. Factor Analysis of Mixed Data (FAMD) was implemented to reduce the dimensionality of available metadata, and the selection algorithm used the first three Principal Component (PC) from FAMD coordinates to hierarchically select candidate locations, maximizing variation between gardens in the available environmental space.
The project included two six-month measurement campaigns, spanning the growing season (spring and summer) of 2021 and 2022. Citizens received a microclimate sensor device (a “lawn dagger”) that was inserted in the soil in the middle of a lawn patch in the garden. The microclimate sensor was a custom-tailored modified version of the well-established TMS logger, which provides a robust and cost-effective device to monitor temperature and soil moisture near the soil surface [27]. The TMS sensors recorded data every 15 min with three temperature sensors (DS7505 digital thermometer), one positioned at 10 cm below the soil surface, one at the soil surface, and one 12 cm above it. In addition, the TMS measures soil moisture using the time domain transmission principle in the top 15 cm of the soil [27]. The device has a large internal storage, allowing it to internally store the data collected over a period up to 10 years.
While regular TMS sensors only allow off-line data collection, a novel TMS-NB version of the instrument was specifically developed for the project, which was equipped with wireless transmission ability (collaboration between University of Antwerp, sensor development company TOMST, and telecom operator Orange Belgium). To this end, the TMS-NB was equipped with a data transfer module (BG77 Quectel with Qualcomm chipset) to send small data packages via the Narrowband Internet of Things (NB-IoT) network hosted by Orange Belgium. Measurements of temperature and soil moisture were recorded every 15 min, and the recorded data were stored in the device’s internal memory. Data collected over one day were sent as one data package via NB-IoT each day at midnight. This data package included additional metrics (e.g., battery status, signal quality) and was transmitted via the LiveObjects platform of telecom operator Orange to a relational database (MS SQL) at the University of Antwerp.
The CNidT project thus gave rise to a large WSN (>4000 nodes) that performed NB-IoT-based environmental monitoring for a period of two summers (April–September). The use of low-cost sensors and reliance on citizen input occasionally led to erroneous values or missing data points due to various factors, including random sensor malfunctioning (e.g., occasional missing data), connectivity issues (i.e., data package not sent over the NB-IoT network), as well errors and accidents by the participants (e.g., sensors damaged by kid’s play or robot lawn mowers). As sensor malfunctioning was virtually absent (<0.01% of data points) and connectivity issues could be solved by reading out the data manually after the end of the project, overall data availability ended up around 90% (see Section 2.2). Although such a percentage might be sufficient for most scientific questions, the project’s goal of reporting back to individual citizens in real time about conditions in their own garden, as well as a subsequent analysis using ML methods, makes a gap-filling exercise especially appealing.

2.2. Dataset and Preprocessing

A subset of the available data was used to evaluate the missing value imputation techniques. Sensor readings were retained for one six-month measurement campaign (starting on 12 April 2021, at 00:00:00 until 30 September 2021, at 23:45:00), measured every 15 min, thus providing a maximum of 16,512 records in each sensor time series (172 days of data collection times 96 readings per day). Before the final construction of the dataset, extensive manual data recovery measurements were done from the TMS-NB sensor to have as complete a dataset as possible. Furthermore, the additional metrics (e.g., battery status, signal quality) from the processed signal were also analyzed to generate as complete a sample as possible. Each sensor reading contained four data records (three temperatures at different heights and soil moisture). Data from the temperature sensor at 12 cm above the soil surface (expressed in degrees Celsius) were selected for the evaluation (and thus the remainder of our analysis), as these readings expressed the highest variability. Data series were available for 4163 sensor locations. Figure 1b illustrates the location of the sensors and whether the recorded time series for the sensor was complete or not. In total, 2978 sensors (or 71.5% of the WSN) had no missing values. Across all 4163 sensors comprising the WSN, 7.8% of records were missing. Although the missing percentage is not extremely high, about one-third of the sensors showed at least a few missing values, with some sensors missing nearly all values.
The geographical coordinates of the sensor location (uncertainty 10 m) are part of the metadata. As some methods utilize geometric distances between sensors, we calculated the haversine distance (as described by [28]) for every sensor combination as follows:
d h a v ( x , y ) : = 2 r arcsin sin 2 y l a t x l a t 2 + cos ( x l a t ) cos ( y l a t ) sin 2 y l o n x l o n 2
where x and y are the coordinates of two different sensors, and r is the radius of the Earth (6371 km). The haversine (or great circle) distance is the angular distance between two points on the surface of a sphere. We used the haversine distance rather than the Euclidean distance to account for the Earth’s curvature given the scale of the measurement domain (Flanders region; ∼300 km).

2.3. Missing Value Imputation

In the literature, various methodologies for imputing missing (sensor) data exist. Below, we provide a concise overview of the methods that were considered in this study. For a more detailed explanation, we refer to their respective papers and code implementations. First, we introduce different types of missing values, after which we detail how to create suitable test datasets for imputation.

2.3.1. Different Types of Missing Values

There are three types of missing data mechanisms: Missing Completely at Random (MCAR), Missing at Random (MAR), and Missing not at Random (MNAR) [29,30,31]. MCAR implies that missingness is independent of observed and unobserved data, making observed data still representative. This assumption is often strong and unrealistic. MAR means missingness is linked to observed but not unobserved data [29]. MNAR occurs when missingness is related to unobserved variables, making it the most challenging scenario to handle and non-ignorable [32].

2.3.2. General Approach

To test the performance of imputation methods, we restricted ourselves to data series that had complete information, as done in other comparison studies [15,33]. From the 2978 sensors available with complete records, we selected a subset of 1500 as our basic dataset. In these complete data series, we artificially introduced data gaps that had to be resolved by imputation. Two separate approaches for missing value creation were applied.
In a first approach, we applied patterns of randomly missing values with increasing fractions of data missing (10%, 20%, 30%, 40%, and 50%) using the numpy.random.choice function [34]. This approach provides missing values that are MCAR, which is the missingness pattern that is most often used in the literature on missing value imputation [19]. Although it would also be possible to introduce MAR missing values, e.g., based on sensor location, we did not consider this option, as it is included in the masked missing scenario introduced below. Finally, MNAR missing values would manifest themselves by removing temperature values based on the actual temperature values themselves (e.g., low temperature values are removed). As our sensor operates within a range of −55 °C to 125 °C [27], but this scenario was not relevant for our dataset.
In reality, missing data patterns are not necessarily random. In order to mimic a more realistic case, we took advantage of the missing data patterns from the sensors that actually had missing data. To this end, missing patterns were imposed from incomplete sensors onto complete sensors, thus imposing a mask with missing data. In this approach, time points for which an incomplete sensor was missing data were imposed as a mask onto a sensor with a complete data series (as illustrated in Figure 2). This approach is further referred to as the masked missing. This approach allows for a more realistic evaluation, as potential issues such as spatial or temporal block missing, for example, due to network failure or sensor failure, will be present in the data.
For all six scenarios (10%, 20%, 30%, 40%, and 50% random missing as well as the masked missing), a 2 × 5 nested Cross-Validation (CV) was run to obtain robust results. Nested CV involves two levels of CV loops: an outer loop and an inner loop. In the outer loop, the dataset is divided into training and testing sets using k-fold CV. Each fold of the outer loop trains the model on the training set and evaluates it on the testing set. Within each fold of the outer loop, an inner CV loop is employed where the training data are split into training and validation sets, also using k-fold CV. The inner loop is responsible for selecting the set of hyperparameters that performs best on the validation set. In our study, we used a 2 × 5 nested CV, i.e., we had two outer loops and five inner loops. For hyperparameter tuning, where applicable, we utilized a randomized grid search strategy. This involved exploring a predefined range of hyperparameters, as listed in Table A1, and selecting the combination that minimized the Root-Mean-Square Error (RMSE). To ensure comprehensive exploration of the hyperparameter space, we conducted tuning across 50 different hyperparameter combinations. Notably, this process was carried out each time on a randomly selected subset of 500 sensors from our dataset to ensure computational efficiency while maintaining representativeness. To facilitate transparency and reproducibility, we have included the best-performing hyperparameters used in our study, which are presented in Table A1.
In this section, we use the following notation, where a given test dataset X is defined as follows:
X : = { X n t } , n { 1 , , N } ; t { 1 , , T }
with N being the number of sensors and T the number of time points. The imputed value for sensor n at time point t is denoted by X ^ n t . A test dataset is schematically depicted in Figure 3. The implementation of all evaluated methods was done in Python, with the list of packages included in Table A6. We conducted our experiments on a Lenovo Thinkpad T495 2019 model with 2.3 GHz AMD Ryzen 7 PRO 3700U CPU consisting of four cores, with 4 MB of L3 cache and 16 GB of RAM.

2.3.3. Mean Imputation

In this approach, missing values are filled using the arithmetic mean. Specifically, within our application, we replace missing data points from a particular sensor with the mean value of that sensor across all available time steps. Mathematically, the imputed value is expressed as shown in Equation (3), where V represents the set of time points with available observations for the given sensor as follows:
X ^ n t = 1 | V | v V X n v .
Mean imputation is effective when there is limited temporal and spatial variability and when the number of missing observations for a sensor is relatively low. Because of its straightforward approach, it serves as a baseline method for comparison within our study.

2.3.4. Spline Imputation

The linear spline imputation method uses temporal correlation within one sensor to impute missing values [8]. An imputed value X ^ n t for sensor n is estimated at time t by applying a linear interpolation based on the closest available time point in both directions, t and t + :
X ^ n t = X n t + X n t + X n t t + t ( t t ) .

2.3.5. K Nearest Neighbor (KNN) Imputation

The KNN method was originally developed to estimate missing values in gene expression microarray experiments, but it can be easily applied to other use cases [9]. During imputation, data points with similar features as the data point with missing values are selected. In our case, the data points are the different sensors, while the features are the values at different time points. Thus, the KNN imputation technique leverages the spatial correlation of the dataset. This method would find k sensors that have a value present for the missing time point, where the values of the other time points are most similar to those of the sensor with the missing value. Afterward, a weighted average of the k “closest” sensors is calculated to estimate the missing value. The contribution of each of the k sensors is weighted by its similarity to the features of the sensor with missing values, where the similarity is quantified using the Euclidean distance.

2.3.6. Multivariate Imputation by Chained Equations (MICE)

The previously introduced methods all involve replacing missing values with a single estimation, disregarding the uncertainty and variability of the missingness. MI is a statistical technique used to handle missing data that generates multiple plausible imputations based on the distribution of the observed data [10,11]. Estimating multiple imputations, as opposed to just one imputation, accounts for (part of) the statistical uncertainty in the imputations [35]. MICE is an example of an MI technique and generally operates under the assumption that the missing data are MAR or MCAR [11,35,36,37]. When the data are not MAR, the application of MICE could result in biased or inaccurate estimates. The chained equations process used in MICE consists of the following steps [35,37]:
  • Make an initial guess about the missing values using a simple imputation method, such as mean imputation.
  • Set the missing values for one feature f back to missing. The observed values for f are then regressed using (all) other features in the dataset.
  • Make a prediction for the missing values of f using the regression model from the previous step.
  • Repeat steps 2 and 3 for all features that contain missing values. At the end of this step, all features with missing values have been imputed.
  • Repeat steps 2, 3, and 4 for a number of cycles and update the imputations in each cycle.
The number of cycles can be chosen by the user and is task-dependent. The final imputation is the imputation found in the final cycle [35].

2.3.7. Markov Chain Monte Carlo (MCMC) Imputation

Another MI technique is MCMC [12], based on the Bayesian framework. In essence, MCMC leverages the principles of a Markov Chain and Monte Carlo simulation to approximate missing values by iteratively sampling from a probability distribution. The main focus is finding the desired posterior distribution defined by a set of parameters θ , from which the unobserved values X u can be predicted using the conditional density of the observed observations X o  [12,38]. The method starts from an initial, plausible approximation of the missing readings X u . In the next step, the MI technique starts. Given certain parametric assumptions, the θ i can be estimated from the posterior distribution f ( θ i | X o , X u ) , with θ i being the estimated parameter values in step i. In a second step, the predictive distribution can be used to obtain the improved predicted values X u i at iteration i.
X u i + 1 f ( X u | θ i , X o )
In the next step, the θ parameter values can again be estimated from the complete data posterior distribution using the newly acquired values.
θ i + 1 f ( θ | X o , X u i + 1 )
These last two steps are iteratively executed until gradually converging to the true distribution. Due to the sequential sampling from two distributions, a Markov Chain is made, and the use of simulations renders the MCMC name [12,38,39]. MCMC imputation offers several advantages, including the ability to handle complex data structures and missingness patterns, as well as the flexibility of incorporating prior knowledge or constraints. The technique tends to be computationally intensive for large datasets, and it often requires careful tuning.

2.3.8. Matrix Completion (MC) Imputation

Another imputation technique based on iterative MI is matrix completion, based on [14]. The method uses the spatial and temporal correlations of the data to impute missing values. The main idea of MC is to handle missing values in a data matrix by imputing them with estimates based on the observed values and the low-rank structure of the data matrix. More precisely, missing readings get replaced iteratively with those obtained from a soft-thresholded singular value decomposition [14]. First, a singular value decomposition is applied to the incomplete matrix with soft-thresholding, where the nuclear norm of the matrix is used as a regularizer. In the next step, the modified singular value matrices are used to reconstruct the data matrix. Then, these two steps are iterated until convergence of the imputed values is reached. The matrix completion technique as discussed by [14] is well-suited for situations where the data matrix can easily be approximated by a lower-rank matrix, rendering an effective solution to the missing value problem for large and sparse matrices.

2.3.9. Data Estimation Using Statistical Model (DESM) Imputation

Similar to previously discussed methods, DESM uses temporal and spatial correlations between sensors to impute missing values [22]. The method is specifically developed for WSNs, with the sensor data specific characteristics in mind. More specifically, DESM relies on historical values of the sensor for which a value is missing (sensor n), as well as the values of the sensor spatially located the closest (sensor m), with the requirement that the latter sensor does not have missing values around the time point that needs to be imputed. Missing values are then estimated according to
X ^ n t = ( 1 α ) Y ^ + ( α ) Z ^ .
In Equation (7), Y ^ is the imputed value at the previous time point X ^ n ( t 1 ) , and Z ^ is defined as follows:
Z ^ = X n ( t 1 ) 1 + X m t X m ( t 1 ) X m ( t 1 ) .
DESM leverages both the temporal influence of one sensor on itself, included in Y ^ , as well as the spatial attributes of the other sensors (in this case m), represented by Z ^ , to impute the missing readings. The α in Equation (7) is the Pearson correlation coefficient between two sensors X n and X m , which serves as a weight parameter that evaluates the effects of Y ^ and Z ^ on the estimated value. Equation (8) is based on the assumption that the data collected by sensors X n and X m are approximately similar, as they are spatially close to each other.

2.3.10. Applying k Nearest Neighbor Estimation (AKE) Imputation

Sensors that are located in close spatial proximity to other sensors will yield very similar measurements, which means that it is possible to impute missing values based on the neighboring sensor values for the same time point. As the exact functional relationship between two nearby sensors is unknown, the AKE method assumes that this relationship can be approximated linearly in a short time period [6]. Under this assumption, we can estimate missing values X n t m from a neighbor sensor X m t for any time t using linear regression
X ^ n t m = α + β X m t ,
where α and β are estimated using all non-missing ( X n t , X m t ) pairs. In total, k linear regression models will be fitted for every sensor, where k is a tunable hyperparameter. To obtain an imputed value, we have to combine the estimations from all k neighboring sensors. While using the arithmetic mean of all imputations is a valid option, this would disregard the strength of the linear correlation between two sensors. For this reason, AKE uses a weighted average of all k estimated values
X ^ n t = m = 1 k w n m · X ^ n t m ,
where w n m is the weight, for which 0 w n m 1 and m = 1 k w n m = 1 . As we can assess the performance of a linear regression by using the determination coefficient r 2 , we define the weight w n m as the normalized determination coefficient as follows:
w n m = r n m 2 j = 1 k r n j 2 .

2.3.11. MissForest Imputation

The MissForest method is a non-parametric imputation method that uses an RF to impute missing values [13]. To start, an initial guess is made for the missing values, using mean imputation or another imputation method. Afterward, the RF is trained on all data, including the initial guess for the missing values. The trained RF is then used to impute the missing values again. This procedure is repeated for a fixed number of iterations or until a stopping criterion is reached, whichever comes first. The stopping criterion is met as soon as the difference between the previously imputed values ( X ^ o l d ) and the newly imputed values ( X ^ n e w ) increases for the first time, i.e., the imputation has converged. The difference Δ is defined as follows, with p being the total number of missing values:
Δ = i = 1 p ( X ^ i n e w X ^ i o l d ) 2 i = 1 p ( X ^ i n e w ) 2 .
Due to the use of random forests, the method is relatively robust against outliers; however, it can become computationally expensive on large datasets.

2.3.12. Multiple Imputation Using Denoising Autoencoders (MIDA)

MIDA is another MI technique that uses overcomplete Denoising Autoencoders (DAEs) to impute missing values [16]. An overcomplete DAE is a DAE where the input data are projected to a higher-dimensional subspace, from which the missing values are recovered. The input layer has T nodes, assuming the data have T features (time points in our use case). Then, each successive hidden layer adds Θ nodes, where Θ is a tunable hyperparameter. This is done for j encoding layers, after which j 1 decoding layers are added, which decrease the number of nodes from T + j Θ to T for the output layer. Empirically, Θ = 7 and j = 3 have been found to be a good choice, but both of these parameters can be seen as tunable hyperparameters [16]. The MI part of MIDA is established by initializing the model with a different set of random weights in multiple runs, thereby providing multiple predictions. By leveraging the representational learning capabilities of denoising autoencoders, MIDA can capture the underlying patterns in the data, thus potentially generating more realistic results. However, these results heavily depend on the quality and similarity between the unknown and known observations in the training data. Furthermore, the use of autoencoders and MI also make this method computationally expensive.

2.3.13. Bidirectional Recurrent Imputation for Time Series (BRITS)

There are several methods based on an RNN for missing value imputation, such as BRITS [17]. The imputation method tries to learn the missing values in a bidirectional recurrent dynamical system, without any specific assumptions [17]. The method was originally developed for missing value imputation in multiple correlated time series, which we extend to the WSN framework. In BRITS, an RNN is used directly for predicting missing values, meaning that missing values are regarded as variables of the bidirectional RNN graph, leveraged in the back-propagation of the neural network. This approach ensures that missing values receive delayed gradient updates in both the forward and backward directions, with consistency constraints. The model architecture can also be leveraged for simultaneous regression or classification jointly in one graph, rather than pure missing imputation. This can mitigate a part of the error propagation in subsequent modeling tasks. The main advantages of the BRITS imputation method is the application to general missing settings, as well as the ability to handle correlated time series and nonlinear dynamics within the data.

2.3.14. Multi-Directional Recurrent Neural Network (M-RNN) Imputation

Closely related to BRITS is M-RNN, as introduced by [15]. M-RNN imputes values both within and across data-streams, thus both in a temporal and spatial fashion. The original method was developed for clinical applications, yet it can be easily applied to other scenarios. The imputation technique contains both an interpolation block (temporal) and an imputation block (spatial), which are trained simultaneously. The interpolation block uses an adjusted bi-directional RNN with a lagged timing for the inputs into the hidden layers in the forward direction and advanced in the backward direction [15]. The imputation block is then a fully connected neural network with dropout. Similarly to BRITS, the method can also be used for a subsequent modeling task directly. Notably, the M-RNN tends to be less affected by both the quantity and specific nature of missing data.

2.4. Empirical Evaluation

To compare the performance of the different methods, it is important to have a predetermined set of performance metrics. In this study, we use the RMSE and MAE to assess the accuracy of the imputed values. Based on previous studies, we also evaluate the percentage of cases in which a missing value can be estimated, i.e., the Prediction Coverage Error (PCE), as defined in Equation (13) [40,41].
PCE = number of successful imputations number of missings × 100 %
The PCE is necessary to be able to interpret the RMSE and MAE fairly; indeed, the RMSE and MAE ignore missing data points that the model was unable to impute. As a result, the RMSE and MAE might be underestimated for some methods that are not able to impute all samples, resulting in an overstatement of the model performance. For these cases, the PCE provides additional context.

3. Results and Discussion

3.1. Random Missings

The performance of the various imputation methods, evaluated using the metrics defined in Section 2.4, is depicted in Figure 4 and Figure 5. Exact values for the RMSE, MAE, PCE, and execution times are detailed in Table A2, Table A3, Table A4 and Table A5. As can be seen in Figure 4a,b and Table A2 and Table A3 the MC imputation method achieves the best performance (smallest RMSE and MAE) for all degrees of missingness. Generally, all methods consistently outperform the baseline mean imputation, except for the MIDA method, where performance diverges for higher degrees of missingness (see Section 3.5). In the results, a noticeable trend is observed where methods considering the spatial features of the data generally outperform others. Methods such as MissForest and MCMC obtain a good performance and even outperform DESM, which is specifically tailored for WSNs problems. AKE, another WSN-specific method, has a very good performance and is only outperformed by MC and MICE. For nearly all methods, the performance gets worse with increasing degrees of missingness, which is expected and is also commonly observed in the literature [15,41]. A notable exception to this rule is M-RNN, as its performance remains stable with increasing degrees of missingness. Although unexpected, we consider this result to be less significant, as the difference is relatively small, and the absolute performance of M-RNN is among the worst for our specific use case. In addition, M-RNN can efficiently handle higher degrees of missingness, explaining the result.

3.2. Masked Missings

As random missings do not accurately represent real-life missing scenarios, we also evaluated all methods on a realistic missings dataset obtained by creating masks from real, observed missing patterns. From Figure 5a,b and Table A2 and Table A3, we can conclude that the MC method performs best, as was also the case in the random missing scenario. Similarly, AKE, DESM, MCMC, and MissForest are again among the top performing methods, indicating the better performance of spatial methods on our dataset. As expected, the performance in the masked missings scenario is generally worse than for random missings, as so-called “block” missings frequently cause a lack of “nearby” data points, which are often used to impute missing values. In particular, the performance of spline imputation and MICE is significantly worse in the masked scenario. This finding was expected for spline imputation, as it applies a linear interpolation based on the closest time points surrounding the missing value, which will often be missing itself in the case of block missings. For MICE, this can be attributed to a worse convergence of the model due to a poor initial guess. More specifically, the initial guess for the MICE algorithm was made using mean imputation. From Table A2 and Table A3, we can clearly see that mean imputation performs significantly worse for masked missings when compared to its performance for random missings and thus probably did not provide an accurate initial guess.

3.3. Prediction Coverage Error

We should, however, interpret the results for AKE and DESM in the masked missings scenario with caution. In fact, Figure 6a and Table A4 show that these methods do not achieve a PCE of 1, meaning that they were unable to impute all missing values. Indeed, AKE and DESM were only able to impute 95.3% and 98.5% of the missing values, respectively. As AKE requires at least one sensor (in a group of k nearby sensors) to have an observed value in the given time step and DESM requires one sensor to have no missing values around the considered time step, this is likely attributable to large-scale network outages. Indeed, these outages result in spatially nearby points failing simultaneously. Although the number of unfilled missings is quite small, an end user might opt to address the remaining gaps by employing a straightforward imputation method such as mean imputation.

3.4. Execution Time

A final criterion to consider when evaluating the imputation methods is the time needed to train the method (if necessary) and make the imputations, which is combined in the execution time. In Figure 6b and Table A5, the execution time averaged over all degrees of missingness and masked missings is shown. As expected, the deep learning-based methods (M-RNN, MIDA, and BRITS) were very computationally intensive to train. The MCMC method was also very computationally expensive, which was expected, as the authors state that this is the case for large datasets [12]. Finally, even though the MC method is specifically designed for large matrices, it is interesting to note that it achieved a very small execution time while also providing the best imputed values in both missing scenarios.

3.5. Discussion

Within our empirical evaluation, spatial methods tend to outperform others across all imputation scenarios. This can be attributed to the large number of sensors as well as the spatial proximity of the sensors. Furthermore, the evaluated time steps were limited, thus impacting the performance. In general, these results are in line with the literature. However, the deep learning-based methods exhibited poor performance on our dataset, regardless of the missing pattern. We attribute this to several factors. First, the relatively short time series considered in these data may limit the ability of certain deep learning-based methods, such as BRITS, to accurately capture temporal dependencies and patterns. Additionally, most deep learning applications tend to evaluate datasets with either a higher frequency of measurements or a higher number of measured objects or sensors. Furthermore, our dataset considers the behavior of small microclimates, in which local effects may cause temperature peaks that do not appear for other sensors, which complicates generalization to other sensors. Nevertheless, deep learning-based methods do exhibit good results for other datasets and should therefore not be ignored as a possible imputation technique [15,16,17]. They also offer the added advantage of making supplementary predictions using the acquired structure [15,17] or by training new models based on the already learned weights. For datasets where deep learning-based methods perform very well or at least comparably to other methods, this added benefit could be the deciding factor.

4. Conclusions

During the last decade, sensors have become increasingly important across scientific fields and industries. Unfortunately, sensor data often contain missing values, which can significantly hamper the interpretation and possible analysis of the collected data. Consequently, the importance of methods capable of imputing these missing values with accurate estimates has grown considerably. In this study, we conducted a comparison of twelve imputation methods on a unique environmental microclimate monitoring dataset collected by the CNidT citizen science project. We extend the current literature by providing an extensive comparison of different missing value imputation methods originating from different backgrounds and imputation strategies. In addition, our work considers a real WSN dataset with a large number of sensors (1500), which is uncommon in the literature. Furthermore, we extend the evaluation of the evaluated techniques from random missings to masked missings, which provides a highly realistic evaluation scenario for practical implementations.
We evaluated the imputation methods for two different missing patterns: random missings, with the degree of missingness ranging from 10% to 50%, and masked missings, which were obtained using realistic missing value patterns. For all missing patterns, the MC method outperformed all other methods. MissForest and MCMC also performed relatively well in both scenarios, while MICE only achieved good results for random missings. The methods that are designed for WSNs specifically also performed well in both scenarios; however, they were not able to provide imputations for all missing values in the masked missings scenario. Finally, the deep learning-based methods, M-RNN, MIDA, and BRITS, performed poorly for both missing patterns, which can be attributed to the characteristics of our dataset. We can conclude from the results obtained that the methods that exploit spatial correlations within the dataset tend to perform better than the other methods. This can be explained by the relatively small distance between sensors, as well as the granularity of the temporal component. Moreover, since the data encompassed the period from April to September, temperatures predominantly experienced an upward trajectory, making it challenging to discern a clear trend in the temporal aspects of the data. These results can be extrapolated to similar scenarios where the number of sensors is high and densely distributed with a comparable length of time. The success of methods such as MC, MissForest, and MCMC, particularly in capturing spatial correlations within the dataset, suggests that they would generalize well to such environments. Despite challenges posed by masked missing values, these methods still demonstrated robust performance, implying their potential applicability in scenarios with complex missing data patterns.
Future research can expand upon our study with a more detailed assessment of (other) methods on different datasets. More specifically, different numbers of sensors and temporal granularity can be evaluated to more clearly identify the impact of these dataset specific features on the evaluated models. This can aid in the identification of a general best imputation technique across different WSNs. Furthermore, in future studies concerning missing data imputation for WSNs, additional features of the sensors or locations can be used to address missing values, such as the type of microclimate location, or other measured variables, such as the humidity in our specific use case. Also, the development of novel WSN-specific methods that efficiently exploit all structures (spatial and temporal) that are available in the data, carry significant potential. For example, a method could use an MI approach by first imputing all missing values using temporal correlations and subsequently using these imputations to obtain a more accurate spatial imputation, or vice versa. Additionally, cost-sensitive methods for missing value imputation can be evaluated, where over- or underestimations of the actual value can be penalized more heavily. Moreover, the evaluation of the temporal and spatial granularity and its impact on the imputation performance for various methods could be a valuable addition. Finally, our comparative study focuses on daily temperature values, whereas it may be interesting to evaluate it per 15-min interval or hourly and assess the imputation performance.
In conclusion, we were able to successfully impute missing values in our unique environmental monitoring dataset and provided guidelines for researchers who want to impute missing values in a similar dataset. Ultimately, we found that the best method to impute missing values is often dataset-specific and should be identified using a set of artificially induced missings, preferably both randomly generated and based on a realistic missing pattern.

Author Contributions

Conceptualization, T.D. and S.M.; methodology, T.D. and S.M.; software, T.D. and S.M.; validation, T.D. and S.M.; formal analysis, T.D. and S.M.; investigation, J.J.L. and F.J.R.M.; resources, J.J.L., F.J.R.M., S.L., E.M. and T.V.; data curation, J.J.L. and F.J.R.M.; writing—original draft preparation, T.D., S.M. and J.J.L.; writing—review and editing, F.J.R.M., S.L., E.M. and T.V.; visualization, T.D. and S.M.; supervision, F.J.R.M., S.L., E.M. and T.V.; project administration, J.J.L. and F.J.R.M.; funding acquisition, F.J.R.M., S.L., E.M. and T.V. All authors have read and agreed to the published version of the manuscript.

Funding

The project “Curieuze-Neuzen in de Tuin” was funded by a citizen science grant from the Department of Economy, Science, and Innovation (Flanders) and through a partnership between UAntwerpen, De Standaard, Rabobank Belgium, Orange Belgium, Flanders Environment Agency (VMM), Departement Omgeving (Flanders), and the Flemish Institute for Technological Research (VITO), with additional financial contributions from Aquafin and Bioplanet.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. The data can be found at: https://www.vlaanderen.be/datavindplaats/catalogus/curieuzeneuzen-in-de-tuin (accessed on 27 February 2024). Data can also be found in the SoilTemp database available at: https://www.soiltempproject.com/ [7]. Code for the replication of our study is available on GitHub: https://github.com/ThomasDecorteUA/Missing_Imputation_Sensors (accessed on 5 April 2024).

Acknowledgments

We thank Iris Janssens and Thomas Servotte for their valuable feedback. We thank the members of the Curieuze-Neuzen project team (Camille Allonsius, Sanne De Rooij, Stijn Van de Vondel, Johan Meuris, Mikhael Regni, Joke Van den Berge, and Ivan Nijs) and the 4400 participating citizen scientists for their amazing data collection effort.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analysis, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
AKEApplying k-Nearest Neighbor Estimation
BRITSBidirectional Recurrent Imputation for Time Series
CNidTCurieuze-Neuzen in de Tuin
CVCross-Validation
DAEDenoising Autoencoder
DESMData Estimation using Statistical Models
FAMDFactor Analysis of Mixed Data
IoTInternet of Things
KNNk Nearest Neighbors
M-RNNMulti-directional Recurrent Neural Network
MAEMean Absolute Error
MCMCMarkov Chain Monte Carlo
MCMatrix Completion
MCARMissing Completely at Random
MARMissing at Random
MICEMultiple Imputation using Chained Equations
MIDAMultiple Imputation using Denoising Autoencoders
MIMultiple Imputation
MNAR   Missing not at Random
MLMachine Learning
NB-IoTNarrowband Internet of Things
PCPrincipal Components
PCEPrediction Coverage Error
RNNRecurrent Neural Network
RFRandom Forest
RMSERoot-mean-square Error
WSNWireless Sensor Network

Appendix A. Hyperparameter Optimization and Results

Table A1. The hyperparameter grid search space for each evaluated model.
Table A1. The hyperparameter grid search space for each evaluated model.
MethodHyperparameter Search Space Optimal Hyperparameter Value
AKE [6]k: [2–25]15
timesteps: [5–30],18
BRITS [17]hidden dimensions: [5–15],8
learning rate: [0.0001,0.001,0.01,0.1]0.001
KNN [9]N neighbors: [2–25]5
MICE [11]N nearest features: [2–25]24
MIDA [16] Θ : [5,6,7,8,9]7
sequence length: [5–30],5
MRNN [15]hidden dimensions: [5–25],7
learning rate: [0.0001,0.001,0.01,0.1]0.01
Table A2. The RMSE scores for all imputation models, for different degrees of missingness and masked missings. The best performances are indicated in bold.
Table A2. The RMSE scores for all imputation models, for different degrees of missingness and masked missings. The best performances are indicated in bold.
Method10%20%30%40%50%Masked
AKE [6]0.4990.5080.5140.5160.5200.614
BRITS [17]2.8292.8592.9283.1223.5232.837
DESM [22]0.8570.8620.8660.8700.8750.901
KNN [9]1.7111.6951.6801.6481.6281.748
MC [14]0.3040.3220.3540.3720.3980.493
MCMC [12]0.8470.8560.8580.8610.8640.858
MICE [11]0.4140.4220.4370.4400.4481.195
MIDA [16]1.2002.3263.8505.2566.5931.294
MRNN [15]2.5582.4722.3942.3262.2682.023
Mean Imputation4.2594.2504.2414.2394.2455.703
MissForest [13]0.6640.6690.6730.6740.6750.714
Spline [8]1.3351.4061.4861.5921.7243.796
Table A3. The MAE scores for all imputation models, for different degrees of missingness and masked missings. The best performances are indicated in bold.
Table A3. The MAE scores for all imputation models, for different degrees of missingness and masked missings. The best performances are indicated in bold.
Method10%20%30%40%50%Masked
AKE [6]0.3700.3750.3760.3790.3840.444
BRITS [17]2.2552.2772.3212.4602.7442.265
DESM [22]0.6170.6210.6220.6270.6310.630
KNN [9]1.3201.3041.2931.2671.2541.371
MC [14]0.2260.2400.2550.2720.2940.341
MCMC [12]0.6400.6450.6440.6480.6500.650
MICE [11]0.3110.3180.3220.3280.3350.594
MIDA [16]0.9462.0823.5744.9596.2331.019
MRNN [15]1.9381.8481.7681.6981.6351.429
Mean Imputation3.4363.4243.4143.4133.4184.534
MissForest [13]0.5190.5200.5200.5220.5230.554
Spline [8]1.0131.0591.1101.1831.2722.520
Table A4. The PCE scores for all imputation models, for different degrees of missingness and masked missings. The best performances are indicated in bold.
Table A4. The PCE scores for all imputation models, for different degrees of missingness and masked missings. The best performances are indicated in bold.
Method10%20%30%40%50%Masked
AKE [6]1.0001.0001.0001.0001.0000.953
BRITS [17]1.0001.0001.0001.0001.0001.000
DESM [22]1.0001.0001.0001.0000.9990.985
KNN [9]1.0001.0001.0001.0001.0001.000
MC [14]1.0001.0001.0001.0001.0001.000
MCMC [12]1.0001.0001.0001.0001.0001.000
MICE [11]1.0001.0001.0001.0001.0001.000
MIDA [16]1.0001.0001.0001.0001.0001.000
MRNN [15]1.0001.0001.0001.0001.0001.000
Mean Imputation1.0001.0001.0001.0001.0001.000
MissForest [13]1.0001.0001.0001.0001.0001.000
Spline [8]1.0001.0001.0001.0001.0001.000
Table A5. The execution times for all imputation models, for different degrees of missingness and masked missings. The lowest execution times are indicated in bold.
Table A5. The execution times for all imputation models, for different degrees of missingness and masked missings. The lowest execution times are indicated in bold.
Method10%20%30%40%50%Masked
AKE [6]13.118.028.126.546.212.4
BRITS [17]52.481.7103.0123.6303.836.1
DESM [22]10.219.844.247.092.48.8
KNN [9]15.321.727.634.241.611.8
MC [14]1.52.03.93.76.22.2
MCMC [12]537.0552.1572.3567.6881.5435.9
MICE [11]23.525.732.625.841.623.4
MIDA [16]70.171.2106.081.5108.668.6
MRNN [15]1180.51193.61640.91278.71811.71144.0
Mean Imputation0.30.30.50.40.60.3
MissForest [13]1.31.41.81.62.31.3
Spline [8]0.40.40.70.40.70.4

Appendix B. Package Versions

Table A6. The versions of the Python packages used in the project.
Table A6. The versions of the Python packages used in the project.
PackageVersionReference
python3.11.5[42]
scikit-learn1.3.2[43]
fancyimpute0.7.0[44]
geopy2.4.1[45]
keras2.12.0[46]
missforest2.3.1[13]
numpy1.23.5[34]
pandas2.1.4[47]
scipy1.11.4[48]
seaborn0.13.1[49]
tensorflow2.12.0[50]
tensorflow-probability0.20.0[50]
torch2.1.2[51]

References

  1. Jamshed, M.A.; Ali, K.; Abbasi, Q.H.; Imran, M.A.; Ur-Rehman, M. Challenges, Applications, and Future of Wireless Sensors in Internet of Things: A Review. IEEE Sens. J. 2022, 22, 5482–5494. [Google Scholar] [CrossRef]
  2. Achyutha Prasad, N.; Chaitra, H.V.; Manjula, G.; Shabaz, M.; Martinez-Valencia, A.B.; Vikhyath, K.B.; Verma, S.; Arias-Gonzáles, J.L. Delay optimization and energy balancing algorithm for improving network lifetime in fixed wireless sensor networks. Phys. Commun. 2023, 58, 102038. [Google Scholar] [CrossRef]
  3. Jino Ramson, S.R.; Moni, D.J. Applications of wireless sensor networks—A survey. In Proceedings of the International Conference on Innovations in Electrical, Electronics, Instrumentation and Media Technology (ICEEIMT 2017), Coimbatore, India, 3–4 February 2017; pp. 325–329. [Google Scholar] [CrossRef]
  4. Kandris, D.; Nakas, C.; Vomvas, D.; Koulouras, G. Applications of Wireless Sensor Networks: An Up-to-Date Survey. Appl. Syst. Innov. 2020, 3, 14. [Google Scholar] [CrossRef]
  5. Kong, L.; Xia, M.; Liu, X.Y.; Chen, G.; Gu, Y.; Wu, M.Y.; Liu, X. Data loss and reconstruction in wireless sensor networks. IEEE Trans. Parallel Distrib. Syst. 2014, 25, 2818–2828. [Google Scholar] [CrossRef]
  6. Pan, L.; Li, J. K-Nearest Neighbor Based Missing Data Estimation Algorithm in Wireless Sensor Networks. Wirel. Sens. Netw. 2010, 2, 115–122. [Google Scholar] [CrossRef]
  7. Lembrechts, J.J.; Aalto, J.; Ashcroft, M.B.; De Frenne, P.; Kopeckỳ, M.; Lenoir, J.; Luoto, M.; Maclean, I.M.; Roupsard, O.; Fuentes-Lillo, E.; et al. SoilTemp: A global database of near-surface temperature. Glob. Chang. Biol. 2020, 26, 6616–6629. [Google Scholar] [CrossRef] [PubMed]
  8. Kreindler, D.M.; Lumsden, C.J. The effects of the irregular sample and missing data in time series analysis. In Nonlinear Dynamical Systems Analysis for the Behavioral Sciences Using Real Data; CRC Press: Boca Raton, FL, USA, 2016; pp. 149–172. [Google Scholar]
  9. Troyanskaya, O.; Cantor, M.; Sherlock, G.; Brown, P.; Hastie, T.; Tibshirani, R.; Botstein, D.; Altman, R.B. Missing value estimation methods for DNA microarrays. Bioinformatics 2001, 17, 520–525. [Google Scholar] [CrossRef]
  10. Rubin, D.B. Multiple imputation. In Flexible Imputation of Missing Data, 2nd ed.; Chapman and Hall/CRC: Boca Raton, FL, USA, 2018; pp. 29–62. [Google Scholar]
  11. White, I.R.; Royston, P.; Wood, A.M. Multiple imputation using chained equations: Issues and guidance for practice. Stat. Med. 2011, 30, 377–399. [Google Scholar] [CrossRef]
  12. Schunk, D. A Markov chain Monte Carlo algorithm for multiple imputation in large surveys. AStA Adv. Stat. Anal. 2008, 92, 101–114. [Google Scholar] [CrossRef]
  13. Stekhoven, D.J.; Bühlmann, P. MissForest—non-parametric missing value imputation for mixed-type data. Bioinformatics 2012, 28, 112–118. [Google Scholar] [CrossRef]
  14. Mazumder, R.; Hastie, T.; Edu, H.; Tibshirani, R.; Edu, T.; Jaakkola, T. Spectral Regularization Algorithms for Learning Large Incomplete Matrices. J. Mach. Learn. Res. 2010, 11, 2287–2322. [Google Scholar] [PubMed]
  15. Yoon, J.; Zame, W.R.; Van Der Schaar, M. Estimating Missing Data in Temporal Data Streams Using Multi-Directional Recurrent Neural Networks. IEEE Trans. Biomed. Eng. 2019, 66, 1477–1490. [Google Scholar] [CrossRef] [PubMed]
  16. Gondara, L.; Wang, K. MIDA: Multiple imputation using denoising autoencoders. In Proceedings of the PAKDD 2018: Advances in Knowledge Discovery and Data Mining, Melbourne, Australia, 3–6 June 2018; Springer: Berlin/Heidelberg, Germany, 2018; Volume 10939 LNAI, pp. 260–272. [Google Scholar] [CrossRef]
  17. Cao, W.; Wang, D.; Li, J.; Bytedance, H.Z.; Lab, A.I.; Li, Y.; Lab, B.A.; Li, L. BRITS: Bidirectional Recurrent Imputation for Time Series. In Advances in Neural Information Processing Systems, Proceedings of the NeurIPS, Montreal, ON, Canada, 3–8 December 2018; Neural Information Processing Systems Foundation, Inc. (NeurIPS): San Diego, CA, USA, 2018; Volume 31. [Google Scholar]
  18. Jadhav, A.; Pramod, D.; Ramanathan, K. Comparison of performance of data imputation methods for numeric dataset. Appl. Artif. Intell. 2019, 33, 913–933. [Google Scholar] [CrossRef]
  19. Jäger, S.; Allhorn, A.; Bießmann, F. A Benchmark for Data Imputation Methods. Front. Big Data 2021, 4, 693674. [Google Scholar] [CrossRef] [PubMed]
  20. Khayati, M.; Lerner, A.; Tymchenko, Z.; Cudré-Mauroux, P. Mind the gap. Proc. VLDB Endow. 2020, 13, 768–782. [Google Scholar] [CrossRef]
  21. Yozgatligil, C.; Aslan, S.; Iyigun, C.; Batmaz, I. Comparison of missing value imputation methods in time series: The case of Turkish meteorological data. Theor. Appl. Climatol. 2013, 112, 143–167. [Google Scholar] [CrossRef]
  22. Li, Y.; Ai, C.; Deshmukh, W.P.; Wu, Y. Data Estimation in Sensor Networks Using Physical and Statistical Methodologies. In Proceedings of the 2008 The 28th International Conference on Distributed Computing Systems, Beijing, China, 17–20 June 2008; pp. 538–545. [Google Scholar] [CrossRef]
  23. Lembrechts, J.; Van de Vondel, S.; Allonsius, C.; Nijs, I.; Meysman, F. CurieuzeNeuzen in de tuin: Eerste ervaringen met de microklimaatnetwerken van de toekomst. Bodem: Tijdschrift voor informatie-uitwisseling en discussie over duurzaam bodembeheer.-Alphen aan den Rijn, 1991, currens 2022, 33. Available online: https://edepot.wur.nl/651758 (accessed on 5 February 2024).
  24. CurieuzeNeuzen in de Tuin (Curious Noses In The Garden). 2021. Available online: https://curieuzeneuzen.be/home-en/ (accessed on 7 February 2024).
  25. Lembrechts, J.J.; van den Hoogen, J.; Aalto, J.; Ashcroft, M.B.; De Frenne, P.; Kemppinen, J.; Kopeckỳ, M.; Luoto, M.; Maclean, I.M.; Crowther, T.W.; et al. Global maps of soil temperature. Glob. Chang. Biol. 2022, 28, 3110–3144. [Google Scholar] [CrossRef]
  26. Lembrechts, J.J.; Lenoir, J.; R Scheffers, B.; De Frenne, P. Designing countrywide and regional microclimate networks. Glob. Ecol. Biogeogr. 2021, 30, 1168–1174. [Google Scholar] [CrossRef]
  27. Wild, J.; Kopeckỳ, M.; Macek, M.; Šanda, M.; Jankovec, J.; Haase, T. Climate at ecologically relevant scales: A new temperature and soil moisture logger for long-term microclimate measurement. Agric. For. Meteorol. 2019, 268, 40–47. [Google Scholar] [CrossRef]
  28. Chopde, N.R.; Nichat, M. Landmark based shortest path detection by using A* and Haversine formula. Int. J. Innov. Res. Comput. Commun. Eng. 2013, 1, 298–302. [Google Scholar]
  29. Little, R.J.; Rubin, D.B. Statistical Analysis with Missing Data; John Wiley & Sons: Hoboken, NJ, USA, 2019; Volume 793. [Google Scholar]
  30. Rubin, D.B. Inference and missing data. Biometrika 1976, 63, 581–592. [Google Scholar] [CrossRef]
  31. Schafer, J.L. Analysis of Incomplete Multivariate Data; CRC Press: Boca Raton, FL, USA, 1997. [Google Scholar]
  32. Van Buuren, S. Flexible Imputation of Missing Data; CRC Press: Boca Raton, FL, USA, 2018. [Google Scholar]
  33. Lokupitiya, R.S.; Lokupitiya, E.; Paustian, K. Comparison of missing value imputation methods for crop yield data. Environ. Off. J. Int. Environ. Soc. 2006, 17, 339–349. [Google Scholar] [CrossRef]
  34. Harris, C.R.; Millman, K.J.; van der Walt, S.J.; Gommers, R.; Virtanen, P.; Cournapeau, D.; Wieser, E.; Taylor, J.; Berg, S.; Smith, N.J.; et al. Array programming with NumPy. Nature 2020, 585, 357–362. [Google Scholar] [CrossRef] [PubMed]
  35. Azur, M.J.; Stuart, E.A.; Frangakis, C.; Leaf, P.J. Multiple imputation by chained equations: What is it and how does it work? Int. J. Methods Psychiatr. Res. 2011, 20, 40–49. [Google Scholar] [CrossRef] [PubMed]
  36. van Buuren, S. Multiple imputation of discrete and continuous data by fully conditional specification. Stat. Methods Med. Res. 2007, 16, 219–242. [Google Scholar] [CrossRef] [PubMed]
  37. van Buuren, S.; Groothuis-Oudshoorn, K. mice: Multivariate Imputation by Chained Equations in R. J. Stat. Softw. 2011, 45, 1–67. [Google Scholar] [CrossRef]
  38. Kong, A.; Liu, J.S.; Wong, W.H. Sequential imputations and Bayesian missing data problems. J. Am. Stat. Assoc. 1994, 89, 278–288. [Google Scholar] [CrossRef]
  39. Geman, S.; Geman, D. Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images. IEEE Trans. Pattern Anal. Mach. Intell. 1984, 6, 721–741. [Google Scholar] [CrossRef] [PubMed]
  40. Le Gruenwald, M.H. Estimating missing values in related sensor data streams. In Proceedings of the COMAD, Goa, India, 5–8 January 2005; pp. 83–94. [Google Scholar]
  41. Deng, Y.; Han, C.; Guo, J.; Sun, L. Temporal and Spatial Nearest Neighbor Values Based Missing Data Imputation in Wireless Sensor Networks. Sensors 2021, 21, 1782. [Google Scholar] [CrossRef]
  42. Van Rossum, G.; Drake, F.L. Python 3 Reference Manual; CreateSpace: Scotts Valley, CA, USA, 2009. [Google Scholar]
  43. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  44. Rubinsteyn, A.; Feldman, S. fancyimpute: An Imputation Library for Python. Available online: https://github.com/iskandr/fancyimpute (accessed on 5 February 2024).
  45. Gonzalez-Nieto, P.L.; Flechoso, M.G.; Mocoroa, M.A.; Martin, A.M.; Lorenzo, M.G.; Gomez, G.C.; Gomez, J.A.; Fraile, A.C.; Dagan, J.O.; Palomares, R.M.; et al. Design and Development of a Virtual Laboratory in Python for the Teaching of Data Analysis and Mathematics in Geology: Geopy. In Proceedings of the INTED2020 Proceedings, Valencia, Spain, 2–4 March 2020; IATED: Valencia, Spain, 2020; pp. 2236–2242. [Google Scholar]
  46. Chollet, F. Keras. 2015. Available online: https://keras.io (accessed on 7 February 2024).
  47. Wes McKinney. Data Structures for Statistical Computing in Python. In Proceedings of the 9th Python in Science Conference, Austin, TX, USA, 28–30 June 2010; van der Walt, S., Millman, J., Eds.; 2010; pp. 56–61. [Google Scholar] [CrossRef]
  48. Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef] [PubMed]
  49. Waskom, M.L. Seaborn: Statistical data visualization. J. Open Source Softw. 2021, 6, 3021. [Google Scholar] [CrossRef]
  50. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. 2015. Available online: https://www.tensorflow.org/ (accessed on 12 February 2024).
  51. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32; Curran Associates, Inc.: Sydney, Australia, 2019; pp. 8024–8035. [Google Scholar]
Figure 1. (a) The TMS-NB microclimate sensor was used in a large-scale citizen science project on microclimate monitoring. The sensor measures temperature at three heights, as well as soil moisture. Data transmission occurred via NB-IoT. (b) The WSN covered 4400 gardens across Flanders. Sensor locations are colored based on whether time series were complete (green) or had missing records (red).
Figure 1. (a) The TMS-NB microclimate sensor was used in a large-scale citizen science project on microclimate monitoring. The sensor measures temperature at three heights, as well as soil moisture. Data transmission occurred via NB-IoT. (b) The WSN covered 4400 gardens across Flanders. Sensor locations are colored based on whether time series were complete (green) or had missing records (red).
Sensors 24 02416 g001
Figure 2. Example of missing data pattern in a representative sensor time series, which includes 15-min temperature recordings over a six-month period. (a) Time series for a sensor with complete data. (b) The same time series but with missing data artificially imposed. The missing time points are based on a mask derived from a different sensor with actual missing data.
Figure 2. Example of missing data pattern in a representative sensor time series, which includes 15-min temperature recordings over a six-month period. (a) Time series for a sensor with complete data. (b) The same time series but with missing data artificially imposed. The missing time points are based on a mask derived from a different sensor with actual missing data.
Sensors 24 02416 g002
Figure 3. Schematic representation of a test dataset used in imputation analysis. A network of N sensors is providing data readings over T time points. Artificially induced missings are indicated by the orange fields.
Figure 3. Schematic representation of a test dataset used in imputation analysis. A network of N sensors is providing data readings over T time points. Artificially induced missings are indicated by the orange fields.
Sensors 24 02416 g003
Figure 4. (a) The RMSE for all models for different degrees of missingness. (b) The MAE for all models for different degrees of missingness.
Figure 4. (a) The RMSE for all models for different degrees of missingness. (b) The MAE for all models for different degrees of missingness.
Sensors 24 02416 g004
Figure 5. (a) The RMSE for all models for the masked data. (b) The MAE for all models for the masked data.
Figure 5. (a) The RMSE for all models for the masked data. (b) The MAE for all models for the masked data.
Sensors 24 02416 g005
Figure 6. (a) The PCE for all models for the masked data. (b) The execution time (on a logarithmic scale) for all models, averaged over all degrees of missingness, including masked missings.
Figure 6. (a) The PCE for all models for the masked data. (b) The execution time (on a logarithmic scale) for all models, averaged over all degrees of missingness, including masked missings.
Sensors 24 02416 g006
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Decorte, T.; Mortier, S.; Lembrechts, J.J.; Meysman, F.J.R.; Latré, S.; Mannens, E.; Verdonck, T. Missing Value Imputation of Wireless Sensor Data for Environmental Monitoring. Sensors 2024, 24, 2416. https://doi.org/10.3390/s24082416

AMA Style

Decorte T, Mortier S, Lembrechts JJ, Meysman FJR, Latré S, Mannens E, Verdonck T. Missing Value Imputation of Wireless Sensor Data for Environmental Monitoring. Sensors. 2024; 24(8):2416. https://doi.org/10.3390/s24082416

Chicago/Turabian Style

Decorte, Thomas, Steven Mortier, Jonas J. Lembrechts, Filip J. R. Meysman, Steven Latré, Erik Mannens, and Tim Verdonck. 2024. "Missing Value Imputation of Wireless Sensor Data for Environmental Monitoring" Sensors 24, no. 8: 2416. https://doi.org/10.3390/s24082416

APA Style

Decorte, T., Mortier, S., Lembrechts, J. J., Meysman, F. J. R., Latré, S., Mannens, E., & Verdonck, T. (2024). Missing Value Imputation of Wireless Sensor Data for Environmental Monitoring. Sensors, 24(8), 2416. https://doi.org/10.3390/s24082416

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop