Next Article in Journal
The Probability Distribution of Worldwide Forest Areas
Next Article in Special Issue
Indoor and Outdoor Air Quality for Sustainable Life: A Case Study of Rural and Urban Settlements in Poor Neighbourhoods in Kenya
Previous Article in Journal
Clay-Based Products Sustainable Development: Some Applications
Previous Article in Special Issue
Analysis of the Effectiveness of Air Pollution Control Policies Based on Historical Evaluation and Deep Learning Forecast: A Case Study of Chengdu-Chongqing Region in China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Air Pollution Prediction Using an Ensemble of Dynamic Transfer Models for Multivariate Time Series

1
H. Milton Stewart School of Industrial & Systems Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA
2
Industrial Engineering, Hanyang University, Seoul 04763, Korea
*
Author to whom correspondence should be addressed.
Sustainability 2021, 13(3), 1367; https://doi.org/10.3390/su13031367
Submission received: 7 December 2020 / Revised: 20 January 2021 / Accepted: 25 January 2021 / Published: 28 January 2021
(This article belongs to the Special Issue Air Pollution Monitoring and Environmental Sustainability)

Abstract

:
Entering a new era of big data, analysis of large amounts of real-time data is important, and air quality data as streaming time series are measured by several different sensors. To this end, numerous methods for time-series forecasting and deep-learning approaches based on neural networks have been used. However, they usually rely on a certain model with a stationary condition, and there are few studies of real-time prediction of dynamic massive multivariate data. Use of a variety of independent variables included in the data is important to improve forecasting performance. In this paper, we proposed a real-time prediction approach based on an ensemble method for multivariate time-series data. The suggested method can select multivariate time-series variables and incorporate real-time updatable autoregressive models in terms of performance. We verified the proposed model using simulated data and applied it to predict air quality measured by five sensors and failures based on real-time performance log data in server systems. We found that the proposed method for air pollution prediction showed effective and stable performance for both short- and long-term prediction tasks. In addition, traditional methods for abnormality detection have focused on present status of objects as either normal or abnormal based on provided data, we protectively predict expected statuses of objects with provided real-time data and implement effective system management in cloud environments through the proposed method.

1. Introduction

Massive real-time data storage and real-time data visualization are available in many industries, and have improved data analysis techniques for real-time data. In other words, among various versions of time-series analysis, the prediction in real time is the one of main interests in the field. Also, much log big data has been produced between web or mobile applications and server systems because of developments of web and IoT systems, etc. Thus, analyzing this kind of data is becoming increasingly important these days [1,2]. In particular, APM (Application Performance Management) is a real-time log big data analysis system that collects and manages performance information of a server system between usages of user applications and services of a server system such as web application server or data base server.
Various extended models such as VAR (Vector Autoregressive) and VARMA (Vector Autoregressive Moving Average) have been used for multivariate time-series analysis. VAR and VEC (Vector Error-correction) were exploited for long-term prediction based on multivariate time-series data [3], and VARMA was used for multivariate time-series data in financial services [4]. Study in [5] researched prediction performance of the VAR model using direct multi-step estimation for both stationary and non-stationary time series generated in economic activities, and study in [6] forecasted price of electricity by VAR model using fractional cointegration. One drawback of VAR model is that the number of parameters to be estimated can become large [7]. Recently, predictions for Fuzzy time series were performed using a multivariate heuristic model [8], and a new method using Fuzzy relation based on a neural network algorithm was suggested for high-dimensional time series data [9]. However, these models need to satisfy too many conditions and constraints; the VARMA model is slow with complicated data [10]. In addition, even if the models handle multivariate time-series data, they are usually not suitable to forecast a certain dependent variable of the data. Thus, given many situations, traditional single models for forecasting multivariate time series have limits.
As one possible approach to these problems, an ensemble method combining many models have been proposed for time-series data in diverse situations. In particular, as solutions based on neural network algorithm have been in fashion, [11] suggested a new type of ensemble method using nonlinear weights with ARIMA, ANN, and RNN models. Despite some advantages, however, the method is not adequate for real-time forecasting and has a limitation of increasing complexity as the number of models increases. Similarly, based on research using a neural network algorithm, these kinds of multivariate time-series forecasting methods are not suitable for real-time analysis because of long training time with high complexity [12]. Recently, for real-time prediction, the ELMK method combining ELM (Extreme Learning Machine) and kernel methods was proposed [13]. It showed increased real-time forecasting performance on non-stationary time series by applying a fixed memory-based prediction algorithm that eliminates training data of the past. In addition, [14] suggested an online prediction model by applying versions of Newton and gradient descent algorithms into a loss function of the ARMA model. Although these kinds of studies on real-time prediction of time series have been conducted, there are few studies focusing on real-time analysis for multivariate time series. A new real-time multivariate forecasting model is needed to replace the univariate one to handle many time-series features collected online.
In this paper, we propose multivariate ensemble method based on dynamic transfer model for stable real-time prediction and verify its performance by applying it to predict failures with performance log data generated in a server system. Time-series prediction algorithms using a transfer model are a form of lagged regression where input variables X t affect autoregressive input variables of response variables Y t . There are various ways of building and identifying lagged regression models according to relationships between input X t and output Y t . One disadvantage of a lagged regression model is that it is difficult to consider uncertainty of a dynamic input variable process [15]. Also, when one attempts to find lags from cross correlation between input and output variables, the model can be unclear and too empirical [16]. Although some solutions such as using Monte-Carlo-based analysis [17] have been proposed to solve this problem, we suggest a novel ensemble-based method. An ensemble approach incorporates various features from various locations and scales that can be beneficial in many fields [18]. The proposed method selects various input variables for each lagged regression and generate diverse dynamic transfer models based on the selected input variables, with autoregressive time-series analysis of output variables. We perform this approach to build a basis ensemble model with offline training data and then forecast response variables in real time by updating the basis model online. Since there are many dynamic transfer basis models generated from diverse combinations of input variables, autoregressive orders, and difference orders, and because the best model is selected from a set of bases by an ensemble method, we obtain the best offline model that is suitable in as many situations as possible. Next, we update weights for each basis model and regression coefficients in real time, which enables the proposed approach to be faster, more accurate, and more stable for predictions in many environments. We consistently manage the suggested model by periodically offline updating the basis model to reflect unpredictable environmental changes that are possible in industry. Moreover, we effectively use memory since online updating does not require much information.
We verified significant performance of the proposed real-time forecasting algorithm based on the ensemble of dynamic transfer models for multivariate time series by analyzing simulated dataset. Next, we suggested a way of using the proposed method for server performance data in server systems. For server system data, we predict failures with real-time data from a performance management system. Given plenty of multivariate applications, servers, and databases in the field of performance management, it is important not only to assess status of the present system, but also to correctly predict failures. The precautionary action with high prediction accuracy enables us to consistently operate services in a cloud environment of a server system because of directly allocating resources such as CPU or memory. We showed that the provided method can achieve good prediction performance with the real-time multivariate time-series data in a server system.

2. Preliminaries: ARIMA, Transfer Function, Ensemble

In this section, we provide a summary of classical time-series models and an ensemble model for the prediction model proposed in the next section. A time series is a sequence of quantitative observations at successive time points. We denote x t to be the observation at time t. A sophisticated, classical, and widely used model is autoregressive integrated moving average (ARIMA) with order p, d, and q, ARIMA( p , d , q ):
( 1 i = 1 p α i B i ) ( 1 B ) d x t = ( 1 + i = 1 q ϕ i B ) a t ,
in which a t is the zero-mean white noise at time t and B is a backward shift operator. Autoregressive and moving average, ARMA, is a special case of ARIMA( p , d , q ) with a difference order d of zero. ARMA is a statistical model for time series to describe a weakly stationary stochastic process with autoregression and moving average terms. The model ARIMA( p , d , q ) is ARIMA( p , 0 , q ) for d-step differential process ( 1 D ) d x t . If the moving average polynomial, ( 1 + i = 1 q ϕ i B ) , is invertible, the model can be represented as an infinite autoregressive model, x t = i = 1 α i x t i + a t . A transfer function model represents the relation between input x t and output Y t that reflects input-output delays [19]. The use of moving average polynomials δ ( B ) and ω ( B ) for both x t and Y t in the model enables determination of the causal relationship between two time series: for delay b and noise n t , δ ( B ) Y t = ω ( B ) x t b + n t . If multiple inputs, x 1 , t , , x m , t exist, the model is
Y t = δ ( B ) 1 ω 1 ( B ) x 1 , t b 1 + + δ ( B ) 1 ω m ( B ) x m , t b t + n t ,
where n t is noise [20]. Modeling transfer functions with multiple inputs and reflecting nonlinearity between input and output are challenging. In light of this observation, we propose a dynamic transfer model. An ensemble method is a way of integrating multiple machine-learning algorithms to obtain better performance than one algorithm alone. A key component in ensembles is to have multiple base learners exceeding random algorithm and including diversity. To promote diversity in base learners and numerically avoid overfitting, bootstrap aggregating is widely adopted by using a randomly drawn subset of a training set. In bootstrap aggregating, an ensemble classifier f is composed of base learners, f ^ i , i = 1 , , B , with associated weight w i as follows:
f ^ = w 1 f ^ 1 + w 2 f ^ 2 + + w B f ^ B .
Weight w i is often determined according to the performance of f ^ i .

3. Proposed Method

The proposed method for multivariate time series consists of two components, dynamic transfer model and an ensemble of dynamic transfer model. First, we describe the formulation of dynamic transfer model and point out both similarities and differences between the existing models and the proposed one. Then, we describe the integration of dynamic transfer models for reliable and accurate prediction in multivariate time series.

3.1. Dynamic Transfer Models

In this section, we explain a dynamic transfer model using multivariate time series and expound an ensemble of such models for more accurate and stable performance. We consider two versions, one that stores a fixed number of past observations and another that does not need to store past observations but update the model parameters on a real-time basis. The basic idea of a dynamic transfer model is to combine a dynamic lagged regression model with a transfer function model to allow the model to adapt its parameters in an online fashion.
Consider a time series Y t corresponding to a response variable and independent time-series signals x 1 , t , x 2 , t , , x m , t , t = 1 , , T . We start with a dynamic regression model for response time series Y t ,
( 1 B ) d Y t = β 0 + k = 1 p β k Y t k + j = 1 m I j α j x j , t + ϵ t ,
of autoregressive order p, difference order d, and input selector I of m × 1 , where I j is 1 if the corresponding x j , t is included and zero otherwise and ϵ t is white noise. For example, for p = 2 , d = 1 , and I = [ 1   0   0   1 ] T , the corresponding model is
Y t Y t 1 = β 0 + β 1 Y t 1 + β 2 Y t 2 + α 1 x 1 , t + α 4 x 4 , t + ϵ t .
As shown in the example, for d p , the model with p, d, and I produces the same result as the model with p, 0, and I.
The model attempts to determine the effect of input signals, x j , t , j = 1 , , m , on response signal Y t in such a way that the dynamic response of Y t may vary in x j , t using input selector I. We maintain several input selectors, I ( ) , = 1 , , L , in an ensemble stage, which will be explained in the following section. The model can capture the effect of x j , t that declines gradually to zero as in lagged regression models. For example, if Y t is modeled by an infinite number of lagged values x 1 , t , among m = 4 input signals, with an exponentially decaying effect.
Y t = w { x 1 , t + β x 1 , t 1 + β 2 x 1 , t 2 + } + ϵ t .
Then the model is rewritten as Y t = w 1 β B x 1 , t + ϵ ( t ) for | β | < 1 and is simplified as a lagged dependent variable model,
Y t = β Y t 1 + α x 1 , t + ϵ t ,
which is an instance of the proposed model in (1) with p = 1 , d = 0 , and I = [ 1   0   0   0 ] T .
The dynamic transfer model shares similarity with an autoregressive moving average with external terms (ARMAX) model. ARMAX is an application of ARMA with external explanatory variables. However, we mention the following two differences between the two models. First, the proposed model considers input signal x j , t to be stochastic, whereas ARMAX models consider external inputs as deterministic. We separately build the time-dependent model of input signal x j , t . Second, the model allows the dynamics of a response signal by selection of input signals, using input selector I, in various ways. This approach enables the model to generate several meaningful results that will be used and aggregated in the ensemble stage.
Indeed, the proposed model faces a few computational challenges. For model training, it attempts to find the relationship between response signal Y t and current input signal x j , t only at time t. Given the input signals as a stochastic process, we try to explain Y t using input signals at a certain fixed time, usually current time t. This approach is considered a model simplification, and will be compensated by the ensemble stage in the next section. Difference order is also challenging, and several strategies such as empirical analysis of autocorrelation, numerical search of the lowest standard deviation, and theoretical testing of stationarity can be adopted in practice. For model identifiability, we need to fix p , d , and I in the model. Identification of optimal parameters is nontrivial but challenging. For example, to find the optimal difference order d, one can adopt several strategies such as empirical analysis of autocorrelation, numerical search of the lowest standard deviation, and theoretical testing of stationarity. To cope with this challenge, we set difference order d from 1 to D, assessing its performance in the following ensemble stage. The same strategy is applied to p and I. For fixed p , d , and I in the model, the model is identifiable, and we estimate the parameters using least-squares. The estimates are consistent but possibly biased in some cases [21]. We similarly build an autoregressive model for input x j , t of autoregressive order p j and difference order d j ,
x j , t ( 1 B ) d j = ϕ j , 0 + k = 1 p j ϕ j , k x j , t k + ϵ j , t .
We represent the prediction of input signal x j , t + h , h 1 by available measurements up to current time t when predicting Y t + h in Equation (1). This procedure is commonly used to replace an unobserved, usually random input with its expectations from an auxiliary model [22]. We denote the h-step-ahead prediction of Y t + h using observations up to t by Y ^ t + h | t . For example, suppose the exemplary model in (2) is used, and the following autoregressive models for x 1 , t and x 4 , t were obtained, respectively: x 1 , t = ϕ ^ 1 , 0 + ϕ ^ 1 , 1 x 1 , t 1 + ϵ 1 , t and x 4 , t = ϕ ^ 4 , 0 + ϕ ^ 4 , 1 x 4 , t 1 + ϕ ^ 4 , 2 x 4 , t 2 + ϵ 4 , t . Then, the one-step-ahead prediction Y t + 1 | t is obtained as follows:
Y ^ t + 1 | t = β ^ 0 + β ^ 1 Y t + β ^ 2 Y t 1 + α ^ 1 ( ϕ ^ 1 , 0 + ϕ ^ 1 , 1 x 1 , t ) + α ^ 4 ( ϕ ^ 4 , 0 + ϕ ^ 4 , 1 x 4 , t + ϕ ^ 4 , 2 x 4 , t 1 ) .

3.2. Ensemble of Dynamic Transfer Models

For accurate and reliable online prediction using multivariate time series, we propose an ensemble approach that generates numerous dynamic transfer models, described in the previous section and used as a base learner in the ensemble method, and select a few models that work well in predefined prediction tasks. Then, we update the selected models on a real-time basis using either a fixed amount of past data or recursive least-squares. To build an ensemble model, we conduct the following three steps:
Step 1. 
Generate numerous candidate models for response Y t ( ) as in (1) with p , d , and I , = 1 , , L , using training observations.
Step 2. 
Choose the top-K models as base learners among the L candidate models in terms of minimum prediction error.
Step 3. 
Generate prediction models for input variables, x j , t , using training observations.
The above steps are performed in an offline training session for a h-step-ahead prediction task. Specifically, we consider models of autoregressive orders from 1 to P, difference orders from 1 to D, and input variables no greater than q among m variables, which brings the total number of considered models to P D j q m j in Step 1. If P D j q m j models are computationally feasible to handle, we set L = P D j q m j ; otherwise, L is a large number that can be handled computationally. It is possible that different models for x j , t are used in Step 3. We apply a model-weighting strategy based on prediction error [23]. For a candidate model Y t ( ) , the model-performance weight, w , in training is calculated and normalized as follows:
w = 1 t = T G + 1 h T h ( Y t Y ^ t + h | t ( ) ) 2 ,                 w w l = 1 L w l ,
in which parameter G is the number of time observations for performance evaluation in offline training. Since the weight is reciprocal to the sum of squared error, the larger is w , the better is the model. Then, we choose the top-K models, Y ^ t ( 1 ) , , Y ^ t ( K ) as base learners in terms of w , and we build the final ensemble model as the following:
Y ^ t = w 1 Y ^ t ( 1 ) + w 2 Y ^ t ( 2 ) + + w K Y ^ t ( K ) .
Similarly, for a prediction model of each input x j , t in Step 3, we consider models of autoregressive orders from 1 to P , and difference orders from 1 to D and construct an ensemble of them similarly. In short, the offline training session produces K feasible models of input x j , t for the h-step-ahead prediction task, yielding x ^ j , t + h . Then Y ^ t + 1 | t and Y ^ t + 2 | t that is Y ^ t + h | t in (1), are constructed using x ^ j , t + h .
When online streaming observations of Y t , x 1 , t , x 2 , t , , x m , t , t > T are available, we update the coefficients ( β 0 , β k , and α j in (1)) in each of the top-K models. We have two options in updating. One is to use a fixed number of past observations from t and refit each model to calculate the coefficients. This option is reasonable in that we cannot indefinitely store streaming data in memory due to fixed memory size. Denote this version as EDT-w (ensemble of dynamic transfer model using fixed window). The second option is to apply recursive least-squares to update the coefficients. The second option also makes sense in that not only does the recursive least-squares (RLS) approach enable fast updating, but also makes possible the use of all past observations without storing them in memory. Denote this version as EDT-r (ensemble of dynamic transfer model using RLS). Then we update model-performance weight w using the most recent prediction task as follows:
w = 1 t = t G + 1 h t h ( Y t Y ^ t + h | t ( ) ) 2 ,                 w w l = 1 L w l .

4. Experiments

For all experiments, we deployed six traditional time-series forecasting methods, ARIMA, ARIMAX, ANN (Artificial Neural Network), ANNX (Artificial Neural Network with extra predictor), RNN (Recurrent Neural Network), VAR (Vector Autoregressive), and the two proposed methods, EDT-w and EDT-r. ANN is a collection of nodes, also known as neurons, with a layer structure to map inputs to outputs. ANNX is an application of ANN with external explanatory variables. RNN is a class of artificial neural networks with internal, hidden and autoregressive, states, and VAR is a multivariate autoregressive version of ARIMA. For the proposed methods, we set the parameters as P = 9 ,   D = 3 ,   P = 5 ,   D = 1 ,   K = 40 , and window size = 200 . To compare short-term and long-term forecasting performance, we provide four forecasting errors for t + 1 ,   t + 3 ,   t + 6 , and t + 12 in terms of RMSE (root mean square error) and repeat the computation three times. Also, computing time in seconds is provided as reference, and all of them are in a reasonable range. The computing time of only two proposed methods are different for t + 1 ,   t + 3 ,   t + 6 , and t + 12 because they select the local optimal parameters while training depending on the time window.

4.1. Simulation 1

The first simulation dataset is obtained using VAR.sim function from package multivar in R that aims to generate VAR-based simulation data. Since real-life datasets in the later sections are eight-dimensional, we generated the same eight-dimensional time-series data, ( x 1 , t , , x 8 , t ) , and set the last one, x 8 , t as a target time-series, y t , for forecasting. One example is provided in Figure 1. The length of each time-series is 5000 with the first 4000 data used for training and the remaining 1000 data for testing. There is no anomaly in this first simulation with stationarity compared with simulations without stationarity in the next section.
Outputs for the first simulation are shown in Table 1. For t + 1 , VAR is the most likely to show the best performance since the first simulation data were generated by the VAR model of VAR.sim function in R. Indeed, VAR was superior for short-term prediction. However, the proposed EDT-w method dominates the other methods for t + 3 ,   t + 6 , and t + 12 with longer-term.

4.2. Simulation 2

We used arima.sim function from package stats in R to generate the second simulation dataset. First, we generated seven independent time-series, ( x 1 , t , , x 7 , t ) , as a form of ARIMA ( m , n , 0 ) where 1 m 5 and 0 n 2 . We also pre-set a range of autoregressive coefficients to be between −1 and 1. Then, we defined y t as a target time-series that is dependent on ( x 1 , t , , x 7 , t ) . We first generated independent time-series with the same rule for ( x 1 , t , , x 7 , t ) and define it as z t . Then, we added a linear combination of ( x 1 , t , , x 7 , t ) at time t with coefficients between −1 and 1 to z t and defined it as the final target time-series, y t . For example, y t as a form of ARIMA ( 2 , 2 , 0 ) can be repreresented as
y t = z t + α 1 x 1 , t + + α 7 x 7 , t
where z t = β 1 z t 1 + β 2 z t 2 + θ 1 ϵ t 1 + θ 2 ϵ t 2 + ϵ t and 1 α , β , θ 1 . As in the previous simulations, the length of each time-series is 5000 items, and the first 4000 were used for training and the later 1000 for testing. One example is visualized in Figure 2.
The output in Table 2 is very similar to that of the previous simulation because this second simulation data were created using arima.sim function in R. Thus, ARIMA shows the lowest error for t + 1 and t + 3 . However, as in the previous case, the proposed EDT-w method is always better than the other methods for t + 6 and t + 12 with longer-term.

4.3. Application 1: Air Quality Data

The one-year dataset (https://archive.ics.uci.edu/ml/datasets/Air+Quality) from March 2004 to February 2005 [24], registered at UCI Machine-Learning Repository, has 9358 samples of hourly averaged responses obtained by five metal oxide chemical sensors embedded in an Air Quality Chemical Multisensor Device. Originally, this was used to prove cross-sensitivities and improve sensor capabilities of concentration estimation [25]. For application to the proposed method, we selected eight columns, CO(GT), PT08.S1(CO), C6H6(GT), PT08.S2(NMHC), NOx(GT), PT08.S3(NOx), PT08.S4(NO2), and PT08.S5(O3), and set the C6H6(GT) representing hourly averaged Benzene concentration as a target value for time-series forecasting because this is one of the most important target values in the previous reference. This is visualized in Figure 3.
We use four types of experiments, and the first 5000, 6000, 7000, and 8000 data are used as training data for each. Then, the next immediate 1000 data are used as testing data for each scenario. Missing data are originally tagged with a −200 value and were replaced with na.approx function in R.
Outputs for the air quality application are shown in Table 3. EDT-w always shows the best performance except for two cases for t + 1 . In addition, forecasting performance in terms of RMSE is usually more than twice that of the other methods. Thus, this experiment provides evidence of the strength of the proposed method.
In conclusion, EDT-w usually shows the best performance for long-term time-series forecasting such as t + 6 and t + 12 . Also, it shows not the best but very high performance even for short-term forecasting such as t + 1 and t + 3 . Therefore, the proposed EDT-w can be a good option for accurate and stable forecasting performance for longer-term.

4.4. Application 2: APM Data

Various events occur while applications are in use. Modifying an existing application, sudden heavy inflow of users, and delayed processing are representative examples of such events. They can be stored in a form of log data in connected database or server systems, and APM (application performance management) is an analytic solution that can detect and resolve failure or trouble in web applications by analyzing such log data.
The given APM dataset provides 54,995 items of time-series data collected every 10 s for eight log data of Used memory, System CPU, User CPU, Count, Num http, Num javad, Num jdbc, and Response, as in Table 4, and they are visually presented in Figure 4.
We set the last variable, Response, as a target time-series to predict and the remaining seven variables as its covariates. The first 5000 data items were used for training, and eight manually selected subsets of 1000 items of time-series data are used for testing. We divide Response into eight subsets for comparison: four are stationary according to a stationarity test and the other four are non-stationary. Response without any system failure or trouble is usually in the normal range regardless of time point.
Outputs for the APM application are shown in Table 5. APM based on time-series forecasting is reasonable because its computing time is usually less than 1 or 2 s, which is much smaller than the 10-second interval of collecting data. Next, ANN or VAR shows better performance for t + 1 and t + 3 , while the proposed EDT-r and EDT-w do so for t + 6 and t + 12 . The most interesting thing is that VAR shows extremely bad performance for t + 6 and t + 12 , indicating that VAR is not good for time-series with anomalies, while the proposed EDT-r and EDT-w are very stable in such a situation. Lastly, the second and third non-stationary cases for t + 12 are the only two where EDT-r achieves the best performance.

5. Conclusions

For the prediction task for streaming multivariate time series, numerous models are possible, ranging from classical models to modern deep-learning models. Effective models need to be accurate in a certain h-time ahead task by Y ^ t + h and consistent and reliable in several h-time head tasks. Transfer models together with ARIMA are nonlinear, rather classical, and established forecasting models. We aimed to improve the models to make them dynamic by including variable selectors and stable by constructing an ensemble of them. Though many sub-models participate in the final model, each of which is easy to modify in the face of new online data. From a computational viewpoint, model building requires a comparable amount of time in comparison to several existing models such as VAR and RNN, and computation time will decrease if parallel processing is in use for the building of each sub-model. The prediction performance of the proposed model is promising and effective not only in two adopted models, but also in two real-life multivariate time series. For the prediction task of atmospheric pollution, the data are interrelated and stochastic by nature, so the need for multivariate time series, including external variables, escalates. The current study only included seven internal chemicals for the target variable of air pollution. In future, we hope to include external factors in the proposed model and to verify the model capacity. We plan to refine the model by adding a forgetting factor to updating sub-models and by reflecting theoretical aspects of the models.

Author Contributions

Conceptualization, D.C. and K.L.; Investigation, T.K. and D.C.; Methodology, T.K., D.C. and K.L.; Software, T.K..; Resources, G.L. and T.K.; Data curation, T.K. and D.C.; Writing—original draft, D.C. and K.L.; Writing—review and editing, T.K., G.L. and K.L.; Supervision, K.L.; Funding acquisition, K.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Ministry of SMEs and Startups (S2858156).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author while restrictions may apply to the availability of the APM data.

Conflicts of Interest

The authors declare that there is no conflict of interest regarding publication of this paper.

References

  1. Barford, P.; Kline, J.; Plonka, D.; Ron, A. A signal analysis of network traffic anomalies. In Proceedings of the 2nd ACM SIGCOMM Workshop on Internet Measurment, Marseille, France, 6–8 November 2002; pp. 71–82. [Google Scholar]
  2. Kruegel, C.; Vigna, G. Anomaly detection of web-based attacks. In Proceedings of the 10th ACM conference on Computer and Communications Security, Washington, DC, USA, 27–30 October 2003; pp. 251–261. [Google Scholar]
  3. Quenouille, M.H.; Quenouille, M.H. The Analysis of Multiple Time-Series; No. 519.232 Q84; Griffin: London, UK, 1957. [Google Scholar]
  4. Veenstra, A.W.; Haralambides, H.E. Multivariate autoregressive models for forecasting seaborne trade flows. Transp. Res. Part E Logist. Transp. Rev. 2001, 37, 311–319. [Google Scholar] [CrossRef]
  5. Chevillon, G.; Hendry, D.F. Non-parametric direct multi-step estimation for forecasting economic processes. Int. J. Forecast. 2005, 21, 201–218. [Google Scholar] [CrossRef] [Green Version]
  6. Haldrup, N.; Nielsen, F.S.; Nielsen, M.Ø. A vector autoregressive model for electricity prices subject to long memory and regime switching. Energy Econ. 2010, 32, 1044–1058. [Google Scholar] [CrossRef] [Green Version]
  7. Gong, C.; Tang, P.; Wang, Y. Measuring the network connectedness of global stock markets. Phys. Stat. Mech. Its Appl. 2019, 535, 122351. [Google Scholar] [CrossRef]
  8. Huarng, K.H.; Yu, T.H.K.; Hsu, Y.W. A multivariate heuristic model for fuzzy time-series forecasting. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 2007, 37, 836–846. [Google Scholar] [CrossRef] [PubMed]
  9. Egrioglu, E.; Aladag, C.H.; Yolcu, U.; Uslu, V.R.; Basaran, M.A. A new approach based on artificial neural networks for high order multivariate fuzzy time series. Expert Syst. Appl. 2009, 36, 10589–10594. [Google Scholar] [CrossRef]
  10. Isufi, E.; Loukas, A.; Perraudin, N.; Leus, G. Forecasting time series with varma recursions on graphs. IEEE Trans. Signal Process. 2019, 67, 4870–4885. [Google Scholar] [CrossRef] [Green Version]
  11. Adhikari, R.; Agrawal, R.K. A novel weighted ensemble technique for time series forecasting. In Pacific-Asia Conference on Knowledge Discovery and Data Mining; Springer: Berlin/Heidelberg, Germany, 2012; pp. 38–49. [Google Scholar]
  12. Borovykh, A.; Oosterlee, C.W.; Bohté, S.M. Generalization in fully-connected neural networks for time series forecasting. J. Comput. Sci. 2019, 36, 101020. [Google Scholar] [CrossRef]
  13. Wang, X.; Han, M. Online sequential extreme learning machine with kernels for nonstationary time series prediction. Neurocomputing 2014, 145, 90–97. [Google Scholar] [CrossRef]
  14. Anava, O.; Hazan, E.; Mannor, S.; Shamir, O. Online learning for time series prediction. In Proceedings of the 26th Annual Conference on Learning Theory, Princeton, NJ, USA, 12–14 June 2013; pp. 172–184. [Google Scholar]
  15. Kuiper, R.M.; Ryan, O. Meta-analysis of lagged regression models: A continuous-time approach. Struct. Equ. Model. A Multidiscip. J. 2020, 27, 396–413. [Google Scholar] [CrossRef] [Green Version]
  16. Schoenherr, D.; Paulick, J.; Strauss, B.M.; Deisenhofer, A.K.; Schwartz, B.; Rubel, J.A.; Altmann, U. Identification of movement synchrony: Validation of windowed cross-lagged correlation and-regression with peak-picking algorithm. PLoS ONE 2019, 14, e0211494. [Google Scholar] [CrossRef]
  17. Keele, L.; Kelly, N.J. Dynamic models for dynamic theories: The ins and outs of lagged dependent variables. Political Anal. 2006, 14, 186–205. [Google Scholar] [CrossRef]
  18. Adhvaryu, P.; Panchal, M. A review on diverse ensemble methods for classification. IOSR J. Comput. Eng. 2012, 1, 27–32. [Google Scholar] [CrossRef]
  19. Okiy, S.; Nwobi-Okoye, C.C.; Igboanugo, A.C. Transfer Function Modelling: A Literature Survey. Res. J. Appl. Sci. Eng. Technol. 2015, 11, 1265–1279. [Google Scholar] [CrossRef]
  20. Box, G.E.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  21. Davidson, R.; MacKinnon, J.G. Estimation and Inference in Econometrics. OUP Catalogue; Oxford University Press: Oxford, UK, 1993. [Google Scholar]
  22. Murphy, K.M.; Topel, R.H. Estimation and inference in two-step econometric models. J. Bus. Econ. Stat. 2002, 20, 88–97. [Google Scholar] [CrossRef]
  23. Timmermann, A. Forecast combinations. Handb. Econ. Forecast. 2006, 1, 135–196. [Google Scholar]
  24. De Vito, S.; Massera, E.; Piga, M.; Martinotto, L.; Francia, G. UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/ml/datasets/Air+quality (accessed on 28 January 2021).
  25. De Vito, S.; Massera, E.; Piga, M.; Martinotto, L.; Francia, G. On field calibration of an electronic nose for benzene estimation in an urban pollution monitoring scenario. Sens. Actuators B Chem. 2008, 129, 750–757. [Google Scholar] [CrossRef]
Figure 1. An example of the first simulation dataset. The bottom center is the target time-series.
Figure 1. An example of the first simulation dataset. The bottom center is the target time-series.
Sustainability 13 01367 g001
Figure 2. An example of the second simulation dataset. The bottom center is the target time-series.
Figure 2. An example of the second simulation dataset. The bottom center is the target time-series.
Sustainability 13 01367 g002
Figure 3. Eight time-series in air quality datasets are displayed. The bottom center is the target time-series, C6H6(GT).
Figure 3. Eight time-series in air quality datasets are displayed. The bottom center is the target time-series, C6H6(GT).
Sustainability 13 01367 g003
Figure 4. Eight time-series in the APM dataset. The bottom center is the target time-series, Response.
Figure 4. Eight time-series in the APM dataset. The bottom center is the target time-series, Response.
Sustainability 13 01367 g004
Table 1. RMSE and computing time for datasets from the first simulation.
Table 1. RMSE and computing time for datasets from the first simulation.
Target TimeIteration ARIMAARIMAXANNANNXRNNVAREDT-rEDT-w
t + 1 1sttime0.015.690.33.022.450.010.361.98
RMSE0.52450.36070.47050.35950.48100.13560.76190.8498
2ndtime0.534.760.9218.494.020.230.230.81
RMSE0.58240.40020.52170.43610.55580.18150.80430.4107
3rdtime0.070.780.272.580.610.060.780.53
RMSE0.57710.58230.56600.56490.73060.24951.46170.5189
t + 3 1sttime0.015.690.33.022.450.010.260.72
RMSE0.61630.43410.54950.44510.46680.28780.56420.3754
2ndtime0.534.760.9218.494.020.230.280.86
RMSE0.63480.46270.65160.49020.54360.39680.85470.1924
3rdtime0.070.780.272.580.610.060.470.36
RMSE0.62410.72740.63060.64360.71850.45980.90370.2411
t + 6 1sttime0.015.690.33.022.450.010.221.13
RMSE0.66040.41910.63380.44390.46370.36780.62020.3311
2ndtime0.534.760.9218.494.020.230.320.75
RMSE0.66380.47470.67930.53540.54690.44380.72950.2470
3rdtime0.070.780.272.580.610.060.260.38
RMSE0.71000.79610.72490.73210.72080.60280.98210.2087
t + 12 1sttime0.015.690.33.022.450.010.171.61
RMSE0.67810.45260.71370.52240.46580.42450.61700.2047
2ndtime0.534.760.9218.494.020.231.192.32
RMSE0.65600.50390.73860.57460.55190.54200.68070.1731
3rdtime0.070.780.272.580.610.060.280.2
RMSE0.75650.76350.79580.78510.71630.65550.92840.4405
Table 2. RMSE and computing time for datasets from the second simulation.
Table 2. RMSE and computing time for datasets from the second simulation.
Target TimeIteration ARIMAARIMAXANNANNXRNNVAREDT-rEDT-w
t + 1 1sttime0.255.120.813.932.830.151.362.44
RMSE0.00490.00660.00520.00650.20600.00550.05890.0537
2ndtime1.455.173.2210.673.260.80.391.89
RMSE0.04150.04240.09300.05020.56860.04380.16120.0863
3rdtime0.110.780.221.281.250.090.490.56
RMSE0.03220.03510.03750.04330.39660.03280.13680.0982
t + 3 1sttime0.255.120.813.932.830.150.652.31
RMSE0.00790.00920.00910.01000.16460.00780.01180.0040
2ndtime1.455.173.2210.673.260.80.281.48
RMSE0.07370.08270.17990.10460.56830.08690.19820.0943
3rdtime0.110.780.221.281.250.090.370.6
RMSE0.03980.05300.05020.06330.39350.05130.08550.0501
t + 6 1sttime0.255.120.813.932.830.150.412.03
RMSE0.01210.01290.01470.01410.16260.01080.02630.0026
2ndtime1.455.173.2210.673.260.80.383.19
RMSE0.12980.14990.26380.19210.57740.16300.42440.0701
3rdtime0.110.780.221.281.250.090.50.39
RMSE0.05380.07850.07200.08800.39660.07960.16950.0304
t + 12 1sttime0.255.120.813.932.830.151.162.69
RMSE0.01930.01630.02380.01860.16110.01540.03300.0030
2ndtime1.455.173.2210.673.260.80.51.38
RMSE0.23670.21900.35310.28280.59280.24700.31310.0794
3rdtime0.110.780.221.281.250.090.430.44
RMSE0.07470.12160.10380.13030.40930.12490.20650.0406
Table 3. RMSE and computing time for air quality data.
Table 3. RMSE and computing time for air quality data.
Target TimeIteration ARIMAARIMAXANNANNXRNNVAREDT-rEDT-w
t + 1 5000 to 6000time0.060.710.222.790.510.040.470.53
RMSE0.42820.42590.43120.41341.11390.45610.66090.4821
6000 to 7000time0.030.70.251.860.770.020.290.41
RMSE0.33460.33940.35370.33520.98680.35790.45700.2243
7000 to 8000time0.050.470.311.340.560.030.370.6
RMSE0.30390.30670.31940.30820.95220.32150.52610.3211
8000 to 9000time0.070.730.87.70.460.040.30.56
RMSE0.30250.30660.26810.25450.81330.29560.79460.1912
t + 3 5000 to 6000time0.060.710.222.790.510.040.470.41
RMSE0.83770.85620.85590.79541.10690.49671.75840.3451
6000 to 7000time0.030.70.251.860.770.020.390.48
RMSE0.65350.66030.70720.64950.98270.68221.64100.2156
7000 to 8000time0.050.470.311.340.560.030.40.53
RMSE0.59410.60210.60370.59780.94460.62521.12720.1168
8000 to 9000time0.070.730.87.70.460.040.340.59
RMSE0.55800.58410.45080.44650.78820.53841.32600.1284
t + 6 5000 to 6000time0.060.710.222.790.510.040.340.53
RMSE1.02021.05901.09061.02821.11811.00861.12110.5270
6000 to 7000time0.030.70.251.860.770.020.370.52
RMSE0.85210.87110.97100.89720.98290.83850.83620.2502
7000 to 8000time0.050.470.311.340.560.030.540.48
RMSE0.78230.79590.77160.75970.94630.78990.71070.2854
8000 to 9000time0.070.730.87.70.460.040.450.57
RMSE0.61620.64930.53820.50990.79190.61140.61270.1762
t + 12 5000 to 6000time0.060.710.222.790.510.040.310.41
RMSE1.05531.07761.19291.09211.11871.23600.94300.4948
6000 to 7000time0.030.70.251.860.770.020.360.68
RMSE0.91820.95181.17171.03770.98080.96360.82820.2899
7000 to 8000time0.050.470.311.340.560.030.590.6
RMSE0.86720.87160.87010.82080.95630.96390.72100.4015
8000 to 9000time0.070.730.87.70.460.040.50.54
RMSE0.61220.63930.57210.54110.79870.57850.64550.2702
Table 4. Summary of data attributes in the APM data.
Table 4. Summary of data attributes in the APM data.
Log DataDescription
Used memorymemory occupancy for a target application
System CPUCPU occupancy for a target application
User CPUCPU occupancy for all application
CountThe number of simultaneous access
Num httpThe number of thread for a target application about http requests
Num javadThe number of thread for a target application about javad
Num jdbcThe number of thread for a target application about jdbc
ResponseResponse time of a target application to a request of users
Table 5. RMSE and computing time for APM data.
Table 5. RMSE and computing time for APM data.
Target TimeIteration ARIMAARIMAXANNANNXRNNVAREDT-rEDT-w
t + 1 1st non-stationarytime0.164.230.252.391.750.061.941.14
RMSE0.10450.04120.02320.04230.34340.01070.10030.0909
2nd non-stationarytime0.196.430.1314.962.360.110.720.72
RMSE0.02630.02000.01000.03100.30710.00400.05110.0348
3rd non-stationarytime0.11.360.181.60.510.010.340.44
RMSE0.10380.37720.11680.15081.35400.05070.42860.3403
4th non-stationarytime0.141.340.311.30.450.020.390.23
RMSE0.14570.12970.03120.06280.42850.01530.13440.1257
1st stationarytime0.141.270.241.670.780.010.270.37
RMSE0.00140.00140.00020.00030.15550.00030.00750.0075
2nd stationarytime0.141.280.131.431.220.080.380.34
RMSE0.00180.00160.00030.00030.16410.00030.00810.0082
3rd stationarytime0.091.280.161.592.720.050.371.69
RMSE0.00140.00130.00030.00040.16510.00030.00760.0076
4th stationarytime0.151.040.21.640.350.070.180.48
RMSE0.00130.00110.00040.00050.17100.00050.00770.0077
t + 3 1st non-stationarytime0.164.230.252.391.750.060.652.24
RMSE0.16930.13130.15700.20120.28590.12800.16620.1558
2nd non-stationarytime0.196.430.1314.962.360.110.170.56
RMSE0.07160.06440.08390.08980.25570.04040.13630.1286
3rd non-stationarytime0.11.360.181.60.510.010.20.56
RMSE0.45010.64500.71570.72661.29540.66881.12111.0639
4th non-stationarytime0.141.340.311.30.450.020.360.27
RMSE0.23250.27910.20830.28990.35100.19280.21180.1886
1st stationarytime0.141.270.241.670.780.010.140.42
RMSE0.00600.00640.00220.00260.11350.00240.01300.0112
2nd stationarytime0.141.280.131.431.220.080.180.43
RMSE0.00710.00610.00250.00290.11670.00240.01280.0122
3rd stationarytime0.091.280.161.592.720.050.532.7
RMSE0.00530.00500.00260.00300.11700.00250.02300.0237
4th stationarytime0.151.040.21.640.350.070.190.43
RMSE0.00490.00460.00280.00340.11970.00330.01740.0171
t + 6 1st non-stationarytime0.164.230.252.391.750.061.350.51
RMSE0.26170.31440.32080.32040.28970.89600.30720.1404
2nd non-stationarytime0.196.430.1314.962.360.110.270.6
RMSE0.14700.12870.13940.13110.25570.22620.11410.0519
3rd non-stationarytime0.11.360.181.60.510.010.160.27
RMSE1.21911.07481.57191.88001.29837.44171.28831.1557
4th non-stationarytime0.141.340.311.30.450.020.280.25
RMSE0.35110.53300.44260.38290.35001.61000.27390.2432
1st stationarytime0.141.270.241.670.780.010.250.35
RMSE0.01440.01500.00890.00990.11300.00920.01550.0080
2nd stationarytime0.141.280.131.431.220.080.30.39
RMSE0.01600.01360.00950.01110.11540.00920.01820.0087
3rd stationarytime0.091.280.161.592.720.052.162.58
RMSE0.01280.01180.00980.01090.11860.00910.01850.0090
4th stationarytime0.151.040.21.640.350.070.280.23
RMSE0.01250.01220.01040.01170.11750.01290.01830.0101
t + 12 1st non-stationarytime0.164.230.252.391.750.060.670.49
RMSE0.37180.55710.54940.52920.290714.66590.25390.2141
2nd non-stationarytime0.196.430.1314.962.360.110.390.68
RMSE0.22630.18130.28710.16490.26072.48230.11900.1212
3rd non-stationarytime0.11.360.181.60.510.010.180.36
RMSE2.75561.80324.05936.15311.29954286.30251.51281.8513
4th non-stationarytime0.141.340.311.30.450.020.210.39
RMSE0.48620.80090.57620.58870.355446.56820.34110.2814
1st stationarytime0.141.270.241.670.780.010.240.31
RMSE0.02150.02030.01690.01910.11310.01540.01540.0144
2nd stationarytime0.141.280.131.431.220.080.370.36
RMSE0.02120.01900.01670.01840.11800.01580.01560.0143
3rd stationarytime0.091.280.161.592.720.051.351.45
RMSE0.01820.01790.01770.01850.11670.01580.01720.0152
4th stationarytime0.151.040.21.640.350.070.20.27
RMSE0.01850.01900.01850.02060.11620.80720.01660.0159
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kong, T.; Choi, D.; Lee, G.; Lee, K. Air Pollution Prediction Using an Ensemble of Dynamic Transfer Models for Multivariate Time Series. Sustainability 2021, 13, 1367. https://doi.org/10.3390/su13031367

AMA Style

Kong T, Choi D, Lee G, Lee K. Air Pollution Prediction Using an Ensemble of Dynamic Transfer Models for Multivariate Time Series. Sustainability. 2021; 13(3):1367. https://doi.org/10.3390/su13031367

Chicago/Turabian Style

Kong, Taewoon, Dongguen Choi, Geonseok Lee, and Kichun Lee. 2021. "Air Pollution Prediction Using an Ensemble of Dynamic Transfer Models for Multivariate Time Series" Sustainability 13, no. 3: 1367. https://doi.org/10.3390/su13031367

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop