Next Article in Journal
Emotions, Illness Symptoms, and Job Satisfaction among Kindergarten Teachers: The Mediating Role of Emotional Exhaustion
Previous Article in Journal
Simulation of Freshwater Ecosystem Service Flows under Land-Use Change: A Case Study of Lianshui River Basin, China
Previous Article in Special Issue
Comparison of H2S Gas Sensors: A Sensor Management Procedure for Sewer Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Attention-Based Distributed Deep Learning Model for Air Quality Forecasting

School of Computer Science and Engineering, Pusan National University, Busan 46241, Korea
*
Author to whom correspondence should be addressed.
Sustainability 2022, 14(6), 3269; https://doi.org/10.3390/su14063269
Submission received: 10 January 2022 / Revised: 25 February 2022 / Accepted: 3 March 2022 / Published: 10 March 2022
(This article belongs to the Special Issue Real-Time Air Quality Monitoring Systems)

Abstract

:
Air quality forecasting has become an essential factor in facilitating sustainable development worldwide. Several countries have implemented monitoring stations to collect air pollution particle data and meteorological information using parameters such as hourly timespans. This research focuses on unravelling a new framework for air quality prediction worldwide and features Busan, South Korea as its model city. The paper proposes the application of an attention-based convolutional BiLSTM autoencoder model. The proposed deep learning model has been trained on a distributed framework, referred to data parallelism, to forecast the intensity of particle pollution ( P M 2.5 and P M 10 ). The algorithm automatically learns the intrinsic correlation among the particle pollution in different locations. Each location’s meteorological and traffic data is extensively exploited to improve the model’s performance. The model has been trained using air quality particle data and car traffic information. The traffic information is obtained by a device which counts cars passing a specific area through the YOLO algorithm, and then sends the data to a stacked deep autoencoder to be encoded alongside the meteorological data before the final prediction. In addition, multiple one-dimensional CNN layers are used to obtain the local spatial features jointly with a stacked attention-based BiLSTM layer to figure out how air quality particles are correlated in space and time. The evaluation of the new attention-based convolutional BiLSTM autoencoder model was derived from data collected and retrieved from comprehensive experiments conducted in South Korea. The results not only show that the framework outperforms the previous models both on short- and long-term predictions but also indicate that traffic information can improve the accuracy of air quality forecasting. For instance, during P M 2.5 prediction, the proposed attention-based model obtained the lowest MAE (5.02 and 22.59, respectively, for short-term and long-term prediction), RMSE (7.48 and 28.02) and SMAPE (17.98 and 39.81) among all the models, which indicates strong accuracy between observed and predicted values. It was also found that the newly proposed model had the lowest average training time compared to the baseline algorithms. Furthermore, the proposed framework was successfully deployed in a cloud server in order to provide future air quality information in real time and when needed.

1. Introduction

Industrialization has brought us luxuries in life but, at the same time, many challenges too. Air pollution is one of the biggest challenges we are facing in the world today. The biggest contributors to air pollution each year are industries excreting dangerous gases, rapid urbanization, exponential fuel vehicles, and fossil fuel consumption. Air quality is getting worse. Several research reports have shown that higher atmosphere contamination levels and harmful particles impact the population’s health, resulting in respiratory infections, cardiovascular diseases, and lung cancer. Therefore, air pollution research is critical and is consistently viewed as an important topic in environmental protection. There is a growing need to evaluate and forecast the quality of air because, if people already know about air quality, they can take more precautions to prevent disease. Predicting air quality is also critical for emergency management in any administration. Projections aid the government in taking suitable contingency measures to minimize air pollution, such as limiting motor vehicles and lowering emissions from highly polluting companies. Unfortunately, predicting air quality is a difficult process, and improving forecast accuracy and reducing training time are urgent and difficult problems that need to be addressed in air pollution prevention.
Air quality prediction is a challenging task, but a lot of research has been done over the last decade to predict and address air quality challenges. This study uses an experimental setup to focus on two different forms, or models, of air quality forecasting. A knowledge-based model is one type, whereas a data-driven model is another. Knowledge-based paradigms focus on physical and chemical theories to demonstrate the transmission and deflection of air pollution particles. Several proposed knowledge-based methods have been studied in the literature survey to evaluate their impact. As expected, the successful implementation of knowledge-based methods involves intensive knowledge and research on environmental and atmospheric sciences. Using a knowledge-based model under varied conditions can also cause the chemistry and transfer rules to alter, resulting in inaccurate findings. Some scholars in the literature have employed statistical prediction approaches, including Hidden Markov Models [1], ARIMA models [2], and the LASSO model [3], to overcome this problem. The statistical prediction technique that uses mathematical reasoning and regression analysis, on the other hand, has two fundamental drawbacks: (1) a lack of precision and (2) time and energy consumption that is introduced from the examination of long-term historical monitoring data. In conclusion, statistical forecasting approaches [4] may accurately predict air quality, but a variety of air pollution factors make accurate forecasting challenging.
The other approach for air quality prediction involves evaluating data and is known as the data-driven model. Thanks to the big data analysis and artificial intelligence boom, this approach has recently been examined and utilized in various forecasting frameworks. The machine learning models used in this air quality prediction have been helpful to overcome the statistical and numerical drawbacks of previous approaches. These algorithms have established themselves as the backbone of air quality forecasting research. Machine learning methods for estimating air quality have so far yielded positive results. Wang et al. [5] proposed an online SVM model to forecast air pollution concentrations in downtown Hong Kong. They ran a comparative experiment between standard SVM and online SVM. Results showed that the online SVM proposed by Wang et al. [5] outperformed the standard SVM. In research performed in 2018, the air quality prediction in Murcia was conducted to predict ozone levels [6]. Researchers used different traditional machine learning models, out of which the random forest model performed the best [6]. Four regression techniques were used to predict air quality using machine learning algorithms, namely regression ANN, MLP regression, decision trees, and random forest regression [7]. This comparative study determined the best approach benchmark as processing time and data size. Random forest regression outperformed other algorithms as it optimally predicted air quality with various data sets, sizes, locations, and features.
Deep learning (DL) is commonly used in large data analysis to tackle a variety of issues, such as object recognition [8,9], speech recognition [10], classification [11], and prediction based on time series data [12]. The forecasted accuracy and efficiency of air quality is enhanced with the use of deep learning approaches. With the help of deep learning approaches [13], different air quality features can be obtained using the data-driven method. A wide range of articles in the literature have generated positive outcomes when employing deep learning approaches for air pollution prediction [14]. Zhao et al. [15] introduced a deep learning model named LSTM-Fully Connected Neural Network to predict the concentration level of particle pollution among specific monitoring stations over 48 h. Researchers used previous air quality data as well as meteorological and weather forecast data. Finally, a data collection of records from 36 air quality monitoring stations was used to test the suggested approach. They made a comparative study between ANN and LSTM models. Qi et al. [16] used the deep learning approach to propose a hybrid model incorporating both graph CNN and LSTM networks (GC-LSTM) to model and predict the spatial-temporal fluctuation of particles concentrations. The authors created historical observations as spatiotemporal graph series, in which previous meteorological variables and air quality attributes were defined as graph signals. To evaluate their proposed system, they compared it with the existing methodologies and noticed that their proposed model outperformed the conventional approaches. To forecast the air pollution levels, Wen et al. [17] proposed a convolutional LSTM model. They combined a CNN model with an LSTM network to extract the spatiotemporal features. Model performance was also improved by incorporating meteorological and aerosol data. Authors in [18] conducted another study in which they used Convolutional Neural Network and LSTM to predict the air quality. The evaluation criteria included RMSE, MAE, agreement index, and Pearson correlation coefficient.
An hourly concentration of air quality forecasting system was proposed by Bai et al. [19]. The authors used a deep-stacked autoencoding algorithm. It combined the seasonal air quality analysis and feature learning to predict optimal air quality pollution. Their approach was evaluated based on the real-time dataset from three monitoring stations located at Beijing. A. Heydari et al. [20] proposed a hybrid model based on LSTM and multi-verse optimization algorithm to predict the concentration level of pollutants S O 2 and N O 2 . The authors used a dataset collected from May to September 2019. The dataset included four features: wind speed, temperature, S O 2 , and N O 2 . The result showed the superiority of their approach against four other optimization-based algorithms. A multi-task learning approach based on GRU was proposed in [21] to predict air pollution particles, namely P M 2.5 , P M 10 , and N O 2 . The authors formalized the particles pollution forecasting into a multitask architecture. They showed the superiority of their framework by comparing it with seven different baselines. Xiao et al. [22] proposed a weighted LSTM model to predict the concentration level of P M 2.5 . The authors first generated the historical P M 2.5 time series data by using a sites distances feature, wind condition, and pollution concentration as input to a proposed MLP model. The experiment’s results demonstrated that the proposed weighted LSTM model can improve the prediction of the P M 2.5 particle.
Yue-Shan Chang et al. [23] proposed an aggregated LSTM model to predict the P M 2.5   particle. Their research used a dataset based on 17 attributes which was collected from 2012 to 2015 by the Taiwan Environmental Protection Agency. In the experiments, the authors predicted P M 2.5 for the next 1–8 h. The results showed the superiority of the proposed approach against SVR, GBT, and a standard LSTM model. Another RNN based model was proposed in [24] to forecast the concentration level of P M 10   at different future time steps (6, 12, and 24 h). On the other hand, Guo et al. [25] proposed a feature engineering pipeline as well as a deep ensemble network algorithm which combines RNN, LSTM, and GRU networks to predict the P M 2.5 concentration of the next hour. During the data analysis step, the authors used the Pearson correlation coefficient to evaluate the correlation of P M 2.5 with meteorological data, season data, and time stamp data. The experiments showed that the proposed model surpassed the performance of the benchmark algorithms.
Attention mechanism is a learning concept which is based on mimicking human behavior by directing attention toward particular events. This notion of imitating human attention started a few years ago with computer vision studies [26,27]. The main goal was to decrease the computational complexity of image processing while enhancing performance by introducing an approach that would only focus on particular regions of images instead of the entire input frame. However, the attention mechanism had been utilized more in depth in the natural language processing field by Bahdanau et al. [28]. The authors in this article proposed an attention-based approach for a machine translation model. Today, attention mechanism has become an important part of deep learning models which is widely used in object detection [29], medical image segmentation [30], time series forecasting [31], natural language processing [32], question-answering [33], and more [34,35,36]. Recently researchers have started to use attention mechanism for air quality forecasting as well. For instance, in [37], the authors used a variational autoencoder with multiple directed attention to predict the concentration of ambient pollutants, namely N O 2 , O 3 , S O 2 ,   and   C O . They evaluated their model by using six different evaluation metrics and outperformed the baseline models. Xiangyu et al. [38] proposed an LSTM-based attention algorithm for air quality prediction. They evaluated their model on a Beijing dataset and then compared it against six baseline algorithms: ARIMA, MFSVR, DeepST, LSTM, GC-LSTM, and ADAIN. The presented results showed how their models outperformed the baselines based on RMSE and R 2 evaluation. In [39], the authors proposed a seq2seq attention-based model for air quality forecasting. They replaced the RNN encoder with a pure attention mechanism with position embeddings. The proposed model was implemented at two different stations in Beijing city: the Dongsi and Olympic center monitoring stations. The experiments results showed that the attention-based GRU seq2seq model was more effective at predicting the particle P M 2.5   for the next 24 h. In the same vein, in [40], a forecasting approach based on dual LSTM was proposed. The authors first obtained each component by using a seq2seq method. Furthermore, they implemented an attention-based LSTM network as a multi-factor forecasting model which was afterwards connected to the seq2seq technology by using an Xgboosting tree in order to make the final prediction. On the other hand, Chen et al. [41] proposed an extreme value attention model based on an autoencoder network for air pollution forecasting. In order to capture long-term dependencies within the feature set, the authors used a temporal attention mechanism at the decoder level.
There have been numerous studies to address air quality forecasting with AI-based models, especially with machine learning and deep learning algorithms. Table 1 provides an overview of some previous approaches to solve the problem of air pollution forecasting.
Recently, the literature has showed that the hybrid model and the attention-based approach are very effective in predicting the air quality; however, they face some challenges:
  • Slow training speed: This issue is caused by the huge amount of data that needs to be trained on a centralized deep learning architecture. In previous research, the authors used a centralized deep learning approach for training. This problem with slow training speed can be bothersome when the data changes. Therefore, models need to be retrained for more accurate predictions.
  • Noisy data: The second disadvantage of these methods is that they dismiss noise effect. Indeed, noise effect is another potential challenge to evaluating air quality and meteorological data. Noise affects the accuracy and performance of the forecasts because the algorithms do not extract the optimal feature and information from pollutant and meteorological data.
  • No real-time traffic data: The third challenge is that these models do not count traffic information in their set of features during the training phase. It is widely recognized that traffic flow is a significant source of air pollutants that could damage air quality. Not considering traffic information while predicting air quality can lead to nonrealistic forecasting.
These challenges show a need to have a more optimal deep learning model to evaluate air quality. We first conducted a prior study for air quality forecasting in Busan City [14]. Compared to our previous work, in this research we propose a deep learning model focusing on the attention-based convolutional BiLSTM autoencoder framework to forecast air quality. A distributed architecture (data parallelization) has been implemented during the training. The dataset used in this research is based on three specific feature sets: particle features, meteorological features, and car traffic-related features. The major contribution of this research includes:
  • The review of new deep learning and machine learning approaches for air quality prediction alongside attention mechanism.
  • The implementation of YOLOv5 model with DeepSort algorithm to collect traffic information data on the road.
  • An extensive feature engineering to find the most important features while predicting both particles P M 2.5 and P M 10 . Knowing a feature’s importance is an essential step in prediction tasks. Machine learning models have been widely used in the literature to find the most relevant features in a given dataset. In this research, we made a comparative study with four different machine learning algorithms to perform this task.
  • The development of an innovative attention-based convolutional BiLSTM autoencoder model to predict air quality. The 1D-CNN layers extract deep spatial correlation and local patterns features from the air pollutants. The autoencoder model encodes the meteorological and the traffic data. The Bi-LTSM layer interprets the obtained features, then passes them to several attention layers to make the final optimal prediction.
  • The evaluation of the proposed framework which is based on two specific steps. Firstly, we trained the deep learning framework in centralized architecture using a single training server. This trained model has been compared with seven state-of-the-art algorithms, including two attention-based deep learning approaches. Secondly, we trained the proposed model using a parallelization training approach. We evaluated the improvement in its accuracy and the reduction of training time during this phase.
  • The creation of a deployment pipeline to run online inference based on the proposed deep learning model. We noticed that most of the previous studies only focused on offline learning and inference. In our research, we are not only considering offline prediction but also online prediction based on real-time data.
The structure of the rest of this research paper is as follows. In Section 2, we present the data collection and methods. In Section 3, we introduce our deep learning framework. We explain the data pre-processing step, the feature correlation and selection, as well as the proposed algorithm. The experiments are described in Section 4. The conclusion and future work are discussed in Section 5.

2. Data Collection and Methods

2.1. Pollution Particles and Meteorological Data Collection

This research has been conducted based on three main data sources, which are: (1) air quality particle data from Air Korea (Seoul, South Korea) [42], (2) meteorological information from the Korean Meteorological Agency (KMA, Seoul, South Korea) [43], and (3) real-time traffic data. Both (1) and (2) data sources were collected from Korean government websites and contain ten specific features, which are: P M 2.5 , P M 10 , N O 2 (Nitrogen dioxide), CO (Carbon monoxide), and S O 2 (Sulfur dioxide) for air pollution-related features, and temperature, rain precipitation, dew point, wind speed, and humidity for meteorological features. The traffic-based data were collected on the road using a computer vision-based approach explained in the following section.

2.2. Traffic Data Collection—Car Counting

One of the contributions of this research is the utilization of traffic data to improve the prediction of air quality. We collected the traffic data near the monitoring stations. Figure 1 shows a road near Namsan Station in Busan city where we collected the traffic data for the air quality station in Namsan-dong. We used a computer server with an NVIDIA Geforce RTX 2060 GPU and a camera with a resolution of 1080p 30 fps. The traffic count algorithm used the object detection model YOLOv5 [44] and the ‘DeepSort’ algorithm to track objects (cars) to prevent multiple counting for one object. Algorithm 1 presents the process of car counting using DeepSort.
Algorithm 1: Input Data verification
Input: Car /*Object that is tracked by Deep Sort Algorithm. */
1. IF Car > diff /*diff: difference reference line and object’s center point. */
2.  FOR each entry tl TL do
3.    IF Car in TL /*TL: Track List */
4.      THEN pass
5.    ELSE
6.      THEN Add Car to TL
7.  END
8. RETURN length(TL)
END Car Counting
The YOLOv5 is the fifth model of You Only Look Once (YOLO). Compared with the previous version of YOLO and other object detection models, YOLOv5 considerably improves the training speed while maintaining a good detection accuracy. The model divides the input images into grid cells and creates a bounding box for the objects. The score of each grid is calculated by multiplying the probability that an object exists in the bounding box on the grid, intersection over union (IOU), and the probability that the object is the corresponding class. By comparing the score with the threshold, the model can define from which class the bounding box belongs to. The DeepSort algorithm predicts the next location of the object by applying the location information of the object detected in the previous frame with the Kalman filter. The IOU distance of the currently detected object and the object predicted in the previous frame is calculated in the next frame. By applying the calculated result to the Hungarian algorithm, we determined whether it is the same object or not.
The DeepSort algorithm cannot track the object properly if the boundary box is smaller than a specific size. Therefore, as shown in Figure 2a, a virtual baseline is set to apply the algorithm only to the size of the objects. If the center of the bounding box that indicates the vehicle is located below the reference line, the count increases. To prevent multiple counting of a single object, the algorithm keeps the bounding box’s index number in a temporal array and compares it with the new frame’s bounding boxes. The union of YOLOv5 and DeepSort algorithm is used to recognize an object that has been detected before. For example, if Object 1 was already detected in Frame 1, the algorithm will not consider it again if it comes in Frame 2 or 3, as shown in Figure 2b.

3. Proposed Framework

The proposed framework workflow is shown in Figure 3. Its major phases are comprised of the following processes: (i) pre-processing data; (ii) features correlation; (iii) deep learning models formulation; (iv) hyperparameter fine-tuning; and (v) model evaluation.

3.1. Data Preprocessing

There are over 490 air quality monitoring stations in South Korea. Data from several stations encounter the problem of missing values due to some errors with the sensors. To solve this problem, we used a linear interpolation technique. Linear interpolation imputed the missing values using the mean value of the last and first available in the dataset. It is based on the following formula [45]:
f x = f x 0 + f x 1 f x 0 x 1 x 0 × x x 0
In the presented equations, x represents the independent value and x 0 and x 1 are the known values of the independent variable. The air quality is impacted by several factors, including meteorological and traffic data that have their physical properties and dimensions. With accurate analysis and seamless training, these factors can be helpful in identifying the air quality. We have combined all data sources into a single dataset in our framework during the data pre-processing step. Each dataset is obtained by normalizing it using linear scaling based on the following equation [46], where x i represent the normalized data.
x i = x x m i n   x m a x x m i n

3.2. Data Correlation and Analysis

Optimal air pollutants prediction can be achieved by a model that must quantify and evaluate various air quality indexes. The motivation of this research is to predict the intensity of air pollutant particles, namely P M 2.5 and P M 10 . We extend our understanding by finding the relation of these particles with other features, especially meteorological and traffic data. Findings showed that the particles P M 2.5 and P M 10 can be impacted by many measurable factors. However, the challenging part was separating the data that is not beneficial for the prediction task, as irrelevant factors are burdensome for the model and training phase. It was important to find a correlation between different relevant and important particles with meteorological data. To overcome this, a correlation coefficient ( δ ) was calculated to each feature and the target particle based on the following formula [47], supposing that we had an observation vector A = ( a 1 , a 2 ,   a 3 , ,   a n ) with another vector B = ( b 1 , b 2 ,   b 3 , ,   b n ):
δ = n i = 1 n a i b i i = 1 n a i i = 1 n b i n i = 1 n a i 2 i = 1 n a i 2 n i = 1 n b i 2 i = 1 n b i 2
The prior formula presents a positive correlation when 0 <   δ < 1 and a negative correlation when −1 <   δ < 0. Similarly, in a scenario where the value of δ is closer to 1, a small variation between A and B is observed, along with a higher correlation. Figure 4 shows the feature correlation heatmap of datasets used in this research analyzing particles such as P M 2.5 and P M 10 as targets. As shown in Figure 4, P M 2.5 and P M 10 particles are negatively correlated with temperature and wind speed, so in low-temperature conditions, these particles will be more concentrated in the air, resulting in worse air quality prediction. If we consider the wind speed factor, with faster or higher wind speed, the lower concentration of the pollution particles results in good air quality. With humidity factors, the concentration of pollution particles is higher, resulting in bad air quality. Other gases which impact the higher concentration levels of harmful particles relate to high levels of N O 2 , CO, S O 2 , and a low level of O 3 . On the other hand, the traffic data is highly correlated with both particles, which makes us think that high car traffic on a specific location can produce air pollution. Based on this feature correlation heatmap, we can clearly see how the traffic information impacts levels of both particles, as they are highly correlated.

3.3. Features Importance Assessment

Features importance is a significant step in forecasting tasks. It assesses the relevance of a feature from the given feature set F = f 1 ,   f 2 ,   f 3 , ,   f n ; the importance value of each feature is computed by using the permutation importance. In this research, a feature assessment was carried out to rule out the useless features that can increase algorithm complexity. Machine learning algorithms have been widely used in the literature to find the most important feature in a given dataset. In this study, we implemented four different machine learning models to perform this task: random forest, Xgboost, AdaBoost, and gradient boost.
As shown in Figure 5, all the algorithms except Xgboost algorithm selected the cars traffic as one of the most important features.
Pollution gas such as CO, N O 2 , and SO2 are also labelled as relevant features for the prediction of particle matter ( P M 2.5 and P M 10 ). We then evaluated the error rate of each of these models while predicting P M 2.5 in order to select the best features that will be used to train the deep learning algorithms presented in the experiments. Based on the results presented in Figure 6, we decided to focus only on the feature importance of random forest, Xgboost, and gradient, since these algorithms had the lowest error rate (MAE, RMSE, and SMAPE); this led us to remove the rain feature and keep the remaining ones for the next training step. One reason that explains the poor performance of AdaBoost in this experiment is the presence of noise in the air quality dataset. As discussed in previous studies [48,49], AdaBoost has been proven to be a very efficient ensemble learning model, which recurrently generates a set of diverse weak learners and combines their outputs using the weighted majority voting rule as the final decision. However, in some cases, AdaBoost leads to overfitting, especially when the training dataset is noisy, engaging in both its reduced generalization performance and non-robustness.

3.4. Deep Learning Model

The standard deep learning model’s effectiveness can be enhanced using different hybrid learning approaches. The proposed framework consists of a stacked deep autoencoder, a convolutional neural network, and an attention-based bi-directional LSTM layer. Together these layers of the framework calculate the concentration level of P M 2.5 and P M 10 based on data collected in Busan city. The deep features were extracted in two stages based on the proposed deep learning architecture. The first stage extracted all suitable features from both particles of P M 2.5 and P M 10 along with time series data, whereas a stacked autoencoder layer was used to encode meteorological, gas, and traffic-related patterns during the second stage. Figure 7 shows the architecture of the proposed model.

3.4.1. One-Dimensional CNN for Local Features Mining

Deep convolutional networks are special neural networks that are based on a convolution operation. Convolution operation is frequently used for extracting the frequency or temporal features from specific signals in machine learning. The convolution of two functions f t and g t can be represented as follows [50]:
f × g t = + f ϑ g t ϑ d ϑ
In the above equation, g t is called kernel and * represents the convolution operator. The time series data for air quality particles is fed into the 1D CNN network to extract local features of particles. Considering the time series of the i t h time window as:
X i = x i 1 ,   x i 2 , ,   x i l ,   x i     R d
where l represents the window size, x i t is the multivariate time series obtained at time step t and d represents the dimension. The temporal convolution layer is the first layer of the model. It is able to automatically extract local features of multivariate time series. In this setup, the convolution operation is represented as follows:
C τ = R e L U   w τ × x + b
where C τ is the result feature map of the τ t h kernel, w τ and b represent the weight and the deviation parameters, respectively. The 1D CNN model constantly moves the matching convolution step from the start of the time series to complete the convolution operation of the entire data. The features obtained from the convolution layer are represented as follows:
C i = C 1 ,   C 2 ,   C 3 , ,   C l m + 1
In the above equation, m denotes the size of the convolution kernel. As depicted in Figure 8, which represents the overall diagram of one-dimensional convolution computing, the input size is represented by n × l × d , in which n denotes the number of input samples, the output feature map size is computed by l m s + l × τ   , with τ representing the number of kernels. In order to complete the convolution operation at each time step to obtain the local features, we set the kernel size of m to 1.
The detailed parameter settings of our 1D CNN layers are presented in Table 2.

3.4.2. Deep Autoencoders Model for Traffic and Meteorological Data Encoding

Air pollution forecasting requires several crucial indicators to be properly monitored. For example, to predict the particles P M 2.5 and P M 10 , it is necessary to consider the traffic features and meteorological data, as they will help predict accurate air quality and precise decision-making. Failing to consider every relevant factor causes poor accuracy and bad decision-making. To improve the accuracy for predicting the concentration of P M 2.5 or P M 10 particles, we extract the relevant data from meteorological datasets and road traffic datasets. To achieve this, we implemented a deep autoencoder to encode information from such features. This autoencoder serves as a neural network with one hidden layer, which sets the output value equal to the input. This encoder compresses the input into the hidden layer, then reconstructs the output from that display. Autoencoder is an unsupervised model consisting of two steps: encoding and decoding. This keystroke enables this neural network to learn more abstract features unsupervised. To predict P M 2.5 or P M 10 particles, we present a vector representation for the traffic and meteorological data.

3.4.3. BiLSTM Layers for Spatiotemporal Features Representation

Statistical-based approaches such as ARIMA and traditional learning models do poorly in predictive tasks because they dismiss the long-term dependence on time series data. The LSTM model is a specific RNN which was proposed by Hochreiter et al. [51]. This model has been found to be the best solution to address the problem of long-term information dependence on time series data. RNN networks are based on a chained network module. In a standard RNN, this repeating module has a very basic structure, such as a tanh layer. LSTM networks have the same structure, but the core of the structure differs from standard RNN. As shown in Figure 9, LSTM networks have four modules playing different roles. The core of an LSTM is composed of an input gate i t , an output gate o t , and a forgetting gate f t .
The input gate decides the information that needs to be stored or updated in the cell. The input gate has a further two layers: the sigmoid layer responsible for deciding the value needed to be stored and the tanh layer responsible for building a new candidate value vector c ˜ t and insert it in the state. The input gate is based on the following equation [50]:
i t = σ W i   · x t   h t 1 + b i  
The forgetting gate decides what information must be deleted from the cellular state at t−1 time with h t 1 as the output value. It is based on the following equation [50]:
f t = σ W f   · x t   h t 1 + b f  
In this equation, W f is the weight matrix and b f is the bias matrix of the forget gate. h t 1 is the historical data, σ represents the activation function, and x t represents the current input of new data and determines which previous information is discarded. After getting the coefficient’s values of the input gate and the forget gate, the following equation [50] will be used to update current cell state:
c t = f t   °   c t 1 + i t   °   c ˜ t  
c ˜ t = t a n h W c x t + b c h t 1  
The output gate determines the current stage of output information. The sigmoid layer of output gates decides the section of cell state that will be output. The obtained value will be between −1 and 1 through the tanh layer and multiplied with the value of the first step output. The output gate is based on the following equations [50]:
o t = σ W o + x t h t 1 + b o  
h t = o t   °   tan h c t  
A disadvantage of the traditional LSTM model is that it is only operating in one way and may discard important information when extracting deep significant features. BiLSTM, on the other hand, is able to process time series data in a bidirectional way simultaneously. This leads to deeper functions for better features extraction and prediction. BiLSTM is formed with a combination of forward and backwards operations for LSTM (as shown in Figure 10). The sequence of information processing is performed under the forward state. The forward state LSTM output is completely isolated with the input of backward LSTM and vice versa. The BiLSTM gathers hidden LSTM states of opposite directions to the same output represented as follows [13]:
  y t = σ W   °     h t   , h t ,   x t + b  
The output layer receives information from both future and past states based on this architecture. The hidden states of the forward and backward layers are measured using the following equations:
h f = o f   °   tan h c f
h b = o b   °   tan h c b

3.4.4. Attention Layers

An attention layer is responsible for assigning weights to sample data and extracting essential information which can impact the features analysis and forecasting process. This process provides a better estimate of the final output and reduces the resource consumption of the computer. Moreover, it is also responsible for decreasing the redundant and noisy data while assigning them weights during the sampling and training phases. This mechanism is used to remember the long source of input. The attention layer creates the context vector, a weighted combination of input states to remember the important sequence while training. These weighted input states are different from the autoencoder model, which uses the context vector of the last hidden state. Therefore, the context vector can process the entire input sequence, thus eliminating the problem of forgetting. The Figure 11 shows the bidirectional LSTM architecture.
Consider a scoring function f defined: f : m × m     , which determines the assigned weight to its vector and indicates the important vectors. Therefore, the context vector ( c t ) is a weighted sum of hidden states H = { h 1 ,   h 2 , ,   h t 1 } and represents the important information for the current time step as depicted in the following equations [52]:
c t = i = 1 t 1 β i h i
β i = exp f h i ,   h t j = 1 t 1 exp f h j ,   h t
h i = h i T   ;   h i T T   ,   i = 1 , ,   n
The context vector is then unified with the previous hidden state unit S t 1 and previous target output y t 1 to predict the current hidden state S t as depicted in the following equation [52]:
S i = g S i 1 ,   y i 1   ,   C i

3.4.5. Distributed Training Approach

Research lacks evidence towards a deep distributed learning approach for air quality prediction, especially with training time and memory consumption investigations. We propose a novel approach by using distributed deep learning data parallelism to address these issues. It is achieved by distributing air quality traffic information and historical meteorological data across multiple training workers. Figure 12 shows the proposed training process.
We implemented a distributed training process called data parallelization with n worker machines to mitigate the training time and achieve significant computing performance. It helped us to achieve our training goals more quickly by processing n different partitions in the parallel data training approach. We distributed n copies of the model over n processing workers in this setup. The machines are trained locally under parallel mode to reduce the overall training time. A parameter server is used to collect the updates and request parameters from each training station.

4. Experimental Results and Discussion

In the following, we demonstrate the usefulness and the efficiency of the proposed deep learning framework by conducting extensive experiments.

4.1. Experiments Environment Setup

This section unravels the hardware and software environment of the experiments. The detailed configuration is presented in Table 3 and Table 4.
The proposed model has been built using python programming language. TensorFlow, a deep learning library, was used to implement our learning model and the other deep learning state-of-the-art algorithms. The library PySpark was used to create a distributed computing system. We made a comparative study between our model and seven state-of-the-art algorithms presented in Table 5.
In this experiment, the performance evaluations were carried out by calculating the Mean Absolute Error (MAE), Root Mean Square Error (RMSE), and Symmetric mean absolute percentage Error (SMAPE) [56].
M A E = 1 n i = 1 n y i y ^ i
R M S E = 1 n i = 1 n y i y ^ i 2
S M A P E = 1 2 n i = 1 n y i y ^ i y i + y ^ i × 100 %
However, our evaluations were divided into two steps. The first step was the training performed in a centralized deep learning environment. This approach was similar to other literature studies and evaluated by comparing it with other state-of-the-art models. The Algorithm 2 shows the steps of phase 1 experimentations.
Algorithm 2: Centralized training process
Input: Historical Data
Output: Error Rates
1. data ← MissingValuesFunc (Input)
2. data ← NormalizationFunc (data)
3. learning rate α = 10 3
4. Initialize F(x) = Proposed model for N pollutants
5. Split data into train, test and validation sets
6. For epoch = 1 to N do
7.   proposed_model ← fit_model (train, epoch, learningRate)
8. For t ← 1 … T do  
9.  Receive instance x t
10.   y ^ t ← forecast_model (f_model, x t )
11.   y t t e s t y
12.   Error_rate = evaluate_fit ( y t , y ^ t )
13. RETURN Error_rate
END
We used a distributed training environment called data parallelism to train our model in the second step. In this step, we employed ten worker machines to train ten different partitions of the air quality, meteorological, gas, and road traffic data in parallel. Each worker held a replica of the proposed algorithm. We then utilized a parameter server to aggregate the model’s updates and parameter requests coming from each worker.

4.2. Experiment Results

4.2.1. Step 1

We employed four input encoding layers and four decoding layers with Relu as the activation function. The deep learning autoencoder and decoder have several hidden layers; for autoencoder, the hidden layers were {128, 100, 64,32} and, for the decoder, they were {32, 64,100, 128}. The traffic and meteorological data were used to pre-train the algorithm; the resulting data was a compressed hidden layer of the encoder. To make the final prediction, this encoded layer of data was concatenated with the output layer of BiLSTM.
Table 6 and Table 7 summarize the proposed model’s comparative results with the different baseline algorithms. It is based on the observed P M 2.5 and P M 10 particles with the corresponding predicted values for 1 h, 2 h, 4 h, 8 h, 10 h, 12 h, 24 h, and 48 h time spans.
  • P M 2.5 forecasting results analysis: The experimental results show that our proposed framework outperformed the baseline models both on short- and long-term predictions. As can be seen, for the forecasting results based on both 1 h (short-term) and 48 h (long-term) time steps, the proposed attention-based model obtained the lowest MAE (5.02 and 22.59, respectively, for short-term and long-term predictions), RMSE (7.48 and 28.02) and SMAPE (17.98 and 39.81) among all the models. This indicates strong agreement between observed and predicted values. The recurrent neural network had the lowest prediction accuracy both on short- and long-term predictions. It had the highest MAE, RMSE, and SMAPE values, which were, respectively, 8.33, 12.17, and 25.11 for the short-term prediction and 30.08, 37.21, and 45.78 for the long-term prediction. All hybrid LSTM based models performed better than the standard LSTM, which recorded 7.96, 11.31, and 24.87, respectively, for MAE, RMSE, and SMAPE evaluation metrics for the short-term prediction, and 29.11, 36.92, and 46.03 for the long-term prediction. Among the hybrid LSTM baseline models, the attention-based CNN-LSTM recorded the second lowest error rates after the proposed approach; these second lowest error rates were 5.21, 7.40, and 18.15 for MAE, RMSE, and SMAPE for the short-term prediction and 22.63, 28.05, and 39.77 for the long-term prediction. The ConvBiLSTM autoencoder model came right after as it obtained 5.34, 7.92, and 19.27 for MAE, RMSE, and SMAPE for the short-term prediction and 23.96, 28.17, and 40.17 for the long-term prediction. The ConvBiLSTM autoencoder model was followed by the attention-based LSTM algorithm with 5.83, 8.14, and 20.16 for the short-term prediction and 24.14, 29.10, and 41.05 for the long-term prediction.
  • P M 10 forecasting results analysis: As depicted in Table 7, our proposed model still performs better than the baseline models both on short- and long-term forecasting. It achieved the minimal MAE (7.38 and 28.89, respectively, for 1 h and 48 h future time steps), RMSE (9.71, and 25.09, respectively, for 1-h and 48-h future time steps), and SMAPE (19.57, and 34.26, respectively, for 1-h and 48-h future time steps). On the other hand, the recurrent neural network captured the highest error rates, which were 12.20, 15.38, and 29.74, respectively, for the MAE, RMSE, and SMAPE metrics through 1h time step prediction and 37.19, 43.28, and 48.17 for the 48h time step prediction. The Attention CNN-LSTM model has the second-best result, as it recorded 8.15, 9.81, and 20.48, respectively, for the MAE, RMSE, and SMAPE metrics on the 1h time step prediction. For the 48 h time step, it achieved 26.39 for MAE, 25.93 for RMSE, and 34.33 for SMAPE. The ConvBiLSTM autoencoder recorded 8.29, 11.27, and 22.79 for MAE, RMSE, and SMAPE on 1h time step prediction, and on 48 h time step, it achieved 27.39, 31.09, and 39.83 for the same evaluation metrics. The other hybrid LSTM baseline models have similar results as they both outperformed the standard LSTM model on both 1 h (short-term) and 48 h (long-term) time steps.
The following images shown in Figure 13 illustrate how the proposed model outperformed all baseline algorithms through the eight different future time steps. In Figure 13, we have graphed each evaluation metric (MAE, RMSE, and SMAPE) for each short- and long-term prediction of both particles P M 2.5 and P M 10 . As shown in the figure, the performances of P M 2.5 and P M 10   long-term predictions are significantly lower than that of the short-term predictions. As the forward prediction size increases, the forecasting performances of these models gradually decrease. Nonetheless, even with long-term forecasting, the proposed model is still better than the baseline methods. This is mainly because the attention mechanism improves the long-term prediction performance of P M 2.5 and P M 10 .
A detailed comparison between the proposed deep learning air quality model was conducted with baseline models of RNN, LSTM, LSTM autoencoder, CNN+LSTM, Attention-based LSTM, ConvBiLSTM autoencoder, and attention-based CNN LSTM to forecast air quality particles of P M 2.5 and P M 10 . The factual results are presented in Figure 14 and Figure 15 as the comparison between all models to predict the air quality values for P M 2.5 and P M 10 . The x-coordinates represent observation time steps, and y-coordinates represent the P M 2.5 values in Figure 14 and P M 10 values in Figure 15. We decided to focus on 1-h short-term prediction only. As represented in these figures, the results show that the prediction pattern of the proposed deep learning air quality model is similar to the real pattern of time series data for both particles.
The proposed model performed better in forecasting as it predicted the pollution particles accurately and quickly, verifying the feasibility of our algorithm to capture the temporal variations, whereas other algorithms, especially RNN, were unable to track to the trends of P M 2.5 and P M 10 . The attention-based models also produced good forecasting trends. In this experiment, the wave peak period highlighted the difference of prediction and ground truth pattern more accurately.
Moreover, we evaluated the training time of all models as shown in Figure 16, and we found that the proposed framework had not only the best accuracy compared to the baseline models but it also had the best training average times alongside the Attention LSTM and the ConvBiLTM autoencoder model, as they all averaged 310/s during the training process for the prediction of both particles P M 2.5 and P M 10 .
The RNN model took the longest training time with 482 s, followed by the autoencoder LSTM with 479 s. The standard LSTM and the convolutional LSTM models had almost the same average training time. In step 2 we will show how the distributed architecture can reduce the training time of the proposed model and be further optimized to perform better.

4.2.2. Step 2

We distributed the architecture and employed ten specific training workers to perform the task to achieve better performance. Each training worker was provided with the same model parameters, and the algorithm was performing with the same capacity. We intended to predict the same particles using the same data as step 1. For this step, we only focused on the 1 h, 24 h, and 48 h future time step predictions to evaluate how the proposed model will perform both on short- and long-term predictions while being trained using ten worker nodes. For short-term prediction, as shown in Figure 17a,b, the proposed model performed the best for both particles P M 2.5 and P M 10 when it is using five and six workers, respectively, during the training process. As depicted in that figure, at five workers for P M 2.5 forecasting, we can see how the error rates decrease suddenly. The same observation can be seen at six workers during P M 10 forecasting. In Figure 17c,d, which correspond to t+24 h future time step forecasting, the proposed model achieved its best performance for both particles P M 2.5 and P M 10 while training it with seven workers. And finally, in Figure 17e,f, which represent to t + 48 h future time step prediction, the best performance results for both particle forecasting was recorded while using eight workers. The shape changes in the rate error are mostly due to the fact that the model had some difficulty generalizing well with a smaller number of training workers, and it was not stable at the beginning of the training. In further research, we are planning to investigate in depth all the potential causes of such changes, which may include the utilization of some optimization techniques. This experiment concluded that the right number of workers for the training of these two air quality particles should be between five and eight nodes.
We were interested in evaluating the reduction in training time with the introduction of distributed architecture and parallel training sessions. As shown in Figure 18, the training time of our framework decreased while the number of workers increased. At five training workers, we reached 250 s, and the training time increased slightly at six workers to 260 s, then decreased again to the lowest recorded time at eight workers, which was 200 s, and, from eight to 10, it was consistent. Compared to the centralized training in step 1, the training time in phase 2 almost decreased twice, making our method more effective and less time-consuming.

4.3. Proposed Model Deployment and Online Inference

With the goal of starting to use the proposed distributed deep learning model for practical decision-making, it needs to be effectively deployed into production. To do so, we created an online inference pipeline based on stream data, cloud server, and GPU instances, as depicted in Figure 19.
  • Input Stream Data: The air quality, meteorological, and traffic data are collected every hour and then preprocessed and stored inside cloud servers.
  • Online Worker: Since we decided to deploy our distributed model based on data parallelism, we have created several data pipelines based on Apache Spark that represent our workers. These training workers will have a model replica and a partition of the stream data. The workers will train their local replica by using the assigned data partition.
  • Parameter Server: The parameter server is responsible for aggregating model updates and parameter requests coming from different online workers.
  • User Interface: After the distributed training, the final model is then used to make the predictions every hour for the next 48 h and show it on the web application for the user to observe the predicted values as presented in Figure 20a. The web application also offers the option to upload a csv file containing the meteorological and road traffic features to predict air quality. This would help the government and the citizens to take all the necessary precautions to either control air pollution or protect themselves from the harmful effects of bad air quality.

5. Conclusions

This study proposed a new attention-based convolutional BiLSTM autoencoder model for air quality prediction. The approach utilized historical and time series data for traffic information, as well as encoded meteorological information in order to perform predictions of air quality at various time points. Car traffic data was collected by implementing a computer vision-based model called YOLO with DeepSort algorithm. Since the DeepSort algorithm could not track the approaching objects (cars) on the road properly if the boundary box was smaller, a virtual baseline or reference line was applied to the tracking algorithm. If the center of the bounding box indicated the vehicle was located below the reference line, the count increased. Multiple counting is prevented by keeping the previous bounding box’s index number in a temporal array and comparing it with the new frame’s bounding boxes. The proposed deep learning approach performed optimally in different atmospheric conditions, i.e., stable and unstable. This was achieved by using a two-stage feature extraction method. The first feature extraction stage was achieved by using a stacked autoencoder to extract information on several features, including pollution, traffic, and various meteorological properties. During the second stage, an attention-based convolutional BiLSTM layer extracted deep spatial correlation features from air quality particles data and performed the final prediction after concatenation with the encoded features from stage 1. The attention mechanism was used to capture the dynamic correlation of both P M 2.5 and P M 10 particles between previous and future time points. A distributed training method, data parallelism, was adopted to train the proposed model in the second phase of the experiments. It used a coordinator server that successfully aggregated the updates on the model and the parameter requests of working nodes. The experimental results showed that the proposed deep learning model outperformed the classical state-of-the-art deep learning architectures. For P M 2.5 prediction, the model obtained the lowest MAE (5.02 and 22.59, respectively, for the 1-h short term and 48-h long-term predictions), RMSE (7.48 and 28.02) and SMAPE (17.98 and 39.81). The proposed model also achieved the best results for P M 10 prediction with the minimum MAE (7.38 and 28.89), RMSE (9.71 and 25.09), and SMAPE (19.57 and 34.26). This indicates a strong agreement between observed and predicted values. For both benchmarks of short-term and long-term predictions, our proposed model performed better. In addition, the distributed training architecture significantly reduced the training time. It is a helpful step towards resources saving, more optimal training, smart training of algorithms, and more precise and accurate prediction. With improved training time, different organizations and governments can mitigate air pollution levels in a timely manner and take the proper steps to fix problems with air pollution.

Author Contributions

A.G.M.M. investigated the theoretical basis for this work, implemented the deep learning models for the experiments as well as the proposed approach, and wrote the manuscript. E.P. and J.J. collected the data used in the experiments section and made some adjustments in the data pre-processing part. Y.Y. supervised the project and helped to write the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Korea Agency for Infrastructure Technology Advancement (KAIA) grant funded by the Ministry of Land, Infrastructure and Transport (Grant 21CTAP-C152124-03), and the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2021R1A2C1095635).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data is available at: https://www.airkorea.or.kr/index%20/ (accessed on 19 December 2021); https://www.weather.go.kr/w/index.do (accessed on 19 December 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dong, M.; Yang, D.; Kuang, Y.; He, D.; Erdal, S.; Kenski, D. PM2.5 concentration prediction using hidden semi-Markov model-based times series data mining. Expert Syst. Appl. 2009, 36, 9046–9055. [Google Scholar] [CrossRef]
  2. Nimesh, R.; Arora, S.; Mahajan, K.K.; Gill, A.N. Predicting air quality using ARIMA, ARFIMA and HW smoothing. Model Assist. Stat. Appl. 2014, 9, 137–149. [Google Scholar] [CrossRef]
  3. Liu, B.; Jin, Y.; Xu, D.; Wang, Y.; Li, C. A data calibration method for micro air quality detectors based on a LASSO regression and NARX neural network combined model. Sci. Rep. 2021, 11, 21173. [Google Scholar] [CrossRef] [PubMed]
  4. Sayegh, A.S.; Munir, S.; Habeebullah, T.M. Comparing the Performance of Statistical Models for Predicting PM10 Concentrations. Aerosol Air Qual. Res. 2014, 14, 653–665. [Google Scholar] [CrossRef] [Green Version]
  5. Wang, W.; Men, C.; Lu, W.-Z.J. Online prediction model based on support vector machine. Neurocomputing 2008, 71, 550–558. [Google Scholar] [CrossRef]
  6. Martínez-España, R.; Bueno-Crespo, A.; Timón, I.; Soto, J.; Muñoz, A.; Cecilia, J.M. Air-pollution prediction in smart cities through machine learning methods: A case of study in Murcia, Spain. J. Univers. Comput. Sci. 2018, 24, 261–276. [Google Scholar]
  7. Ameer, S.; Shah, M.A.; Khan, A.; Song, H.; Maple, C.; Islam, S.U.; Asghar, M.N. Comparative Analysis of Machine Learning Techniques for Predicting Air Quality in Smart Cities. IEEE Access 2019, 7, 128325–128338. [Google Scholar] [CrossRef]
  8. Wang, X. Deep Learning in Object Recognition, Detection, and Segmentation Foundations and Trends R in Signal Processing. Signal Process. 2014, 8, 217–382. [Google Scholar] [CrossRef]
  9. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10778–10787. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Qin, J.; Park, D.S.; Han, W.; Chiu, C.C.; Pang, R.; Le, Q.V.; Wu, Y. Pushing the Limits of Semi-Supervised Learning for Automatic Speech Recognition. arXiv 2020, arXiv:2010.10504. Available online: http://arxiv.org/abs/2010.10504 (accessed on 19 December 2021).
  11. Li, Y.; Wu, C.Y.; Fan, H.; Mangalam, K.; Xiong, B.; Malik, J.; Feichtenhofer, C. Improved Multiscale Vision Transformers for Classification and Detection. arXiv 2021, arXiv:2112.01526. Available online: http://arxiv.org/abs/2112.01526 (accessed on 22 December 2021).
  12. Liu, M.; Zeng, A.; Xu, Z.; Lai, Q.; Xu, Q. Time Series is a Special Sequence: Forecasting with Sample Convolution and Interaction. arXiv 2021, arXiv:2106.09305. Available online: http://arxiv.org/abs/2106.09305 (accessed on 22 December 2021).
  13. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  14. Mengara, A.M.; Kim, Y.; Yoo, Y.; Ahn, J. Distributed Deep Features Extraction Model for Air Quality Forecasting. Sustainability 2020, 12, 8014. [Google Scholar] [CrossRef]
  15. Zhao, J.; Deng, F.; Cai, Y.; Chen, J. Long short-term memory—Fully connected (LSTM-FC) neural network for PM2.5 concentration prediction. Chemosphere 2019, 220, 486–492. [Google Scholar] [CrossRef] [PubMed]
  16. Qi, Y.; Li, Q.; Karimian, H.; Liu, D. A hybrid model for spatiotemporal forecasting of PM2.5 based on graph convolutional neural network and long short-term memory. Sci. Total Environ. 2019, 664, 1–10. [Google Scholar] [CrossRef] [PubMed]
  17. Wen, C.; Liu, S.; Yao, X.; Peng, L.; Li, X.; Hu, Y.; Chi, T. A novel spatiotemporal convolutional long short-term neural network for air pollution prediction. Sci. Total Environ. 2018, 654, 1091–1099. [Google Scholar] [CrossRef] [PubMed]
  18. Huang, C.-J.; Kuo, P.-H. A Deep CNN-LSTM Model for Particulate Matter (PM2.5) Forecasting in Smart Cities. Sensors 2018, 18, 2220. [Google Scholar] [CrossRef] [Green Version]
  19. Bai, Y.; Li, Y.; Zeng, B.; Li, C.; Zhang, J. Hourly PM2.5 concentration forecast using stacked autoencoder model with emphasis on seasonality. J. Clean. Prod. 2019, 224, 739–750. [Google Scholar] [CrossRef]
  20. Heydari, A.; Majidi, M.; Davide, N.; Garcia, A.; Keynia, F.; de Santoli, L. Air pollution forecasting application based on deep learning model and optimization algorithm. Clean Technol. Environ. Policy 2022, 24, 607–621. [Google Scholar] [CrossRef]
  21. Wang, B.; Yan, Z.; Lu, J.; Zhang, G.; Li, T. Deep Multi-task Learning for Air Quality Prediction. Lect. Notes Comput. Sci. Incl. Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinform. 2018, 11305, 93–103. [Google Scholar] [CrossRef] [Green Version]
  22. Xiao, F.; Yang, M.; Fan, H.; Fan, G.; Al-Qaness, M.A.A. An improved deep learning model for predicting daily PM2.5 concentration. Sci. Rep. 2020, 10, 20988. [Google Scholar] [CrossRef] [PubMed]
  23. Chang, Y.-S.; Chiao, H.-T.; Abimannan, S.; Huang, Y.-P.; Tsai, Y.-T.; Lin, K.-M. An LSTM-based aggregated model for air pollution forecasting. Atmos. Pollut. Res. 2020, 11, 1451–1463. [Google Scholar] [CrossRef]
  24. Arsov, M.; Zdravevski, E.; Lameski, P.; Corizzo, R.; Koteli, N.; Gramatikov, S.; Mitreski, K.; Trajkovik, V. Multi-Horizon Air Pollution Forecasting with Deep Neural Networks. Sensors 2021, 21, 1235. [Google Scholar] [CrossRef] [PubMed]
  25. Guo, C.; Liu, G.; Chen, C.-H. Air Pollution Concentration Forecast Method Based on the Deep Ensemble Neural Network. Wirel. Commun. Mob. Comput. 2020, 2020, 8854649. [Google Scholar] [CrossRef]
  26. Heess, N.; Graves, A. Recurrent Models of Visual Attention. Adv. Neural Inf. Processing Syst. 2014, 27, 1–12. [Google Scholar]
  27. Larochelle, H.; Hinton, G. Learning to combine foveal glimpses with a third-order Boltzmann machine. In Proceedings of the Advances in Neural Information Processing Systems 23: 24th Annual Conference on Neural Information Processing Systems 2010, Vancouver, BC, Canada, 6–9 December 2010. [Google Scholar]
  28. Bahdanau, D.; Cho, K.; Bengio, Y. Neural Machine Translation. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015; pp. 1–15. [Google Scholar]
  29. Cao, J.; Chen, Q.; Guo, J.; Shi, R. Attention-guided Context Feature Pyramid Network for Object Detection. arXiv 2020, arXiv:2005.11475. Available online: https://arxiv.org/abs/2005.11475 (accessed on 17 December 2021).
  30. Sinha, A.; Dolz, J. Multi-scale guided attention for medical image segmentation. arXiv 2019, arXiv:1906.02849, 1–10. Available online: https://arxiv.org/abs/1906.02849 (accessed on 17 December 2021).
  31. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting. In Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Virtual Conference, 2–9 February 2021; pp. 1–15. [Google Scholar]
  32. Lin, J.; Su, Q.; Yang, P.; Ma, S.; Sun, X. Semantic-Unit-Based Dilated Convolution for Multi-Label Text Classification. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018. [Google Scholar] [CrossRef] [Green Version]
  33. Dynamic Memory Networks for Visual and Textual Question Answering. PMLR 2015, 48, 2397–2406.
  34. Chaudhari, S.; Mithal, V.; Polatkan, G.; Ramanath, R. An Attentive Survey of Attention Models. ACM Trans. Intell. Syst. Technol. 2021, 12, 1–32. [Google Scholar] [CrossRef]
  35. Lee, J.B. Attention Models in Graphs: A Survey. ACM Trans. Knowl. Discov. Data 2019, 13, 1–25. [Google Scholar] [CrossRef] [Green Version]
  36. Wang, F. Survey on the Attention Based RNN Model and Its Applications in Computer Vision, Computer Vision and Pattern Recognition. Available online: https://doi.org/10.48550/arXiv.1601.06823 (accessed on 16 December 2021).
  37. Dairi, A.; Harrou, F.; Khadraoui, S.; Sun, Y. Integrated Multiple Directed Attention-Based Deep Learning for Improved Air Pollution Forecasting. IEEE Trans. Instrum. Meas. 2021, 70, 1–15. [Google Scholar] [CrossRef]
  38. Zou, X.; Zhao, J.; Zhao, D.; Sun, B.; He, Y.; Fuentes, S. Air Quality Prediction Based on a Spatiotemporal Attention Mechanism. Mob. Inf. Syst. 2021, 2021, 6630944. [Google Scholar] [CrossRef]
  39. Liu, B.; Yan, S.; Li, J.; Qu, G.; Li, Y.; Lang, J.; Gu, R. An Attention-Based Air Quality Forecasting Method. In Proceedings of the 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December 2018; pp. 728–733. [Google Scholar] [CrossRef]
  40. Chen, H.; Guan, M.; Li, H. Air Quality Prediction Based on Integrated Dual LSTM Model. IEEE Access 2021, 9, 93285–93297. [Google Scholar] [CrossRef]
  41. Chen, Z.; Yu, H.; Geng, Y.-A.; Li, Q.; Zhang, Y. EvaNet: An Extreme Value Attention Network for Long-Term Air Quality Prediction. In Proceedings of the 2020 IEEE International Conference on Big Data (Big Data), Atlanta, GA, USA, 10–13 December 2020; pp. 4545–4552. [Google Scholar] [CrossRef]
  42. AIR KOREA. Available online: https://www.airkorea.or.kr/web (accessed on 30 July 2020).
  43. Korea Meteorological Administration. Available online: https://web.kma.go.kr/eng/index.jsp (accessed on 30 July 2020).
  44. Jocher, G. “YOLOv5,” YOLOv5 is a family of object detection architectures and models pretrained on the COCO dataset, and represents Ultralytics open-source research into future vision AI methods. Available online: https://github.com/ultralytics/yolov5 (accessed on 30 November 2021).
  45. Huang, G. Missing data filling method based on linear interpolation and lightgbm. J. Phys. Conf. Ser. 2021, 1754, 012187. [Google Scholar] [CrossRef]
  46. Normalization|Data Preparation and Feature Engineering for Machine Learning|Google Developers. Available online: https://developers.google.com/machine-learning/data-prep/transform/normalization (accessed on 10 February 2022).
  47. Coefficient, C.; Two, B.; Variables, R. 5 Pearson Correlation Coefficient. Noise Reduct. Speech Process. 2009, 2, 1–2. [Google Scholar] [CrossRef]
  48. Banfield, R.E.; Hall, L.O.; Bowyer, K.; Kegelmeyer, W. A Comparison of Decision Tree Ensemble Creation Techniques. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 29, 173–180. [Google Scholar] [CrossRef] [PubMed]
  49. Dietterich, T.G. An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization. Mach. Learn. 2000, 40, 139–157. [Google Scholar] [CrossRef]
  50. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  51. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  52. Brauwers, G.; Frasincar, F. A General Survey on Attention Mechanisms in Deep Learning. IEEE Trans. Knowl. Data Eng. 2021, 4347, 1–20. [Google Scholar] [CrossRef]
  53. Athira, V.; Geetha, P.; Vinayakumar, R.; Soman, K.P. DeepAirNet: Applying Recurrent Networks for Air Quality Prediction. Procedia Comput. Sci. 2018, 132, 1394–1403. [Google Scholar] [CrossRef]
  54. Cai, J.; Dai, X.; Hong, L.; Gao, Z.; Qiu, Z. An Air Quality Prediction Model Based on a Noise Reduction Self-Coding Deep Network. Math. Probl. Eng. 2020, 2020, 3507197. [Google Scholar] [CrossRef]
  55. Li, S.; Xie, G.; Ren, J.; Guo, L.; Yang, Y.; Xu, X. Urban PM2.5 Concentration Prediction via Attention-Based CNN–LSTM. Appl. Sci. 2020, 10, 1953. [Google Scholar] [CrossRef] [Green Version]
  56. Pan, Q.; Darabos, C.; Moore, J. Performance Evaluation: Metrics, Models and Benchmarks; Springer: Berlin/Heidelberg, Germany, 2008; Volume 7246. [Google Scholar]
Figure 1. Traffic collection setup in Namsan Station.
Figure 1. Traffic collection setup in Namsan Station.
Sustainability 14 03269 g001
Figure 2. Traffic Data Collection. (a) Counting process. (b) Car detection process.
Figure 2. Traffic Data Collection. (a) Counting process. (b) Car detection process.
Sustainability 14 03269 g002
Figure 3. Proposed framework workflow.
Figure 3. Proposed framework workflow.
Sustainability 14 03269 g003
Figure 4. Data correlation.
Figure 4. Data correlation.
Sustainability 14 03269 g004
Figure 5. Feature importance representation. (a) Random Forest, (b) Xgboost, (c) AdaBoost, (d) Gradient Boost.
Figure 5. Feature importance representation. (a) Random Forest, (b) Xgboost, (c) AdaBoost, (d) Gradient Boost.
Sustainability 14 03269 g005aSustainability 14 03269 g005b
Figure 6. Feature importance models evaluation.
Figure 6. Feature importance models evaluation.
Sustainability 14 03269 g006
Figure 7. Proposed deep learning architecture.
Figure 7. Proposed deep learning architecture.
Sustainability 14 03269 g007
Figure 8. Representation of a convolution at time window l.
Figure 8. Representation of a convolution at time window l.
Sustainability 14 03269 g008
Figure 9. LSTM model representation.
Figure 9. LSTM model representation.
Sustainability 14 03269 g009
Figure 10. BiLSTM model representation.
Figure 10. BiLSTM model representation.
Sustainability 14 03269 g010
Figure 11. Attention-based BiLSTM with context vector.
Figure 11. Attention-based BiLSTM with context vector.
Sustainability 14 03269 g011
Figure 12. Proposed training process.
Figure 12. Proposed training process.
Sustainability 14 03269 g012
Figure 13. Comparison of evaluation metrics for different future time steps. (a) MAE comparison at different future time steps for P M 2.5 prediction. (b) MAE comparison at different future time steps for P M 10   prediction. (c) RMSE comparison at different future time steps for P M 2.5 prediction. (d) RMSE comparison at different future time steps for P M 10   prediction. (e) SMAPE comparison at different future time steps for P M 2.5 prediction. (f) SMAPE comparison at different future time steps for P M 10   prediction.
Figure 13. Comparison of evaluation metrics for different future time steps. (a) MAE comparison at different future time steps for P M 2.5 prediction. (b) MAE comparison at different future time steps for P M 10   prediction. (c) RMSE comparison at different future time steps for P M 2.5 prediction. (d) RMSE comparison at different future time steps for P M 10   prediction. (e) SMAPE comparison at different future time steps for P M 2.5 prediction. (f) SMAPE comparison at different future time steps for P M 10   prediction.
Sustainability 14 03269 g013aSustainability 14 03269 g013b
Figure 14. Prediction of P M 2.5 . (a) RNN result. (b) LSTM result. (c) Autoencoder LSTM result. (d) ConvLSTM result. (e) Attention LSTM result. (f) ConvBiLSTM result. (g) Attention CNN LSTM result. (h) Proposed model result.
Figure 14. Prediction of P M 2.5 . (a) RNN result. (b) LSTM result. (c) Autoencoder LSTM result. (d) ConvLSTM result. (e) Attention LSTM result. (f) ConvBiLSTM result. (g) Attention CNN LSTM result. (h) Proposed model result.
Sustainability 14 03269 g014aSustainability 14 03269 g014b
Figure 15. Prediction of P M 10 . (a) RNN result. (b) LSTM result. (c) Autoencoder LSTM result. (d) ConvLSTM result. (e) Attention LSTM result. (f) ConvBiLSTM result. (g) Attention CNN LSTM result. (h) Proposed model result.
Figure 15. Prediction of P M 10 . (a) RNN result. (b) LSTM result. (c) Autoencoder LSTM result. (d) ConvLSTM result. (e) Attention LSTM result. (f) ConvBiLSTM result. (g) Attention CNN LSTM result. (h) Proposed model result.
Sustainability 14 03269 g015aSustainability 14 03269 g015bSustainability 14 03269 g015c
Figure 16. Training time evaluation.
Figure 16. Training time evaluation.
Sustainability 14 03269 g016
Figure 17. Evaluation of the proposed model during distributed training over 10 workers at t + 1 h, 24 h, and 48 h future step prediction. (a) Error rate at t + 1 h future time step ( P M 2.5 ). (b) Error rate at t + 1 h future time step ( P M 10 ). (c) Error rate at t + 24 h future time step ( P M 2.5 ). (d) Error rate at t + 24 h future time step ( P M 10 ). (e) Error rate at t + 48 h future time step ( P M 2.5 ). (f) Error rate at t + 48 h future time step ( P M 10 ).
Figure 17. Evaluation of the proposed model during distributed training over 10 workers at t + 1 h, 24 h, and 48 h future step prediction. (a) Error rate at t + 1 h future time step ( P M 2.5 ). (b) Error rate at t + 1 h future time step ( P M 10 ). (c) Error rate at t + 24 h future time step ( P M 2.5 ). (d) Error rate at t + 24 h future time step ( P M 10 ). (e) Error rate at t + 48 h future time step ( P M 2.5 ). (f) Error rate at t + 48 h future time step ( P M 10 ).
Sustainability 14 03269 g017aSustainability 14 03269 g017b
Figure 18. Training time of the proposed model through 10 workers.
Figure 18. Training time of the proposed model through 10 workers.
Sustainability 14 03269 g018
Figure 19. Proposed model deployment pipeline.
Figure 19. Proposed model deployment pipeline.
Sustainability 14 03269 g019
Figure 20. Online prediction results: (a) 18 h real-time prediction. (b) Comparison between predicted values and ground truth.
Figure 20. Online prediction results: (a) 18 h real-time prediction. (b) Comparison between predicted values and ground truth.
Sustainability 14 03269 g020
Table 1. Literature review on air quality forecasting.
Table 1. Literature review on air quality forecasting.
Reference MethodDatasetResult
[1]Hidden Semi-Markov modelsEPA Air Quality System
in Cook County, Illinois (PM2.5
and meteorological data, 2000–2001 period)
Prediction accuracy of 100% with solar radiation, cloudiness,
temperature, pressure,
humidity, wind speed, dewpoint as input parameters
[2]ARIMA, ARFIMA and HW smoothingAQI from Chandigarh including RSPM, SPM, SO2, NO2ARIMA (RMSE: 18.20; MAE: 15.69; MAPE: 26.86)
HW (RMSE: 30.12; MAE: 25.52; MAPE: 37.00)
[3]LASSO regression combined to a nonlinear autoregressive model with exogenous inputs (NARX)1-AQI data from a national monitoring station in Nanjing (from 14 November 2018 to 11 June 2019) 2-Self data collection containing 234,717 raws. The data includes PM2.5, PM10, CO, NO2, SO2, O3. R-square: PM2.5 (0.933); PM10 (0.918); CO (0.899); NO2 (0.90); SO2 (0.941); O3 (0.936)
RMSE: PM2.5 (8.687); PM10 (13.208); CO (0.156); NO2 (7.715); SO2 (4.874); O3 (12.190)
MAE: PM2.5 (5.951); PM10 (8.981); CO (0.098); NO2 (4.806); SO2 (2.464); O3 (7.788) MAPE:
PM2.5 (0.146); PM10 (0.146); CO (0.095); NO2 (0.177); SO2 (0.131); O3 (0.397)
[4]Multiple Linear Regression Model (MLRM) Quantile Regression Model (QRM)
Generalized Additive Model (GAM)
Boosted Regression Trees 1way and 2way
Air quality pollutants data from the city of Makkah (PM10, CO, SO2, NO2, humidity, temperature, wind speed)Mean Bias Error (MBE): GAM (−39.9); MLRM (−29.3); QRM (−1.4); BRT1 (−43.9); BRT2 (−41.1)
MAE: GAM (74.3); MLRM (80.0); QRM (61.0); BRT1 (75.6); BRT2 (80.4)
MAPE: GAM (33.1); MLRM (35.7); QRM (27.2); BRT1 (33.7); BRT2 (35.8)
RMSE: GAM (120.1); MLRM (123.8); QRM (95.6); BRT1 (121.1); BRT2 (125.6)
[5]Online SVMAir pollutant data in Hong KongTesting set
MAE: 19.2902
RMSE: 25.8993
WIA: 0.7880
[6]Bagging algorithm
Random Committee
Random Forest Random Forest KNN
Air quality data in the region of Murcia (NO, NO2, SO2, NOX, PM10, Benzeno, Toluene, and Xileno)Alcantarilla city results:
Random Forest
Year 2013: MAE (7.65); RMSE (10.20)
Year 2014: MAE (7.33); RMSE (9.77)
[7]Decision tree regression
Random forest regression
Gradient boosting regression ANN multi-layer perceptron regression
The dataset consists of five cities of China which include Guangzhou, Chengdu, Beijing, Shanghai, and ShenyangMLP Results:
Shanghai: MAE (13.84); RMSE (0.03)
Guangzhou: MAE (12.2); RMSE (0.045)
Chengdu: MAE (9.8); RMSE (0.108)
Shenyang: MAE (13.65); RMSE (0.062)
Beijing: MAE (21.79); RMSE (0.0806)
[14]Convolutional Bi-Directional LSTM autoencoder modelAir quality data from South Korea (PM2.5, PM10, NO2, CO, O3, SO2, temperature, humidity, and wind speed)PM2.5
MAE: 50.7
RMSE: 6.93
SMAPE:18.27
PM10
MAE: 5.83
RMSE: 7.22
SMAPE: 17.27
[15]LSTM-Fully
connected neural network
AQI dataset
with 36 monitoring stations in Beijing (from 1 May 2014 to 3 April 2015)
1–6 h Prediction:
MAE: 23.97
RMSE: 35.82
7–12 h:
MAE: 38.34
RMSE: 56.03
13–24 h:
MAE: 47.13
RMSE: 65.60
25–48 h:
MAE: 50.13
RMSE: 69.84
[16]Hybrid learning framework based on a Graph Convolutional network and a LSTM modelHourly scaled dataset of pollutants (PM2.5, PM10 NO2, CO, O3, SO2), from 76 stations over Beijing, Tianjin and Hebei.+1 h prediction:
IA: 0.98
MAE: 13.72
RMSE: 22.41
+72 h prediction
IA: 0.92
MAE: 24.21
RMSE: 38.83
[17]A spatiotemporal convolutional LSTM extended modelHourly PM2.5 concentration data collected at 1233 air quality monitoring stations in Beijing and the whole China from 1 January 2016 to 31 December 2017.RMSE: 12.08
MAE: 5.82
MAPE: 17.09
[18]Deep neural network model that integrates the CNN and LSTMPM2.5 dataset of BeijingMAE: 14.63446
RMSE: 24.22874
Pearson Correlation: 0.959986
IA: 0.97831
[19]Stacked autoencoder modelThe data was collection from three monitoring stations in Beijing (Station I, Wangshouxigong; Station II, Nongzhanguan; Station III, Shunyixincheng)Station 1 (Spring):
MAE: 8.01
RMSE: 10.28
R-square: 0.880
[20]LSTM model combined with multi-verse optimization algorithmAQI data from Iran. It composed of wind speed, air temperature, NO2, and SO2 for five months. The data was collected from May–September 2019 with a time step of 3 hWith data type (1): Month of September for NO2 prediction
RMSE: 0.0545
MAE: 0.0465
MAPE: 17.4011
[21]Deep multi-task learning framework based on residual GRUKDD CUP of Fresh AirRMSE: 1.85
MAE: 1.15
[22]Weighted LSTM extended modelDaily pollutants concentration and meteorological data from Beijing–Tianjin–HebeiRMSE: 40.67
MAE: 26.10
Total accuracy index: 0.59
Spatial anomaly correlation: 0.9524
Temporal correlation coefficient: 0.9930
[23]Aggregated LSTM neural networkThe data was collected in Taiwan from 2012 to 2017. It contains 17 attributes based on air pollutants and meteorological information. RMSE: 0.44
MAE: 0.91
MAPE: 16.3
[24]RNN and LSTM modelsAir quality data collected at Skopje 6 h prediction SimpleRNN + Dense with ReLU:
MSE: 0.0007
RMSE: 0.0273
[25]Deep ensemble model which combines RNN, LSTM, and GRU networksPM2.5 concentration and meteorological data collected at 3 stations in Shanghai (From 1 January 2010 to 31 December 2015) Group 1:
MAE: 6.72
MAPE: 19.60%
[37]VAE with multiple directed attention mechanismData from the United States Environmental Protection Agency which includes NO2, SO2, CO, and O3Case of NO2 prediction in Pennsylvania:
RMSE: 12.373
MAE: 10.370
R-square: 0.831
Explained variance: 0.955
MAPE: 13
Mean bias error (MBE): −2.47718
Relative MBE: −3.21039
[38]LSTM model based on a spatiotemporal attention mechanismBeijing air quality dataset (from 1 January 2018 to 31 December 2018)1 h-Prediction:
RMSE: 12.23
R-square with different set of features:
R-square: 0.78
[39]Seq2Seq model with attention mechanismBeijing air quality data from April 2017 to March 2018Olympic Center:
RMSE: 38.119
R-square: 0.493
Dongsi:
RMSE: 59.508
R-square: 0.337
[40]Integrated dual LSTM with attention mechanismBeijing air quality data from 2013 to 2018RMSE: 14.36
MAE: 8.39
MAPE: 35.78
R-square: 0.89
IA: 0.93
[41]Extreme value attention network based on encoder and decoder frameworkAir quality data from Fuzhou (2 November 2017 to 2 October 2018) and Beijing (1 January 2018 to 31 December 2018) Fuzhou dataset:
RMSE: 17.9914
MAE: 9.3818
Beijing dataset
RMSE: 3.2606
MAE: 2.1867
Table 2. 1D CNN Parameters settings.
Table 2. 1D CNN Parameters settings.
1D-ConvolutionSettings
Convolution LayerKernel Size = (15, 1), Filter = 20, Stride = 1
Max Pooling LayerPool-size = (2, 1), Stride = 2
Dropout0.20
Convolution LayerKernel Size = (10, 1), Filter = 40, Stride = 1
Max Pooling LayerPool-size = (2, 1), Stride = 2
Dropout0.20
Convolution LayerKernel Size = (5, 1), Filter = 80, Stride = 1
Max Pooling LayerPool-size = (2, 1), Stride = 2
Dropout0.20
Table 3. Server’s settings.
Table 3. Server’s settings.
ComponentsSpecs
Server CPUAMD Ryzen 7 2700X 3.7 GHz 8 Cores
Graphic CardsNVIDIA GeForce RTX 2080
Memory32 GB
Table 4. Data distribution.
Table 4. Data distribution.
Data SplitDistribution
Training70%
Validation20%
Testing10%
Table 5. Baseline Models.
Table 5. Baseline Models.
AlgorithmsDescription
RNN [53] Recurrent neural networks are a type of neural network that used previous outputs as inputs while having hidden states. They are mostly used in times series forecasting, speech recognition, and natural language processing.
LSTM models [15]LSTM models are a special type of RNN that overcome the long-term dependency problem faced by standard RNN.
Autoencoder [54] LSTMThe autoencoder LSTM is a hybrid model which is implemented with an LSTM encoder and decoder for sequence data. This model has the same structure setup as an autoencoder but is composed of several LSTM layers.
Convolutional LSTM [17]This model is a hybrid framework based on CNN and LSTM models.
ConvBiLSTM autoencoder [14]This model is based on a convolutional BiLSTM concatenated with an autoencoder model.
Attention LSTM [40]This algorithm is a dual LSTM model with attention mechanism.
Attention CNN-LSTM [55]This model combines a one-dimensional convolutional neural network, LSTM network, and attention-based network.
Table 6. Models’ evaluation results for PM2.5 forecasting.
Table 6. Models’ evaluation results for PM2.5 forecasting.
ModelsMetrics+1 h+2 h+4 h+8 h+10 h+12 h+24 h+48 h
RNNMAE8.339.4911.6115.8720.6723.5026.4130.08
RMSE12.1713.9416.7821.1825.4330.0834.5937.21
SMAPE25.1128.5429.9136.1139.0941.1744.9645.78
LSTMMAE7.969.1510.7814.3818.2119.9825.5129.11
RMSE11.3113.9015.1816.8723.1426.7131.1836.92
SMAPE24.8726.4728.3934.1735.2939.4943.4246.03
Autoencoder LSTMMAE6.838.969.9712.2415.8618.7622.2925.87
RMSE8.9411.8713.8315.7918.4823.4427.5330.21
SMAPE22.3925.5728.0932.9334.2835.9639.4944.27
CNN+LSTMMAE6.038.2110.0112.1814.7617.3820.5324.79
RMSE8.2110.7512.6415.3717.9821.8925.2729.63
SMAPE21.0824.1227.7230.3333.0935.0937.8940.99
Attention LSTMMAE5.838.109.9112.0514.5317.3020.4524.14
RMSE8.1410.4312.3715.1417.5121.3824.8029.10
SMAPE20.1623.5326.1930.0233.0535.5737.4841.05
ConvBiLSTM autoencoderMAE5.347.979.8711.6214.2116.9919.8423.96
RMSE7.929.9012.1815.1116.8920.3424.1228.17
SMAPE19.2723.3725.3329.7433.1436.1538.9040.17
Attention CNN-LSTMMAE5.217.159.2710.9914.1016.1319.0722.63
RMSE7.409.6111.9414.9216.3219.8923.9328.05
SMAPE18.1521.2425.0928.1330.2433.9838.2239.77
Proposed ModelMAE5.026.968.5910.9713.8715.8718.7522.59
RMSE7.489.5311.8914.2816.2119.5323.6128.02
SMAPE17.9820.9124.5727.9130.2834.4937.1839.81
Table 7. Models’ evaluation results for PM10 forecasting.
Table 7. Models’ evaluation results for PM10 forecasting.
ModelsMetrics+1 h+2 h+4 h+8 h+10 h+12 h+24 h+48 h
RNNMAE12.2015.7519.6324.7326.2129.8735.5937.19
RMSE15.3819.3222.6825.4129.3734.6339.4443.28
SMAPE29.7433.5236.3740.0842.1945.6046.3248.17
LSTMMAE11.8613.0916.7518.6721.4924.8427.1831.49
RMSE14.5617.5619.3622.4925.6829.0933.1835.85
SMAPE27.4130.2534.3437.5840.1845.0246.8247.32
Autoencoder LSTMMAE11.2113.2516.5419.0521.5323.3126.5430.97
RMSE13.6716.5318.1121.2524.5727.3331.8534.37
SMAPE25.8928.0931.5235.4737.6940.0844.1645.99
CNN+LSTMMAE9.9512.3815.4918.9621.7823.0725.4629.60
RMSE11.5313.5617.1820.3123.4527.2730.7932.40
SMAPE23.9626.5428.4930.0833.1735.5338.2841.72
Attention LSTMMAE9.1712.0514.2117.3121.1122.3525.3028.37
RMSE11.3613.3316.4719.8722.6927.0429.7632.11
SMAPE22.8426.2727.8130.9332.9235.2938.0540.84
ConvBiLSTM autoencoderMAE8.2911.2513.8516.6719.0721.4825.0727.39
RMSE11.2713.9616.3418.7521.4926.0728.9631.09
SMAPE22.7925.3827.7531.9632.5134.0937.6939.83
Attention CNN-LSTMMAE8.1510.1911.6915.2317.4319.3923.4226.39
RMSE9.8112.9314.7017.8220.2822.7124.8025.93
SMAPE20.4823.3725.9729.0530.9730.6933.1334.33
Proposed ModelMAE7.389.5711.1913.9614.3718.5620.7425.89
RMSE9.7112.2714.6816.7417.8220.3323.6625.09
SMAPE19.5723.1724.9927.0729.4930.2832.9934.26
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mengara Mengara, A.G.; Park, E.; Jang, J.; Yoo, Y. Attention-Based Distributed Deep Learning Model for Air Quality Forecasting. Sustainability 2022, 14, 3269. https://doi.org/10.3390/su14063269

AMA Style

Mengara Mengara AG, Park E, Jang J, Yoo Y. Attention-Based Distributed Deep Learning Model for Air Quality Forecasting. Sustainability. 2022; 14(6):3269. https://doi.org/10.3390/su14063269

Chicago/Turabian Style

Mengara Mengara, Axel Gedeon, Eunyoung Park, Jinho Jang, and Younghwan Yoo. 2022. "Attention-Based Distributed Deep Learning Model for Air Quality Forecasting" Sustainability 14, no. 6: 3269. https://doi.org/10.3390/su14063269

APA Style

Mengara Mengara, A. G., Park, E., Jang, J., & Yoo, Y. (2022). Attention-Based Distributed Deep Learning Model for Air Quality Forecasting. Sustainability, 14(6), 3269. https://doi.org/10.3390/su14063269

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop