Next Article in Journal
Commercial and Scientific Solutions for Blood Glucose Monitoring—A Review
Next Article in Special Issue
Pressure Sensor Placement for Leak Localization in Water Distribution Networks Using Information Theory
Previous Article in Journal
Design of a Highly Efficient Wideband Multi-Frequency Ambient RF Energy Harvester
Previous Article in Special Issue
Two Simultaneous Leak Diagnosis in Pipelines Based on Input–Output Numerical Differentiation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Water Quality Indicator Interval Prediction in Wastewater Treatment Process Based on the Improved BES-LSSVM Algorithm

1
School of Electrical and Control Engineering, North China University of Technology, Beijing 100144, China
2
Advanced Control Systems Research Group at Institutde Robòtical, CSIC-UPC, Universitat Politècnica de Catalunya-BarcelonaTech, 08028 Barcelona, Spain
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(2), 422; https://doi.org/10.3390/s22020422
Submission received: 30 October 2021 / Revised: 16 December 2021 / Accepted: 23 December 2021 / Published: 6 January 2022

Abstract

:
This paper proposes a novel interval prediction method for effluent water quality indicators (including biochemical oxygen demand (BOD) and ammonia nitrogen (NH3-N)), which are key performance indices in the water quality monitoring and control of a wastewater treatment plant. Firstly, the effluent data regarding BOD/NH3-N and their necessary auxiliary variables are collected. After some basic data pre-processing techniques, the key indicators with high correlation degrees of BOD and NH3-N are analyzed and selected based on a gray correlation analysis algorithm. Next, an improved IBES-LSSVM algorithm is designed to predict the BOD/NH3-N effluent data of a wastewater treatment plant. This algorithm relies on an improved bald eagle search (IBES) optimization algorithm that is used to find the optimal parameters of least squares support vector machine (LSSVM). Then, an interval estimation method is used to analyze the uncertainty of the optimized LSSVM model. Finally, the experimental results demonstrate that the proposed approach can obtain high prediction accuracy, with reduced computational time and an easy calculation process, in predicting effluent water quality parameters compared with other existing algorithms.

1. Introduction

Nowadays, freshwater is considered one of the most critical resources for humans, since it can ensure the availability of an acceptable quantity of water for livelihoods, health, ecosystems and production. Hence, freshwater plays a key role in poverty and disease burden reduction, economic growth and environmental sustainability [1,2]. This fact has long been acknowledged all over the world. However, due to industrial pollution, rapid population growth and farmland sewage caused by the extensive use of chemical fertilizers, pesticides and herbicides, the shortage of freshwater sources is a serious and challenging issue [3,4].
Wastewater treatment is one key technology to potentially provide additional water supplies, and it is very important for the functioning of the economy and society. Wastewater treatment has been attracting a lot of attention, since it can not only remove organic wastes to reduce the environmental burden, but also offer the advantage of producing a renewable source of water [5,6]. Wastewater treatment is a very complex process with a variety of physical and biochemical reactions since it presents nonlinear dynamic behavior, time delay and uncertainty [7]. In wastewater treatment plant processes, effluent water quality monitoring is an important task that involves measuring the evolution of the quality parameters in time.
Note that most traditional methods of measuring these quality indicators for wastewater treatment processes are based on manual lab-based monitoring approaches, with manual sample collection, long-time transportation and biological/microbial testing in a laboratory, which is cumbersome and time-consuming. Usually, the testing equipment is very expensive and cannot be used online. In addition, since the process of wastewater treatment is complex, some control strategies are necessary and required to be deployed to guarantee that effluent quality indicators behave normally. In recent decades, water quality monitoring has been evolving to the latest wireless sensor networks [8], such that most of the important indicators of effluent water (pressure, pH, level and so on) can be measured by their corresponding sensors online. However, there are still some parameters that cannot be measured quickly due to high costs and the limitations of sensors, such as BOD and NH3-N. Usually, the concentration of the BOD/NH3-N effluent associated with a wastewater treatment process is an important factor to measure the water quality since the discharge of a large amount of NH3-N and BOD wastewater will lead to water eutrophication, which can affect human health. In China’s “Pollutant Discharge Standard for Urban Wastewater Treatment Plants ( G B 18918 - 2002 )”, the Class A standard stipulates that the maximum discharge for NH3-N is 5 mg/L, while for BOD, it is 10 mg/L. Thus, measuring these effluent quality indicators with high accuracy is an important issue.
Researchers have focused on soft-sensing methods to predict these effluent quality indicators and the prediction task is addressed combining data analytics and water quality control. Soft-sensing methods aim to find some certain relationships between easy-to-measure variables and difficult-to-measure variables in the sewage treatment process. Then, a suitable model is established based on these relationships, and difficult-to-measure variables can be predicted based on the soft-sensing models.
Machine learning approaches are usually considered a subset of artificial intelligence. They focus on some statistical models and algorithms to extract patterns from data so that useful inferences can be used to predict new data. Recently, with the development of machine learning, artificial neural network (ANN), support vector machine (SVM), decision tree, random forest, ensemble learning and many other methods have been researched in depth and have a wide range of applications, including text processing, computer vision, healthcare, finance and robotics. They can also be used for socio-economic and environmental studies [9,10,11,12]. In [12], the impacts of flood protection in Bangladesh were evaluated by machine learning methods. In [13], a gray model and ANN method were investigated to predict suspended matter and chemical oxygen demand in the wastewater treatment process. Cong et al. proposed a mixed soft sensor model based on a wavelet neural network and adaptive weighted fusion for the online prediction of effluent COD [14]. M. Hamada carried out the assessment of a wastewater treatment plant’s performance based on ANN and a multiple linear regression method [15]. M. Zeinolabedini et al. proved that applying various parent wavelet functions to the neural network structure can improve the accuracy of predicting the wastewater sludge volume [16]. A. K. Kadam et al. used ANN and multiple linear regression to model and predict water quality parameters in river basins [17]. S. Heddam et al. investigated a generalized regression neural network model to predict the BOD of effluent in wastewater treatment plants [18]. Tan et al. predicted the first weighting from the working face roof in a coal mine based on a GA-BP neural network [19]. V. Nourani et al. proved that the prediction ability of a neural network ensemble is more reliable [20].
Compared with the ANN method, SVM is another important prediction technique, which can effectively solve the problem of high-dimensional data model construction under the condition of limited samples, and has strong generalization ability. Hence, many scholars have carried out a lot of research on SVM-based prediction. Cheng et al. proposed a variety of kernel single-class SVMs to monitor and predict the intake conditions of wastewater treatment plants [21]. Han et al. developed a neural network model for predicting the sludge volume index based on information transfer strength and adaptive second-order algorithms [22]. Wu et al. proposed an adaptive multi-output soft sensor model for monitoring wastewater treatment and made several simulation comparisons to prove the superiority of the algorithm [23]. K. Lotfi et al. used a linear–nonlinear hybrid method to predict the effluent index of a wastewater treatment plant, which improves the prediction ability of the single method [24]. Han et al. proposed a data-based predictive control strategy and proved its superiority through several simulations [25]. In [26], the total solid content of a wastewater treatment plant was predicted by an SVM model, which can enhance performance and durability.
Although SVM is a small-sample learning method and has been widely used to solve the wastewater prediction problem, the calculation process is multifarious, which is difficult to implement for large-scale training samples [27]. To overcome these disadvantages, the least-squares support vector machine (LSSVM) has been proposed. LSSVM improves the performance of the SVM algorithm by solving linear programming rather than quadratic programming. In this way, the calculation process can be reduced and the computation speed greatly improved [28]. Zhang et al. proposed an improved LSSVM model based on SVM to predict river flow [29]. Fei Luo et al. integrated the Gustafson-Kessel algorithm and least-squares support vector machine for line prediction of [30]. D. S. Manu et al. combined SVM and an adaptive neuro-fuzzy reasoning system model to predict the effluent nitrogen content of wastewater treatment plants [31]. Liu et al. investigated the online prediction of effluent COD in an anaerobic wastewater treatment system based on principal component analysis and the LSSVM algorithm [32].
Note that there are some unknown parameters in the kernel functions of LSSVM that need to be selected in advance. Generally, these parameters are determined according to experience, which may be time-consuming, and it is difficult to find the optimal parameters. Nowadays, swarm intelligence optimization algorithms are researched extensively, since the optimal solution can be found by swarm intelligence to perform a collaborative search mechanism. The results of the combination of swarm intelligence optimization algorithms and machine learning methods can be found in a large number of references. In [33], a hybrid model of particle swarm optimization (PSO) and support vector machine is proposed to predict the turbidity and pH value of sand filtered water in irrigation systems. Han et al. use an adaptive PSO algorithm to design self-organizing radial basis function neural networks to improve the accuracy and save time [34]. Chen et al. study the artificial bee colony optimization back-propagation network to predict the water quality of a water diversion project [35]. Fan et al. use the LSSVM model to improve the performance of predicting the safety factor of a circular slope [36]. Mahdi Shariati et al. use the gray wolf algorithm to optimize ELM model parameters to predict the compressive strength of partially replaced cement concrete [37]. However, to the best of the authors’ knowledge, these swarm intelligence methods may fall into local optima and do not find the global optimal solutions.
Most of the above-mentioned methods only focus on point prediction, without providing information regarding accuracy. The prediction results have strong uncertainty that affects the decision-making process, increasing the risk of not making good decisions. Prediction interval (PI) is a standard tool for quantifying prediction uncertainty. PI not only provides the range where the target value is most likely to exist, but also indicates its accuracy. Yao et al. combined the mean variance estimation (MVE) method with a recurrent neural network to measure the uncertainty in prediction [38]. Yuan et al. combined beta distribution with the PSO-LSTM model to obtain the wind power prediction interval with high reliability and a narrow interval width, so as to provide decision support for the safe and stable operation of power systems [39]. Liao et al. combined the bootstrap method with the long and short memory network to realize the uncertain prediction of the remaining service life of the machine [40]. Marin et al. obtained the prediction interval of power consumption by combining the delta method with a fuzzy prediction model [41]. Sun et al. constructed a high-quality prediction interval based on the two-step method of dual ELM and applied it to the scheduling of a gas system [42]. In recent years, a direct interval prediction method called upper and lower bound estimation (LUBE) has been proposed. The main idea of this method is to directly construct the upper and lower bounds of PI by optimizing the coefficients of the neural network according to the interval quality evaluation index. This approach can provide good performance and does not consider strict data distribution assumptions, such that it can provide more information about the prediction results, which motivates the work of this paper.
The main objective of this paper is to obtain a soft-sensor-based interval prediction method with high prediction accuracy and less computational time to predict the effluent water quality parameters, which is significant for water quality monitoring and control. Aiming at the online prediction of BOD/NH3-N effluent in a wastewater treatment plant within a smart data-driven framework, the main contributions of this paper are the following:
  • Data pre-processing methods, i.e., abnormal data elimination and normalization, are taken into consideration after the data and their related auxiliary variables are collected. Then, some key factors of the wasterwater quality indicators are selected based on the gray correlation analysis algorithm.
  • In order to improve the prediction accuracy of BOD/NH3-N effluent, a novel IBES-LSSVM algorithm is proposed, in which an improved bald eagle search (IBES) optimization algorithm is used to find the optimal parameters of the least-squares support vector machine (LSSVM). The superiority of the proposed method is verified by comparing it with the existing soft-sensing models (such as GWO, WOA, PSO, SSA) using some benchmark functions and providing higher prediction accuracy.
  • In order to estimate the uncertainty of the model prediction results and make better decisions, after obtaining the point prediction results, the interval prediction bounds of effluent quality are also generated. Compared with some existing soft-sensing models, the proposed interval prediction method can obtain a more accurate prediction range.
The structure of this paper is as follows: In Section 2, the problem description is given, including the real data collection, data pre-processing and gray-correlation-analysis-based data selection. Section 3 describes the model uncertainty analysis by using the proposed IBES-LSSVM algorithm and LUBE algorithm. In Section 4, the simulation examples are depicted, demonstrating the effectiveness of the proposed method based on the BOD and NH3-N data. Section 5 draws the main conclusions of this paper.

2. Problem Description

In this paper, a soft-sensing-based method is investigated to analyze and predict the water quality indicators, including three main aspects: data collection, data pre-processing and data interval prediction. The main steps of the approach presented in this paper are shown in Figure 1.
Under a smart data-driven framework, in order to predict water quality tendencies and analyze the mechanisms behind the considered data sources, enough relevant experimental data in real time must be collected based on the prediction quality indicators. Most collected data may present several issues, such as data sparsity and data synchronization, among others. After the data are collected, they must be pre-processed in advance by applying several procedures, such as data cleaning, abnormal data elimination or normalization. Then, correlation analysis from different dimensions of water quality indicators should be considered to extract the relations between these auxiliary variables and find the key factors.

2.1. Data Collection

Due to the complexity of the wastewater treatment process and the large number of parameters that need to be set, it is necessary to determine the characteristic variables related to the water quality to be determined as auxiliary variables. The data that can evaluate the quality or impact of water quality in wastewater treatment plants are mainly divided into the following four categories [43]:
  • Physical data: Physical properties are the ones that must be monitored throughout the treatment process, including total suspended solids, temperature, conductivity, transparency, total dissolved solids, etc.
  • Chemical data: Chemical water quality indices of the national comprehensive discharge standard for water pollutants, including: pH, biochemical oxygen demand, biochemical oxygen consumption, heavy metals, nitrates, etc.
  • Biological data: Biomarkers include a variety of microorganisms in the water, such as mayflies, E. coli, etc.
  • Environmental data: Environmental data cover the whole process of water supply, including indexes of weather, hydrology, soil or ecology.
Figure 1. Main steps of the proposed approach.
Figure 1. Main steps of the proposed approach.
Sensors 22 00422 g001
This paper focuses on a real wastewater treatment plant in Beijing, China, from August 2014 to September 2014 [7,44]. Two data sets are collected first, which are used to predict the BOD/NH3-N effluent, separately. (1) BOD data set: containing 360 batches of data with 23 variables (including the BOD effluent parameters)—the detailed information is shown in Table 1; (2) NH3-N data set: including 10 characteristic variables related to NH3-N effluent parameters, as shown in Table 2.

2.2. Elimination of Abnormal Data

Data collected from wastewater treatment plants can contain erroneous values because of improper instrument operation, human or environmental interference and other factors. As a result, we need to analyze the collected data first, and eliminate some abnormal or meaningless data.
In this paper, we use the 3 σ criterion to handle the abnormal data of the two collected data sets. The sample data are denoted as x 1 , x 2 , , x n . η i is used to represent the data residual error. Then, the standard deviation is calculated as follows:
σ = i = 1 n η i 2 n 1
η i = x i x ¯
where n represents the number of elements in the data set, and x ¯ is the data average.
If the residual error of particular data sample x i satisfies
η i > 3 σ
this means that it corresponds to an abnormal sample and needs to be eliminated. Otherwise, x i is accepted.
Table 1. Effluent BOD data set.
Table 1. Effluent BOD data set.
NumberAuxiliary Variable
01Influent pH (IPH)
02Effluent pH (EPH)
03Influent SS(mg/L)
04Effluent SS (ESS)(mg/L)
05Influent BOD (IBOD)(mg/L)
06Influent COD (ICOD)(mg/L)
07Effluent COD (ECOD)(mg/L)
08Sludge settling ratio of biochemical tank(mg/L)
09MLSS in biochemical tank (MLSS)(mg/L)
10Biochemical pool Do(mg/L)
11Influent oil (IOil)(mg/L)
12Effluent oil (EOil)(mg/L)
13Influent NH3-N (INH3-N)(mg/L)
14Effluent NH3-N(mg/L)
15Influent Chroma (IC)(d)
16Effluent Chroma (EC)(d)
17Influent TN (IT)(mg/L)
18Effluent TN(mg/L)
19Influent phosphate concentration (IPC)(mg/L)
20Effluent phosphate concentration(mg/L)
21Inlet water temperature( C)
22Outlet water temperature( C)
23Effluent BOD (EBOD)(mg/L)
Table 2. Effluent NH3-N data set.
Table 2. Effluent NH3-N data set.
NumberAuxiliary Variable
01Effluent TP(mg/L)
02Influent TP (ITP)(mg/L)
03Temperature (T)( C)
04Anaerobic terminal ORP (ATORP)(mv)
05Aerobic front end DO(mg/L)
06Aerobic terminal DO(mg/L)
07Total suspended solids TTS (TTS)(mg/L)
08Effluent PH (EPH)
09Effluent ORP (EORP)(mL)
10Effluent nitrate (EN)(mg/L)
11Effluent NH3-N (ENH3-N)(mg/L)

2.3. Data Normalization

Different variables often have different dimensions and dimensional units. In order to eliminate the dimensional influence between indicators, it is necessary to normalize the data to achieve uniformity among the different data indicators. There are four classes of normalization methods, i.e., rescaling, mean normalization, standardization and scaling to unit length. In this paper, the rescaling method is selected. The normalization formula is as follows:
x ˜ i = x i x i min x i max x i min
where x i is any value of a variable; x i min and x i max are, respectively, the minimum and maximum value of the variable.
After this kind of normalization, all the values of the data are set in the range of 0 , 1 .

2.4. Correlation Degree Analysis

Since different characteristic variables will have different influences on the predicted variables, to obtain a soft-sensing model with a simpler structure, it is necessary to choose the quality indicators with high correlations. Selecting m ´ auxiliary variables from m variable, it has m ´ < m . In practice, the larger m is, the smaller m ´ is compared to m.
In this paper, the gray relational degree analysis method is investigated to select the characteristic variables of BOD and NH3-N effluents. Gray relational degree analysis is a multi-factor statistical method, which describes the strength of the relationship between various factors according to the gray relational degree. This method looks for the inconsistency between quantitative results and quantitative analysis in the traditional mathematical statistics method and reduces the amount of calculation.
The gray correlation coefficient is formulated as follows:
β = x 0 ( k ) x j ( k )
μ j ( k ) = min j min k β + ρ · max j max k β β + ρ · max j max k β
where j means the j-th variable, k is the k-th iteration, x 0 ( k ) is the output variable, x j ( k ) is the input variable, μ j is the gray correlation coefficient and ρ is the resolution coefficient. If ρ is smaller, the difference between correlation coefficients is larger, and the distinguishing ability is stronger.
Then, the gray correlation degree can be calculated as follows:
γ j = 1 n k = 1 n μ j ( k )
where n is the number of variables.
If the gray correlation degree is larger, this means that the corresponding variable has a higher correlation with the effluent quality indicators. Then, according to the gray correlation degree, the characteristic variables are sorted from front to back. Usually, a threshold is determined in advance as ħ, and then the key indicators can be selected as the input of the soft-sensing model if
γ j >
is satisfied.

3. Methodology

In this section, a novel IBES-LSSVM method is proposed to find the optimal kernel function parameters of the LSSVM in Figure 2.

3.1. LSSVM Algorithm

The theory of LSSVM was first proposed by Suykens in 1994. LSSVM is a kernel learning machine following the principle of structural risk minimization and is suitable for analyzing the issue of sample classification and regression estimation [45].
In LSSVM theory, firstly, the sample data are mapped to higher dimensions through nonlinear changes, and linear functions are used for fitting in this high-dimensional feature space:
y ( x ) = w · ϕ ( x ) + b
where y ( x ) is the output variable, x is the input variables, and w and b are weight and bias terms, respectively.
The optimization objectives of the LSSVM regression algorithm can be formulated as
min J ( w , ξ i ) = 1 2 w T w + C 2 i = 1 n ξ i 2 s . t . y i = w · ϕ ( x ) + b + ξ i , i = 1 , 2 , , n
where C is the regularization coefficient, ξ i is the relaxation variable, and i = 1 n ξ i 2 is the experience risk.
By means of Lagrange multipliers α i , (10) can be expressed as:
L ( w , b , ξ i , α i ) = 1 2 w T w + C 2 i = 1 N ξ i 2 i = 1 n α i [ w · ϕ ( x ) + b + ξ i y i ]
According to Karush–Kuhn–Tucker (KKT) optimization conditions:
L b = 0 i = 1 n α i = 0 L w = 0 w = i = 1 n α i ϕ ( x i ) L ξ i = 0 α i = C ξ i L a = 0 w · ϕ ( x i ) + b + ξ i y i
By defining kernel functions, the optimization problem (11) can be transformed into a linear solution issue:
0 1 1 1 K ( x 1 , x 1 ) + 1 C K ( x 1 , x n ) 1 K ( x n , x 1 ) K ( x n , x n ) + 1 C b α 1 α n = 0 y 1 y n
where K ( x , x i ) is the kernel function.
The Lagrange multiplier and its parameters can be obtained from (13). Therefore, the output of LSSVM can be obtained:
y ^ ( x ) = i = 1 n α i K ( x , x i ) + b
For LSSVM, there are many different types of kernel functions, such as linear function, polynomial kernel function, radial basis function (RBF), sigmoid kernel function, etc. Different kernel functions will produce difference types of LSSVM. In this paper, we select RBF as the kernel function of the model:
K ( x , x i ) = exp ( x x i 2 2 σ 2 )
where σ is the variance of RBF.
Through the aforementioned analysis, LSSVM has two tunable parameters (regularization coefficient C and variance of radial basis kernel function σ with RBF), which are important and need to be determined. To obtain the optimal two parameters, the next step is to use an improved PSO algorithm to optimize them.

3.2. IBES-LSSVM Algorithm

The BES algorithm is an optimization algorithm that simulates the hunting strategy of vultures when looking for fish. It can obtain a single optimal solution through multiple iterations and finally obtain the overall optimal solution, such that the position of the optimal solution corresponds to the optimal parameter value.
BES hunting is divided into three stages. In the first stage (selection space), the eagle selects the space with the largest prey number. In the second stage (spatial search), the eagle moves in the selected space to find the prey. In the third stage (dive), the eagle swings from the best position determined in the second stage and determines the best hunting.
In the selection stage, firstly, this paper optimizes the initial prey position and adopts the tent chaos strategy, which has the advantages of simple structure and strong ergodicity. Then, the linear decreasing method is used to improve the control parameters of the vulture iterative update position. The optimal model parameters of the model can be found that improve the quality of the fitting. The tent chaotic mapping function is described as:
P i + 1 = P i / λ , P i [ 0 , λ ) 1 P i / ( 1 λ ) , P i [ λ , 1 ]
where λ is [ 0 , 1 ] .
Then, the vultures hunt for food. The formula is:
P new , i = P best + R 1 · C 1 · P mean P i
where R 1 is a parameter controlling the position change, and C 1 is a random number between ( 0 , 1 ) . P b e s t is the current optimal location. P m e a n is the average distribution location of vultures after the previous search. P i is the location of the i-th vulture.
In the search phase, vultures search for prey in the selected search space and move in different directions in the spiral space to speed up the search. The best position for subduction is:
P i , new = P i + b ( i ) · P i P i + 1 + a ( i ) · P i P mean
where:
a ( i ) = a r ( i ) max ( | a r | )
b ( i ) = b r ( i ) max ( | b r | )
a r ( i ) = r ( i ) · sin [ ( θ ( i ) ) ]
br ( i ) = r ( i ) · cos [ ( θ ( i ) ) ]
r ( i ) = θ ( i ) + R 2 · C 3
θ ( i ) = π · ω · C 2
ω = ( 1 i i m a x ) 2 · ( ω m a x ω m i n ) + ω m i n
where θ ( i ) and r ( i ) are the polar angle and polar diameter of the spiral equation, respectively. ω and R 2 are the parameters controlling the spiral trajectory. C 2 and C 3 are a random number within ( 0 , 1 ) . The a ( i ) and b ( i ) represent the position of the vulture in polar coordinates, and the values are ( 1 , 1 ) .
During the dive phase, vultures swing from the best position in the search space to their target prey. All points also move towards the best point according to
P i , new = C 4 · P best + a 1 ( i ) · P i R 3 · P man + b 1 ( i ) · P i R 4 · P best
where:
a 1 ( i ) = a r ( i ) max ( | a r | )
b 1 ( i ) = b r ( i ) max ( | b r | )
a r ( i ) = r ( i ) · sinh [ ( θ ( i ) ) ]
br ( i ) = r ( i ) · cosh [ ( θ ( i ) ) ]
r ( i ) = θ ( i )
θ ( i ) = π · ω · C 5
where R 3 and R 4 represent the moving speed of the vulture to the optimal point. C 4 and C 5 are random numbers within ( 0 , 1 ) .

3.3. Interval Prediction

The traditional point prediction cannot deal with the uncertainty in the operation of the system. In order to obtain the numerical estimation and its reliability, the practical application requires the calculation of the prediction interval. Interval prediction indicates the estimation interval of the range of predicted values in a certain confidence interval. Therefore, the prediction interval is composed of the upper and lower line of prediction, which provides its accuracy within a certain confidence level. Assuming that the confidence level is ( 1 μ ) %, l and u are the lower and upper limits, respectively, when P ( l < y < u ) = 1 μ %, and PI can be expressed as [ l , u ] . For a given confidence interval, the smaller the range of prediction interval, the smaller the uncertainty of prediction and the higher the accuracy.
The evaluation indexes of interval prediction are as follows [46].
P I C P : The ratio of the real value to the upper and lower bounds of the prediction interval
P I C P = 1 n i = 1 n c i
If the predicted value is within the [ l i , u i ] range, c i is 1. Otherwise, c i is 0. If all predicted values are included in the prediction interval, P I C P = 100%. n is the number of prediction points. In theory, P I C P ( 1 μ ) %; otherwise, PI is invalid or unreliable. When comparing the PIs by the model, the other indexes should be as small as possible under the condition that the P I C P is as close to the confidence level as possible.
P I N A W : The narrow PI has more information and practical value than the wide PI according to
P I N A W = 1 n R i = 1 n u i l i
where R is the range of predicted values, respectively.
P I N R W : Represents the standard square root width of the predicted interval. The expression is:
P I N R W = 1 R 1 n i = 1 n u i l i 2
C W C : In practical application, it is often hoped that a narrow prediction interval width can still be obtained under the condition of high prediction probability, i.e., the prediction interval range probability and interval width will conflict. Therefore, the comprehensive index C W C is proposed:
C W C = P I N A W 1 + ϱ ( P I C P ) · e τ · ( P I C P ( 1 μ ) )
where τ and μ are constants.
When working with training data, the set ϱ ( P I C P ) is 1. In addition, in data verification, ϱ ( P I C P ) is a step function:
ϱ = 0 P I C P 1 μ 1 P I C P < 1 μ
LUBE is a method based on neural networks to directly calculate the lower and upper bound of the prediction interval. Assuming that the two node values of the output layer of the neural network are the upper and lower limits of the interval, respectively, all the predicted values are included in this range at the confidence level ( 1 μ ) %. The training purpose of a neural network is to minimize the objective function C W C . In this way, the probability and width of the prediction interval are considered at the same time, and the advantages and disadvantages of the prediction interval PI can be comprehensively evaluated.
The flow-chart of the proposed IBES-LSSVM algorithm is shown in Figure 2, which mainly includes the procedure presented in Algorithm 1.
Algorithm 1 LUBE interval prediction based on IBES-LSSVM model.
Input:
Measured data of wasterwater treatment plant.
Output:
Prediction interval of BOD/NH3-N effluent.
Step 1:
Abnormal data elimination, normalization of the data according to Equations (1)–(4).
Step 2:
Analyzing and selecting the key indicators with high correlation degree by Equations (5)–(8).
Step 3:
The bald eagle population is initialized by tent chaos strategy based on Equation (16).
Step 4:
Local optimal solution.
1:
for all X i do:
2:
   for all X i do:
3:
      Obtain predicted value by means of Equations (9)–(15), (17).
4:
   end for
5:
   Using confidence, mean, standard deviation and other parameters, the prediction interval is obtained according to n o r m i n v ( ) formula.
6:
   Evaluate interval fitness by means of Equations (33)–(37).
7:
end for
8:
Obtain the local optimal solution.
Step 5:
Global optimal solution.
1:
While t i t e r do:
2:
   for all X i do:
3:
      Update parameter X, C, σ by using Equations (18)–(25).
4:
      Obtain different predictions by using Equations (9)–(15).
5:
   end for
6:
   Using confidence, mean, standard deviation and other parameters, the prediction interval is obtained according to n o r m i n v ( ) formula.
7:
   Judge and update by Equations (33)–(37).
8:
   for all X i do:
9:
      Update parameter X, C, σ by using Equations (26)–(32).
10:
      Obtain different predictions by using Equations (9)–(15).
11:
      Using confidence, mean, standard deviation and other parameters, the prediction interval is obtained according to n o r m i n v ( ) formula.
12:
      Judge and update by means of Equations (33)–(37).
13:
   end for
14:
    t = t + 1
15:
end while
16:
Obtain the global optimal solution.
Step 6:
Return the global optimal prediction interval.
Step 7:
Output C, σ , fitness and other index values by using Equations (33)–(37), (38)–(41).

4. Simulation Results

In this section, the data sets of BOD/NH3-N effluents are collected from a wastewater treatment plant in Beijing and are used to verify the effectiveness of the proposed approach.
The following evaluation indices of several certainty point predictions are evaluated as follows:
M S E = 1 n i = 1 n y ^ i y i 2
R M S E = 1 n i = 1 n y ^ i y i 2
M A E = 1 n i = 1 n y ^ i y i
R 2 = 1 i = 1 n y ^ y i 2 i = 1 n ( y ^ y ¯ ) 2 = i = 1 n y i y ¯ 2 i = 1 n ( y ^ y ¯ ) 2

4.1. Experiment of Benchmark Functions

The proposed approach is based on the six functions listed in Table 3 with the corresponding ranges and parameters. The range is the boundary of the function search space.
In order to verify the superiority of the proposed approach, it is compared with the WOA, GWO, PSO and SSA algorithms. Statistical results are presented in Table 4. Moreover, the iteration process is depicted in Figure 3, Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8. From the results, we can see that the convergence rate of IBES is better than that of the other algorithms and the proposed IBES method is able to provide competitive results on the benchmark functions.

4.2. Experiment of BOD Data

BOD is one of the most important effluent quality indexes and can reflect the water pollution situation [7]. First, the key auxiliary variables are selected for the BOD effluent data set by calculating the gray correlation degree based on (7). The threshold of the gray correlation degree is chosen as 0.8 . Hence, 14 auxiliary variables (as shown in Table 5) are selected as the soft measurement model inputs. Including the output effluent BOD, there are 15 key indicators; the detailed information is shown in Figure 9. Moreover, the description of each datum is given in Figure 10.
In this paper, the BOD effluent data set has 365 sets of data; among them, 335 sets of data are randomly selected as training samples, and the remaining 30 sets of data are treated as the prediction samples. In order to demonstrate the superiority of the proposed IBES-LSSVM method, it is compared with some existing results, i.e., CNN, LSTM, ELMAN, WOA-LSSVM, GWO-LSSVM, PSO-LSSVM and SSA-LSSVM. In the experiments, the initialization conditions are set as: i t e r is 50, n = 30 , ω m a x = 10 , ω m i n = 0 , R 1 = 1.8 , R 2 = 1 , R 3 = 1.5 , R 4 = 1.5 .
From Table 6 and Table 7 and Figure 11, Figure 12 and Figure 13, we can see that, compared with the existing CNN model, LSTM model, ELMAN model, WOA-LSSVM model, GWO-LSSVM model, PSO-LSSVM model and SSA-LSSVM model, the prediction accuracy of the proposed method is better, demonstrating its effectiveness.

4.3. Experiment of NH3-N Data

In this experiment, the NH3-N effluent data set is considered, which has been described in [44]. First, the gray correlation degree is calculated from (7), and the results are presented in Figure 14. In addition, each selected auxiliary datum of the NH3-N data set is shown in Figure 15.
In this example, the threshold of the gray correlation degree is also chosen as 0.8 ; hence, 7 auxiliary variables (as shown in Table 8) are selected as the soft measurement model input. The experimental data of effluent NH3-N used in this paper are from a sewage treatment plant in Beijing. In total, 237 sets of data were obtained, including 200 sets of data that were randomly selected as training samples, and the remaining 37 sets of data were treated as the prediction samples.
In order to demonstrate the superiority of the proposed BES-LSSVM method, it is compared with some existing approaches, i.e., CNN, LSTM, ELMAN, WOA-LSSVM, GWO-LSSVM, PSO-LSSVM and SSA-LSSVM. In the experiments, the parameters are set as follows: i t e r is 50, n = 30 , ω m a x = 10 , ω m i n = 0 , R 1 = 1.8 , R 2 = 1.2 , R 3 = 1.8 , R 4 = 1.8 .
From Table 9 and Table 10 and Figure 16, Figure 17 and Figure 18, we can see that, compared with the existing CNN model, LSTM model, ELMAN model, WOA-LSSVM model, GWO-LSSVM model, PSO-LSSVM model and SSA-LSSVM model, the prediction accuracy of the proposed method is the best, demonstrating its effectiveness.

5. Conclusions

This paper investigates an improved IBES-LSSVM algorithm to predict the effluent water quality indicators of a wastewater treatment plant, in which an improved BES method is proposed to find the optimal LSSVM parameters. To deal with the uncertainties of the data, the prediction interval is generated within a certain confidence level, which could provide the upper and lower bounds of the prediction results. Compared with other existing methods, the proposed approach demonstrates high prediction accuracy, with reduced computational time and an easy calculation process, in predicting effluent water quality parameters. Note that the proposed results can only predict the water quality indicators, but this is not the end work for a wastewater treatment plant process. The application of this work to reliable decision-making and the generation of a suitable control strategy will be our future work.

Author Contributions

Conceptualization M.Z., V.P.; methodology M.Z., J.W., Y.S., V.P.; resources Y.S.; writing-review and editing M.Z., V.P., Y.Z.; supervision M.Z., J.W.; investigation J.W.; formal analysis M.Z.; software and data curation Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Key Research and Development Program (Grant No. 2018YFC1602704), National Natural Science Foundation of China (Grant No. 51805021,61973023), the Fundamental Research Funds for Beijing (No. 110052972027/015) and the Research foundation for Talents of NCUT (No. 213051360020XN173/017), “Science and Technology Innovation” Special Fund Project in Shijingshan District: 4010537621I7.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available upon request.

Acknowledgments

This work was carried out with the help of Dingyuan Chen at Beijing University of Technology.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Julio, N.; Figueroa, R.; Ponce Oliva, R.D. Water resources and governance approaches: Insights for achieving water security. Water 2021, 13, 3063. [Google Scholar] [CrossRef]
  2. Abrams, A.L.; Carden, K.; Teta, C.; Wågsæther, K. Water, sanitation, and hygiene vulnerability among rural areas and small towns in south Africa: Exploring the role of climate change, marginalization, and inequality. Water 2021, 13, 2810. [Google Scholar] [CrossRef]
  3. Shannon, M.A.; Bohn, P.W.; Elimelech, M.; Georgiadis, J.G.; Mariñas, B.J.; Mayes, A.M. Science and technology for water purification in the coming decades. Nature 2008, 452, 301–310. [Google Scholar] [CrossRef]
  4. Olsson, G. ICA and me—A subjective review. Water Res. 2012, 46, 1585–1624. [Google Scholar] [CrossRef]
  5. Thürlimann, C.M.; Dürrenmatt, D.J.; Villez, K. Soft-sensing with qualitative trend analysis for wastewater treatment plant control. Control Eng. Pract. 2018, 70, 121–133. [Google Scholar] [CrossRef]
  6. Vorosmarty, C.J.; Hoekstra, A.Y.; Bunn, S.E.; Conway, D.; Gupta, J. Fresh water goes global. Science 2015, 349, 478–479. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Qiao, J.; Wang, L.; Yang, C.; Gu, K. Adaptive Levenberg-Marquardt algorithm based echo state network for Chaotic time series prediction. IEEE Access 2018, 6, 10720–10732. [Google Scholar] [CrossRef]
  8. Chen, Y.; Han, D. Water quality monitoring in smart city: A pilot project. Autom. Constr. 2018, 89, 307–316. [Google Scholar] [CrossRef] [Green Version]
  9. Lu, H.; Ma, X. Hybrid decision tree-based machine learning models for short-term water quality prediction. Chemosphere 2020, 249, 126169. [Google Scholar] [CrossRef]
  10. Zhou, P.; Li, Z.; Snowling, S.; Baetz, B.W.; Na, D.; Boyd, G. A random forest model for inflow prediction at wastewater treatment plants. Stoch. Environ. Res. Risk Assess. 2019, 33, 1781–1792. [Google Scholar] [CrossRef]
  11. Liu, H.; Zhang, H.; Zhang, Y.; Zhang, F.; Huang, M. Modeling of Wastewater Treatment Processes Using Dynamic Bayesian Networks Based on Fuzzy PLS. IEEE Access 2020, 8, 92129–92140. [Google Scholar] [CrossRef]
  12. Manandhar, A.; Fischer, A.; Bradley, D.J.; Salehin, M.; Islam, M.S.; Hope, R.; Clifton, D.A. Machine learning to evaluate impacts of flood protection in Bangladesh, 1983–2014. Water 2020, 12, 483. [Google Scholar] [CrossRef] [Green Version]
  13. de Canete, J.F.; Orozco, P.D.S.; Baratti, R.; Mulas, M.; Ruano, A.; Garcia-Cerezo, A. Soft-sensing estimation of plant effluent concentrations in a biological wastewater treatment plant using an optimal neural network. Expert Syst. Appl. 2016, 63, 8–19. [Google Scholar] [CrossRef]
  14. Cong, Q.; Yu, W. Integrated soft sensor with wavelet neural network and adaptive weighted fusion for water quality estimation in wastewater treatment process. Measurement 2018, 124, 436–446. [Google Scholar] [CrossRef]
  15. Hamada, M.; Adel Zaqoot, H.; Abu Jreiban, A. Application of artificial neural networks for the prediction of Gaza wastewater treatment plant performance-Gaza strip. J. Appl. Res. Water Wastewater 2018, 5, 399–406. [Google Scholar] [CrossRef]
  16. Zeinolabedini, M.; Najafzadeh, M. Comparative study of different wavelet-based neural network models to predict sewage sludge quantity in wastewater treatment plant. Environ. Monit. Assess. 2019, 191, 1–25. [Google Scholar] [CrossRef]
  17. Kadam, A.; Wagh, V.; Muley, A.; Umrikar, B.; Sankhua, R. Prediction of water quality index using artificial neural network and multiple linear regression modelling approach in Shivganga River basin, India. Model. Earth Syst. Environ. 2019, 5, 951–962. [Google Scholar] [CrossRef]
  18. Heddam, S.; Lamda, H.; Filali, S. Predicting effluent biochemical oxygen demand in a wastewater treatment plant using generalized regression neural network based approach: A comparative study. Environ. Process. 2016, 3, 153–165. [Google Scholar] [CrossRef]
  19. Tan, T.; Yang, Z.; Chang, F.; Zhao, K. Prediction of the First Weighting from the Working Face Roof in a Coal Mine Based on a GA-BP Neural Network. Appl. Sci. 2019, 9, 4159. [Google Scholar] [CrossRef] [Green Version]
  20. Nourani, V.; Elkiran, G.; Abba, S.I. Wastewater treatment plant performance analysis using artificial intelligence—An ensemble approach. Water Sci. Technol. 2018, 78, 2064–2076. [Google Scholar] [CrossRef]
  21. Cheng, T.; Dairi, A.; Harrou, F.; Sun, Y.; Leiknes, T. Monitoring influent conditions of wastewater treatment plants by nonlinear data-based techniques. IEEE Access 2019, 7, 108827–108837. [Google Scholar] [CrossRef]
  22. Han, H.; Liu, H.; Liu, Z.; Qiao, J. Fault detection of sludge bulking using a self-organizing type-2 fuzzy-neural-network. Control Eng. Pract. 2019, 90, 27–37. [Google Scholar] [CrossRef]
  23. Wu, J.; Cheng, H.; Liu, Y.; Liu, B.; Huang, D. Modeling of adaptive multi-output soft-sensors with applications in wastewater treatments. IEEE Access 2019, 7, 161887–161898. [Google Scholar] [CrossRef]
  24. Lotfi, K.; Bonakdari, H.; Ebtehaj, I.; Mjalli, F.S.; Zeynoddin, M.; Delatolla, R.; Gharabaghi, B. Predicting wastewater treatment plant quality parameters using a novel hybrid linear-nonlinear methodology. J. Environ. Manag. 2019, 240, 463–474. [Google Scholar] [CrossRef]
  25. Han, H.; Zhang, L.; Qiao, J. Data-based predictive control for wastewater treatment process. IEEE Access 2017, 6, 1498–1512. [Google Scholar] [CrossRef]
  26. Mateo Pérez, V.; Mesa Fernández, J.M.; Ortega Fernández, F.; Villanueva Balsera, J. Gross Solids Content Prediction in Urban WWTPs Using SVM. Water 2021, 13, 442. [Google Scholar] [CrossRef]
  27. Ribeiro, D.; Sanfins, A.; Belo, O. Wastewater Treatment Plant Performance Prediction with Support Vector Machines. In Advances in Data Mining. Applications and Theoretical Aspects; Springer: Berlin/Heidelberg, Germany, 2013; pp. 99–111. [Google Scholar] [CrossRef]
  28. Baghban, A.; Sasanipour, J.; Habibzadeh, S.; Zhang, Z. Sulfur dioxide solubility prediction in ionic liquids by a group contribution—LSSVM model. Chem. Eng. Res. Des. 2019, 142, 44–52. [Google Scholar] [CrossRef]
  29. Zhang, N.; Kamaha, R.; Behera, P. Prediction of Surface Water Supply Sources for the District of Columbia Using Least Squares Support Vector Machines (LS-SVM) Method. Adv. Comput. Sci. Int. J. 2015, 4, 1–9. [Google Scholar]
  30. Luo, F.; Qiao, X.; Liao, W. Soft-sensing modeling based on GK-LSSVM method for online predictions of BOD in activated sludge process. In Proceedings of the 2017 International Conference on Robotics and Automation Sciences (ICRAS), Hong Kong, China, 26–29 August 2017; pp. 134–138. [Google Scholar]
  31. Manu, D.; Thalla, A.K. Artificial intelligence models for predicting the performance of biological wastewater treatment plant in the removal of Kjeldahl Nitrogen from wastewater. Appl. Water Sci. 2017, 7, 3783–3791. [Google Scholar] [CrossRef] [Green Version]
  32. Liu, Z.; Wan, J.; Ma, Y.; Wang, Y. Online prediction of effluent COD in the anaerobic wastewater treatment system based on PCA-LSSVM algorithm. Environ. Sci. Pollut. Res. 2019, 26, 12828–12841. [Google Scholar] [CrossRef] [PubMed]
  33. Nieto, P.G.; García Gonzalo, E.; Arbat, G.; Duran Ros, M.; de Cartagena, F.R.; Puig Bargués, J. A new predictive model for the filtered volume and outlet parameters in micro-irrigation sand filters fed with effluents using the hybrid PSO-SVM-based approach. Comput. Electron. Agric. 2016, 125, 74–80. [Google Scholar] [CrossRef]
  34. Han, H.; Lu, W.; Hou, Y.; Qiao, J. An adaptive-PSO-based self-organizing RBF neural network. IEEE Trans. Neural Netw. Learn. Syst. 2016, 29, 104–117. [Google Scholar] [CrossRef] [PubMed]
  35. Chen, S.; Fang, G.; Huang, X.; Zhang, Y. Water quality prediction model of a water diversion project based on the improved artificial bee colony–backpropagation neural network. Water 2018, 10, 806. [Google Scholar] [CrossRef] [Green Version]
  36. Zeng, F.; Amar, M.N.; Mohammed, A.S.; Motahari, M.R.; Hasanipanah, M. Improving the performance of LSSVM model in predicting the safety factor for circular failure slope through optimization algorithms. Eng. Comput. 2021, 1–12. [Google Scholar] [CrossRef]
  37. Shariati, M.; Mafipour, M.S.; Ghahremani, B.; Azarhomayun, F.; Ahmadi, M.; Trung, N.T.; Shariati, A. A novel hybrid extreme learning machine–grey wolf optimizer (ELM-GWO) model to predict compressive strength of concrete with partial replacements for cement. Eng. Comput. 2020, 1–23. [Google Scholar] [CrossRef]
  38. Yao, W.; Zeng, Z.; Lian, C. Generating probabilistic predictions using mean-variance estimation and echo state network. Neurocomputing 2017, 219, 536–547. [Google Scholar] [CrossRef]
  39. Yuan, X.; Chen, C.; Jiang, M.; Yuan, Y. Prediction interval of wind power using parameter optimized Beta distribution based LSTM model. Appl. Soft Comput. 2019, 82, 105550. [Google Scholar] [CrossRef]
  40. Liao, Y.; Zhang, L.; Liu, C. Uncertainty prediction of remaining useful life using long short-term memory network based on bootstrap method. In Proceedings of the 2018 IEEE International Conference on Prognostics and Health Management (ICPHM), Seattle, WA, USA, 11–13 June 2018; pp. 1–8. [Google Scholar]
  41. Marín, L.G.; Cruz, N.; Sáez, D.; Sumner, M.; Núñez, A. Prediction interval methodology based on fuzzy numbers and its extension to fuzzy systems and neural networks. Expert Syst. Appl. 2019, 119, 128–141. [Google Scholar] [CrossRef] [Green Version]
  42. Sun, X.; Wang, Z.; Hu, J. Prediction interval construction for byproduct gas flow forecasting using optimized twin extreme learning machine. Math. Probl. Eng. 2017, 2017, 5120704. [Google Scholar] [CrossRef] [Green Version]
  43. Wu, D.; Wang, H.; Seidu, R. Smart data driven quality prediction for urban water source management. Future Gener. Comput. Syst. 2020, 107, 418–432. [Google Scholar] [CrossRef]
  44. Chen, Z.; Yang, C.; Qiao, J. The optimal design and application of LSTM neural network based on the hybrid coding PSO algorithm. J. Supercomput. 2021, 5. [Google Scholar] [CrossRef]
  45. Su, Y.; Han, L.; Wang, J.; Wang, H. Quantum-behaved RS-PSO-LSSVM method for quality prediction in parts production processes. In Concurrency and Computation Practice and Experience; John Wiley & Sons: Hoboken, NJ, USA, 2019; pp. 1–15. [Google Scholar]
  46. Li, R.; Jin, Y. A wind speed interval prediction system based on multi-objective optimization for machine learning method. Appl. Energy 2018, 228, 2207–2220. [Google Scholar] [CrossRef]
Figure 2. Flow chart of IBES-LSSVM model.
Figure 2. Flow chart of IBES-LSSVM model.
Sensors 22 00422 g002
Figure 3. The result of F1.
Figure 3. The result of F1.
Sensors 22 00422 g003
Figure 4. The result of F2.
Figure 4. The result of F2.
Sensors 22 00422 g004
Figure 5. The result of F3.
Figure 5. The result of F3.
Sensors 22 00422 g005
Figure 6. The result of F4.
Figure 6. The result of F4.
Sensors 22 00422 g006
Figure 7. The result of F5.
Figure 7. The result of F5.
Sensors 22 00422 g007
Figure 8. The result of F6.
Figure 8. The result of F6.
Sensors 22 00422 g008
Figure 9. Auxiliary variables of BOD.
Figure 9. Auxiliary variables of BOD.
Sensors 22 00422 g009
Figure 10. Original data of BOD.
Figure 10. Original data of BOD.
Sensors 22 00422 g010
Figure 11. 99 % of BOD.
Figure 11. 99 % of BOD.
Sensors 22 00422 g011
Figure 12. 95 % of BOD.
Figure 12. 95 % of BOD.
Sensors 22 00422 g012
Figure 13. 90 % of BOD.
Figure 13. 90 % of BOD.
Sensors 22 00422 g013
Figure 14. Auxiliary variables of NH3-N.
Figure 14. Auxiliary variables of NH3-N.
Sensors 22 00422 g014
Figure 15. Original data of NH3-N.
Figure 15. Original data of NH3-N.
Sensors 22 00422 g015
Figure 16. 99 % of NH3-N.
Figure 16. 99 % of NH3-N.
Sensors 22 00422 g016
Figure 17. 95 % of NH3-N.
Figure 17. 95 % of NH3-N.
Sensors 22 00422 g017
Figure 18. 90 % of NH3-N.
Figure 18. 90 % of NH3-N.
Sensors 22 00422 g018
Table 3. Benchmark functions.
Table 3. Benchmark functions.
FunctionRangeParameters
F1 F ( x ) = i = 1 10 ( X a i ) ( X a i ) T + c i 1 [ 1 , 10 ] dim = 4 popsize = 100 iteration = 300
F2 F ( x ) = i = 1 7 ( X a i ) ( X a i ) T + c i 1 [ 1 , 10 ] dim = 4 popsize = 100 iteration = 300
F3 F ( x ) = i = 1 5 ( X a i ) ( X a i ) T + c i 1 [ 1 , 10 ] dim = 4 popsize = 100 iteration = 300
F4 F ( x ) = i = 1 4 c i exp j = 1 6 a i j ( x j p i j ) 2 [ 0 , 1 ] dim = 6 popsize = 100 iteration = 200
F5 F ( x ) = i = 1 4 c i exp j = 1 3 a i j ( x j p i j ) 2 [ 1 , 3 ] dim = 3 popsize = 100 iteration = 120
F6 F ( x ) = 1 500 + j = 1 25 1 j + i = 1 2 x i a i j 6 1 [ 65 , 65 ] dim = 2 popsize = 100 iteration = 180
Table 4. Simulation results of algorithms.
Table 4. Simulation results of algorithms.
GWOPSOWOASSAIBESTheoretical Value
F1−10.5364−105364−10.5364−10.5364−10.5364−10
F2−10.4042−10.4029−10.4029−10.4029−10.4029−10
F3−10.1561−10.1532−10.1576−10.1532−10.1532−10
F4−3.3220−3.3311−3.3231−3.3220−3.3220−3
F5−3.8628−3.8628−3.8627−3.8628−3.8628−3
F60.99800.99800.99802.98210.99801
Table 5. Data after processing.
Table 5. Data after processing.
Number of CoefficientAuxiliary VariableCorrelation
1Influent BOD0.9179
2Effluent COD0.9151
3Influent TN0.9119
4Effluent pH0.8878
5Influent NH3-N0.8826
6Influent pH0.8716
7Influent COD0.8676
8Influent Chroma0.8669
9Influent oil0.8562
10Effluent SS0.8556
11Effluent oil0.8519
12Effluent Chroma0.8415
13Influent phosphate0.8397
14MLSS in biochemical tank0.8037
Table 6. Predictive index of BOD.
Table 6. Predictive index of BOD.
ModelMSERMSEMAER 2
CNN0.08470.15000.11150.9503
LSTM0.13100.29850.23300.8132
ELMAN0.24250.31200.25230.7849
GWO-LSSVM0.06590.02170.01820.9889
WOA-LSSVM0.07110.18310.15210.9693
PSO-LSSVM0.05870.10490.08510.9757
SSA-LSSVM0.07260.23710.17070.9758
IBES-LSSVM0.02010.01040.01030.9911
Table 7. PI of BOD.
Table 7. PI of BOD.
μ = 90% μ = 95% μ = 99%
PICPPINRWCWCPINAWTimePICPPINRWCWCPINAWTimePICPPINRWCWCPINAWTime
CNN0.92980.27310.27310.234841.4890.96170.38480.38480.332542.9400.99110.28410.28410.241346.076
LSTM0.91240.36320.36320.311227.4860.96090.37960.37960.325427.7310.99130.35540.35540.302027.821
ELMAN0.90730.29780.29780.2474316.3160.95490.25730.25730.2202241.4460.99090.25710.25710.213290.582
WOA-LSSVM0.91040.26630.26630.23251.6860.96330.26970.26970.23461.8730.99090.26730.26730.22451.654
GWO-LSSVM0.90990.25570.25570.22411.3960.95870.26680.26680.23551.3890.99110.26890.26890.22542.012
PSO-LSSVM0.91110.25190.25190.21981.0290.95440.25960.25960.21550.9670.99080.27730.27730.22770.963
SSA-LSSVM0.90720.29010.29010.25431.4280.95630.31780.31780.26131.4100.99070.29610.26910.22451.599
IBES-LSSVM0.90530.24680.24680.20071.4060.95310.25690.25690.20641.4320.99070.25690.25690.21111.207
Table 8. Data after processing.
Table 8. Data after processing.
Number of CoefficientAuxiliary VariableCorrelation
1Influent TP0.8730
2Anaerobic terminal ORP0.8726
3Effluent PH0.8693
4Temperature0.8659
5Total suspended solids TTS0.8525
6Effluent ORP0.8257
7Effluent nitrate0.8143
Table 9. PI of NH3-N.
Table 9. PI of NH3-N.
μ = 90% μ = 95% μ = 99%
PICPPINRWCWCPINAWTimePICPPINRWCWCPINAWTimePICPPINRWCWCPINAWTime
CNN0.92310.539510.539510.5011129.9910.96190.497760.497760.4685432.4460.99190.520630.520630.4844531.703
LSTM0.91820.494370.494370.4423522.1760.95880.423200.423200.3782422.6370.99210.531850.531850.5011121.272
ELMAN0.90660.386370.386370.342556.6610.95800.376250.376250.321423.1750.99120.420320.420320.387643.120
WOA-LSSVM0.91970.497110.497110.457391.5470.95810.461060.461060.421311.7110.99130.475620.475620.411211.584
GWO-LSSVM0.92270.510670.510670.461741.3460.96010.511170.511170.478941.1660.99130.517760.517760.456691.163
PSO-LSSVM0.92410.482090.482090.453940.9590.96040.478150.478150.427560.7970.99170.492090.492090.464010.801
SSA-LSSVM0.91120.405790.405790.357521.3630.95740.389470.389470.345561.1840.99090.387770.387770.364541.142
IBES-LSSVM0.90370.345310.345310.309891.3540.95560.349060.349060.311281.1810.99070.346770.346770.310011.366
Table 10. Predictive index of NH3-N.
Table 10. Predictive index of NH3-N.
ModelMSERMSEMAER 2
CNN0.18740.17110.14500.8932
LSTM0.11380.21310.16630.7666
ELMAN0.09540.18460.15640.7872
GWO-LSSVM0.09970.08950.06280.7280
WOA-LSSVM0.19290.23710.17090.8959
PSO-LSSVM0.13120.17220.12470.8922
SSA-LSSVM0.11960.19580.20370.8117
IBES-LSSVM0.09170.06450.04500.8967
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhou, M.; Zhang, Y.; Wang, J.; Shi, Y.; Puig, V. Water Quality Indicator Interval Prediction in Wastewater Treatment Process Based on the Improved BES-LSSVM Algorithm. Sensors 2022, 22, 422. https://doi.org/10.3390/s22020422

AMA Style

Zhou M, Zhang Y, Wang J, Shi Y, Puig V. Water Quality Indicator Interval Prediction in Wastewater Treatment Process Based on the Improved BES-LSSVM Algorithm. Sensors. 2022; 22(2):422. https://doi.org/10.3390/s22020422

Chicago/Turabian Style

Zhou, Meng, Yinyue Zhang, Jing Wang, Yuntao Shi, and Vicenç Puig. 2022. "Water Quality Indicator Interval Prediction in Wastewater Treatment Process Based on the Improved BES-LSSVM Algorithm" Sensors 22, no. 2: 422. https://doi.org/10.3390/s22020422

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop