Next Article in Journal
Effectiveness of Torrefaction By-Products as Additive in Vacuum Blackwater under Anaerobic Digestion and Economic Significance
Previous Article in Journal
Experimental Manufacturing of Ferromanganese Alloy from Man-Made Manganese-Containing Wastes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Industrial Process Control Using DPCA and Hierarchical Pareto Optimization

by
Dmitriy Arsenyev
1,
Galina Malykhina
2,* and
Viacheslav Shkodyrev
1
1
Graduate School of Cyber-Physical Systems Control, Institute of Computer Science and Cybersecurity, Peter the Great St. Petersburg Polytechnic University, Saint-Petersburg 195251, Russia
2
Graduate School of Computer Technologies and Information Systems, Institute of Computer Science and Cybersecurity, Peter the Great St. Petersburg Polytechnic University, Saint-Petersburg 195251, Russia
*
Author to whom correspondence should be addressed.
Processes 2023, 11(12), 3329; https://doi.org/10.3390/pr11123329
Submission received: 23 October 2023 / Revised: 20 November 2023 / Accepted: 24 November 2023 / Published: 29 November 2023

Abstract

:
The control of large-scale industrial systems has several criteria, such as ensuring high productivity, low production costs and the lowest possible environmental impact. These criteria must be established for all subsystems of the large-scale system. This study is devoted to the development of a hierarchical control system that meets several of these criteria and allows for the separate optimization of each subsystem. Multicriteria optimization is based on the processing of data characterizing production processes, which makes it possible to organize a multidimensional statistical control process. Using neural networks to model the technological processes of subsystems and the method of dynamic principal component analysis (DPCA) to reduce the dimensionality of control problems allows us to find more efficient solutions. Using the example of a two-level hierarchy, we showed a variant of the connection between two subsystems by parameters.

Graphical Abstract

1. Introduction

Hierarchical optimization is a methodology for managing complex large-scale production systems that consist of individual subsystems. Large-scale systems have many goals: high productivity, low production costs, and possibly a lower environmental impact. A large-scale high-level system can be divided into subsystems belonging to lower levels of the hierarchy. Each subsystem of a lower level forms a control and transmits information about its output parameters to the next level, which takes the received data into account to form its control actions. Moreover, each subsystem can be optimized independently, in accordance with the developed optimization methodology, taking into account the exchange of information between subsystems of different levels.
In hierarchically organized large-scale systems, lower-level subsystems solve problems, the results of which are used by higher-level subsystems. Each subsystem of the lower level, the operation of which is optimized, affects the optimization results of the subsystem of the next level.
The result of multicriteria optimization (MCO) is the selection of an equilibrium solution from a set of optimal solutions. A set is called Pareto optimal if, when moving from this solution to another in the feasible solution space, any improvement in the value of one target criterion leads to a deterioration in at least one of the remaining target criteria.
Multicriteria optimization methods can be divided into four classes: direct methods based on the preferences of the decision maker (DM); a priori methods that allow for an optimal solution that is closest to the preferences of the decision maker to be obtained; a posteriori methods based on obtaining a set of Pareto optimal solutions and choosing one of the decisions of the decision maker; interactive methods that provide multiple Pareto optimal solutions and invite decision makers to iteratively select the optimal solution, providing recommendations at each iteration step.
Classical a priori methods include the weighted sum method, the epsilon constraint method, and the weighted metrics method. These methods have some disadvantages, in particular the heterogeneity of optimal solutions and the inability to obtain solutions for non-convex sets in the space of objective functions.
Multi-objective optimization methods, such as Particle Swarm Optimization (PSO), are based on modeling the social behavior of a group of subjects who iteratively try to improve their position. For example, the Multi-Objective Gray Wolf Optimizer (MOGWO) method is based on modeling the social leadership and hunting techniques of gray wolves.
Evolutionary algorithms (EA) and methods are used to solve complex FEM problems due to their ability to quickly find solutions in a multi-parameter space. They provide effective sampling in the criterion space by modeling the behavior of the system in some modes that cannot be realized in practice and allow for one to obtain many solutions simultaneously. Depending on the definition of elitism, two types of multifunctional advisers are distinguished. The non-dominated sorting genetic algorithm (NSGA) searches for non-dominated solutions while preserving data diversity. The NSGA-II algorithm performs an additional sorting of non-dominated solutions.
NSGA-II performs the following operations: initializing an initial data set based on the model and input data ranges, criteria and constraints; sorting using criteria for non-dominance of the initialized population; dividing data into edges of different ranks and determining the distance of data condensation along the edge. Individuals from a population are selected based on rank and aggregation distance; the N best individuals are selected according to the criterion of non-dominance and combined into a new population, to which recombination and mutation procedures are applied. These actions continue until the specified number of generations has been calculated.
In our study, we use the evolutionary algorithm NSGA-II to construct the Pareto frontier of a technological process, because it allows for the use of a recurrent multilayer perceptron (MLP) to take into account the dynamics of processes and eliminate rare island conditions.
A number of publications are devoted to solving the MCO problem using machine learning methods. The authors of [1] use deep MLP and EA to implement dynamic MCO. Article [2] presents the results of machine learning for the development, control, diagnostics and prediction of cyber-physical systems. The authors of article [3] use hybrid modeling and a comparison of multicriteria learning with and without supervision and an evolutionary algorithm.
The article [4] demonstrates the use of MCO in systems with the visualization of solutions using growing hierarchical self-organizing maps (HSOM), which are applicable even if the dimension of the criterion space is more than three.
When using hierarchical MCO, we assume that each subsystem can be optimized separately and independently. The solutions obtained as a result of optimization of the lower-level subsystem are transferred to the next level, which manipulates them to find the optimal solution for their level or for the entire system. The advantage of the hierarchical optimization method is its ability to simplify a complex system due to modularity and reduce the dimension of each individual subsystem. Pareto boundaries must be calculated for all subsystems that have their own goals. Optimal solutions for subsystems are found by calculating the Pareto frontier for the subsystems and selecting the desired point on the boundaries of the subsystems. Then, the resulting optimal parameter values are transmitted back to the upper level. Such parameters connect the optimal values of subsystems of neighboring levels. Having received new parameter values, a Pareto optimal solution is generated [5].
In large-scale manufacturing systems, separating subsystems does not always achieve significant dimensionality reductions because each subsystem is characterized by a large number of correlated parameters and dynamic behavior. Therefore, the task of reducing the dimensionality and correlation of subsystems remains relevant [6,7].
The goal of our research is to optimize process control, thereby achieving the Stackelberg equilibrium between the desire to increase equipment performance and reduce the costs (increase the efficiency) of production [8,9]. As an example, we will consider a two-level technological process for producing superheated steam for a metallurgical enterprise.

2. Materials and Methods

2.1. Multivariate Statistical Process Control (MSPC)

The statistical control of production processes, which is mainly used in the energy and chemical industries, is based on mathematical methods of modeling and statistical analysis of large data sets. MSPC is a transition to a higher level of processing the accumulated data characterizing production processes, including modeling, clustering, multifactor monitoring and process control, aiming to constantly improve the performance of departments and the enterprise as a whole.
Currently, in the energy sector, the number of measured parameters used to control just one steam boiler exceeds sixty. The parameters have correlation dependencies. As calculations have shown, cross-correlation coefficients range from 0.01 to 0.99. Therefore, to reduce the dimensionality of the modeling and control problem, it is advisable to use the principal components of parameters obtained from the original data. The dependencies between the parameters may be dynamic and are not always linear, so it is advisable to consider the possibility of using the dynamic principal component method of dynamic DPCA.
The curse of dimensionality makes it difficult to model complex dynamical systems. Reducing the dimension of the model allows for one to obtain more stable results; for example, the authors of [10], when describing the dynamics of the turbulence of the velocity field, use the method of principal components of the field parameters, presented on a fine grid, in a multidimensional space. The principal component method allowed for us to obtain a significant reduction in the model dimension.
The authors of [11] use the principal component method to model systems in the chemical industry.
Nonlinear principal component analysis (NLPCA) is a nonlinear generalization of standard PCA using the principal curve technique, which involves moving from straight lines to curves. To implement the NLPCA method, the authors of [12,13] developed an auto-associative neural network and simulated the dynamics of continuous chemical reactors using linear and nonlinear principal component methods. The nonlinear principal components are determined by a feedforward neural network with one hidden layer. The authors analyze the resulting models from the perspective of their application for process control.

2.2. Pareto Optimization of Industrial Processes

Pareto optimization methodology refers to a posteriori method in which a decision is made after finding a set of feasible solutions to multicriteria optimization [14]. Mathematically, if the goal is to maximize or minimize the objective function F x :
max F x = f 1 x , , f K x , x X ,
where F x : X R K is the objective vector function, and where K > 2 ; x * X * x X : x x * :
x x * i 1 . . K , f i x f i x * i 1 . . K ,   f i x f i x * ,
x x * means that x dominates x * .
When controlling technological processes in production, the objective function is difficult to analytically set, since it depends on operating modes characterized by many interrelated parameters that change over time. Therefore, modern production systems implement a multidimensional statistical control process [15].
The control object is represented as a nonlinear dynamic system in state space, having a vector of control parameters u , a state vector x , and a vector of output parameters y . The optimal control vector u is obtained from the following conditions [16,17]:
Maximization of the criteria vector: m a x u F ;
Constraints in the form of inequalities: G u , x , y 0 ;
Constraints in the form of equalities: H u , x , y = 0 ;
Accounting for boundaries: u L u u U ; x L x x U ; y L y y U .
The input of the dynamic system is the vector of control parameters u T ( n ) , which influence the object at time n , which leads to a change in the state vector x T ( n + 1 ) at the next time, n + 1 . At the output of the system, the result of the measurement is observed in the form of a vector y T ( n ) . The model is represented by a nonlinear process equation and a linear measurement equation:
x n + 1 = φ W A x n + W B u n ,
y n = C x n ,
where W A , W B are matrices characterizing the control object, φ T are vector functions characterizing the nonlinearity of the object, and C is a matrix characterizing the measuring system.
A nonlinear state-space dynamic system is characterized by a recurrent artificial neural network (ANN), which has a hidden layer with nonlinear activation functions and a linear output layer. In a more complex case, the number of hidden layers can be increased. The structure of the ANN is shown in Figure 1.
A neural network with one hidden layer is a model of a system in state space, characterized by Equations (2) and (3). Typically, recurrent multilayer perceptron, which has one or more hidden layers, is more efficient; the structure of this is shown in Figure 1.
Increasing the number of object parameters that can be measured allows for the use of multivariate statistical process control (MSPC). To avoid the “curse of dimensionality” when modeling production processes, it is advisable to use principal component analysis (PCA) for linear and nonlinear dependencies between parameters.
Linear SVD. The initial data are presented in the form of a table containing measurement results (matrix X) recorded at a given time interval. The number of columns m in matrix X is equal to the number of parameters, the number of rows n depends on the number of parameter measurements during the observation time, and n > m. When performing PCA, the columns of the matrix X are centered.
The covariance matrix R obtained by formula R = X T X n 1 , n 1 is a symmetric m × m matrix; therefore, it can be diagonalized as R = V L V T , where V is a matrix, each column of which contains its own vectors, L is a diagonal matrix containing the eigenvalues λ i , in descending order, on the main diagonal. Projections of data onto axes are called principal components. To reduce the data dimensionality from m to k < m , the first k columns of V are selected, and the k × k upper left part of the matrix R . Their product V k L k requires an n × k matrix containing the k principal components s.
PCA is preferentially performed via singular value decomposition (SVD) due to its greater numerical stability. The result of the SVD decomposition of the matrix X is the product X = U S V T , where U is a unitary matrix whose columns contain left singular vectors, a diagonal matrix S containing singular numbers s i , and a matrix V whose columns are called right singular vectors. Since the correlation matrix has the expression:
C = V S U T n 1 = V S 2 V T n 1 , n 1 ,
and the eigenvalues of the covariance matrix are related to the singular numbers λ i = s i 2 n 1 , the principal components are determined by Formula (4), X V = U S V T V = U S , using the numerical method of singular value decomposition.
MSPCs used for control have groups of correlated parameters, and therefore contain redundancy. Redundancy can be reduced using principal component analysis (PCA). The property of PCA that can order the principal components of PCA allows for one to exclude components that have an insignificant influence on the optimization criteria.

2.3. Linear Dynamic PCA

The principal components method does not consider the dynamic properties of manufacturing processes. When the multicriteria optimization of complex nonlinear industrial processes is carried out, it is necessary to consider their dynamic behavior. The results of changes in parameter have a strong correlation, which reflects data redundancy, which makes it difficult to quickly process data. Therefore, to eliminate correlation dependencies and reduce the dimension of the control problem, it is advisable to use the method of dynamic principal component analysis (DPCA) proposed by the authors of the article. DPCA considers the fact that controlled processes are characterized by an autocorrelation function. Such processes can be described by an autoregressive model, a moving average model, or a mixed autoregressive-moving average model.
The DPCA method uses the dilated source matrix X ~ , which is the matrix X augmented with time-shifted repeat values of all variables in X . PCA is then performed on this matrix.
DPCA considers that the data have an implicit vector autoregressive model. The KSG-95 method given in [18] and the RR-13 method proposed by Rato and Reisa [19] are known, which use the same lag for all variables, and the lag value determines the order of the autoregression process.
The method somewhat complicates the process of performing optimal multivariate statistical control, since it considers the autocorrelation of the observed processes. More adequate process models that consider autocorrelation are autoregressive, moving average, or mixed autoregressive moving average models. The use of these models allows for one to obtain good results [20].
We used a method that adds the same number of lags for each i -th parameter. As an example, article [21] shows that for two parameters, x 1 . t , x 2 . t , which are characterized by second-order autoregression, the principal components are calculated using current and previous data:
p X i , t = p i 1 x 1 . t + p i 2 x 2 . t + p i 3 x 1 . t 1 + p i 4 x 2 . t 1 , i = 1 , 4 .
Thus, DPCA leads to the expansion of the parameter matrix X by adding delayed data, the number of which is determined by the order of the autoregressive processes: X ~ = [ X ( t ) , X ( t 1 ) , , X ( t l ) ] . Singular value decomposition is performed on the extended covariance matrix obtained for the matrix X ~ .
Nonlinear SAR: Nonlinear dependencies between parameters lead to the need to use nonlinear PCA. We can use SAR with a kernel. This method uses the function φ ( x ) in the feature space. The kernel K x ( i ) , x ( j ) = φ ( x ( i ) ) T φ ( x ( j ) ) represents the inner product of functions in parameter space. Parameters x j , j = 1 . . m have zero mean value. The covariance matrix is as follows:
Σ = i = 1 N φ ( x ( i ) ) φ ( x ( i ) ) T .
The projection in the space φ ( x ) of features u has the form u = i = 1 N a n φ ( x i ) , where a n represents eigenvectors. As in linear PCA, eigenvalues are used to rank eigenvectors based on how much of the variation in the data is captured by each principal component. This PCA approach is used to reduce the dimensionality of the data for kernel PCA.
The basis for nonlinear dimensionality reductions in nonlinear PCA is provided by the fundamental curves and manifolds. This idea has been explored by many authors [22,23].

3. Results

3.1. Multivariate Statistical Model

To construct the Pareto front from data using classical methods, it is necessary to have a sufficient representation of the data in the entire criteria space. If the data are insufficient, the front boundary may have discontinuities. The solution can be improved by methods that use data generation, allowing for one to achieve a more uniform filling of the criteria plane. Evolutionary algorithms are well-suited for filling areas close to the front with data, which makes it possible to simultaneously obtain many solutions to a problem using hidden parallelism. A recurrent neural network is used to generate data in the strategy space and map then to the criteria space for nonlinear systems.
The NSGA-II genetic algorithm is based on the use of two procedures: a procedure for quickly sorting a set of solutions into fronts of solutions that do not dominate each other and have the same level of dominance, and a procedure for estimating the distance characterizing the density (condensation) of solutions belonging to each front. As a result, the Pareto front will be built according to historical data, which may not contain the most profitable operating modes for the equipment. Therefore, it is advisable to build a boundary based on a process model. A nonlinear, rather complex, dynamically changing model was obtained using a neural network.
The experience of using the NSGA-II genetic algorithm to build a Pareto front for our case showed that the front had irregularities and moved too far from the data, which led to very strict recommendations for the control of production processes, which could lead to an increased operating mode that is not applicable.
Most of the reviewed multicriteria control methods are built on the basis of analytical equations arising from the physical processes that determine the dependencies between system parameters and quality criteria to ensure their functioning. Analytical dependencies do not always take into account more subtle features of production processes, for example, the aging of equipment, the appearance of scale in boilers, changes in fuel oil humidity, and weather conditions. Therefore, our approach is based on models of production processes that are updated based on big data that are constantly accumulated in the data lake, which allows for you to create control actions that are adequate for the current situation in production, considering all factors.
The neural network is pre-trained on historical data of the technological process, and retrained when reconfiguring production processes (after the running-in, repair or replacement of equipment). The application of physically informed neural networks for modeling chemical processes is shown in article [24]. The results of neural network modeling are used to generate a population of data used by the genetic algorithm. The use of DPCA makes it possible to simplify the implementation of a genetic algorithm by excluding equations that describe the relationships between parameters from consideration. The Pareto frontier represents a set of criterion values that satisfy the following conditions:
Maximize the vector of criteria: m a x C F ;
Consider the restrictions imposed on the principal components of the production process parameters: c ~ L c ~ c ~ U .
The measurement results obtained during production contain anomalous measurement results and data loss caused by strong production noise. Failures result in non-numeric values. The evolutionary algorithm is very sensitive to anomalous measurement results. If there are noise, failures, and losses in the change results, the Pareto frontier will be constructed incorrectly. Therefore, it is necessary to perform filtering first. Single anomalous measurement results are suppressed using median filters [25,26,27], while packages of strongly correlated anomalous measurement results are detected using finite-difference filters:
d x i n = x i n x i n 1 x i n d ) x i n d 1 > 2 θ i ;   i = 1 . . M , i f θ i > 3 σ d x i ;   t h e n x n d 1 , , x n = x ^ n ,
where d is the expected length of a packet of highly correlated data, d x i n is the second-order finite difference calculated for the length, d , of the packet, and x ^ ( n ) is the predicted value replacing the packet.
The Algorithm 1 for constructing the Pareto front performs the following steps:
Algorithm 1: Description
1.Reading historical data X;
2.Median filtering;
3.Removal of single anomalous measurements and packages of highly correlated anomalous measurements;
4.Formation of multiple criteria;
5.Formation of an extended matrix X ~ of control parameters;
6.Calculation of DPCA parameters X ~ = P C A X ~ ;
7.Construction of a neural network model A N N of the controlled system A N N X ~ ;
8.Initializing a random population P ( t ) by generating data using ANN( X ~ );
9.Assignment of rank P ( t ) using the method of determining the depth of dominance and assessing diversity using the average distance characterizing the proximity to neighboring solutions;
10. W h i l e   t < T , execute loop for generating new data using the neural network A N N X ~
11. M ( t ) : = S e l e c t i o n ( P ( t ) ) —select individuals using depth of dominance and proximity distance.
12. Q ( t ) : = V a r i a t i o n ( M ( t ) ) crossover and evaluation of Q ( t ) target indicators.
13.Saving the corresponding X ~ —principal components of control parameters.
14. N e w   p o p u l a t i o n   P ( t ) = ( P t , Q ( t ) ) —add a new population to the data.
15.fitness P ( t ) —rank assignment based on depth of dominance and diversity.
16. P ( t + 1 ) : = R e v e n a n t ( P ( t ) ) ; select individuals of a new generation.
17. t = t + 1 ; move to the next data (10);
18. E n d   w h i l e .

3.2. Determination of Optimal Control Actions

Article [28] is devoted to the problem of the multicriteria optimization of a control system using PCA. The authors of [29,30] emphasize that multi-objective optimized control methods can improve the efficiency of industrial heating and power generation processes. In publication [31], the authors show the importance of developing methods for managing production processes in industry 4.0.
Figure 2a shows schematically the mapping of the space of dynamic principal components of the measured parameters [ x 1 , x 2 , x 3 ] T of the controlled object into the criteria space [ F 1 , F 2 ] T . ANN neural network is used to obtain this mapping. Figure 2a illustrates the fact that the system has some non-optimal modes, changing which would simultaneously improve two target criteria F 1   a n d   F 2 . After obtaining the Pareto front, which characterizes a set of optimal operating modes of the object and selecting the preferred mode [ F 1 * , F 2 * ] T as a hyperparameter, it is possible to obtain the optimal values of the dynamic principal components of the parameters [ x 1 * , x 2 * , x 3 * ] T . The IANN neural network is used to produce this mapping. Figure 2b illustrates the mapping from the criterion space to the space of dynamic principal component parameters.
The algorithm for determining control actions to achieve a given point on the Pareto front includes the following operations:
  • Formation of a set of values of criteria C r on the Pareto front;
  • Extraction of information about the corresponding principal components of control parameters using I N N C r ;
  • Training of a neural network that maps optimal values on the Pareto front into the space of principal components of control parameters;
  • Restoration of control parameter values from DPCA.
The application of the multicriteria optimization method is considered in [32]. Publication [33] is devoted to the issue of system stability, which is important for control theory, which was solved using Razumikhin’s method.
Let us consider the technological process of preparing fuel oil at a power plant, which includes a receiving and draining device, the main tanks for storing a constant supply of fuel oil, a fuel oil pump, a pipeline system for fuel oil and steam, a group of fuel oil heaters and filters. To pump fuel oil, fill it and drain it from containers, the temperature of the fuel oil must be at least 60–70 °C. The preparation of fuel oil before combustion consists of removing mechanical impurities, increasing the pressure of fuel oil, and heating it, which are necessary to reduce energy losses in the transport of fuel oil to the boilers of the power plant and its fine atomization in the nozzles of burner devices. The temperature of the fuel oil in the tanks is maintained at 60–80 °C at any time of the year due to circulation heating by returning part (up to 50%) of the fuel oil heated in external heaters to the tank.
A complete set of data characterizing the preparation of fuel oil contains 31 parameters, including the pressure of fuel oil, water and steam, the consumption of water, steam and additives, the temperature of fuel oil, water and steam, and fuel oil humidity. The criteria for the quality of the preparation system’s operation are the temperature and consumption of the prepared fuel oil. The historical raw data contain more than 15,000 measurements of 31 parameters.
The use of a large number of model parameters allows for one to obtain high accuracy when assessing the criteria. However, the solution to the inverse problem, which consists of determining a large number of input parameters based on the values of the criteria, becomes unstable. Even the use of a stabilizer did not provide satisfactory results. To achieve a compromise between accuracy in the approximation of criterion values from experimental data and accuracy when determining optimal control actions to transfer the object to the Pareto front mode, the most significant principal components were selected.
The correlation matrix of the principal components of the control parameters has the form:
9.99 0 0 0 0 0 0 0 0 0 4.65 0 0 0 0 0 0 0 0 0 2.80 0 0 0 0 0 0 0 0 0 1.99 0 0 0 0 0 0 0 0 0 1.76 0 0 0 0 0 0 0 0 0 1.44 0 0 0 0 0 0 0 0 0 1.04 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0.0002 .
After removing components with a total energy not exceeding 3% of the total energy of the input parameters, the correlation matrix of the principal components took the form:
9.99 0 0 0 0 0 4.65 0 0 0 0 0 2.80 0 0 0 0 0 1.99 0 0 0 0 0 1.76 .
The number of principal components determines the size of the ANN input layer. Reducing the size of the input layer led to a significant increase in the error in calculating the criteria. To verify this statement, the result of modeling criteria using ANN for five principal components was obtained. The simulation results are shown in Figure 3. The average relative error in simulating fuel oil flow and temperature was 0.03 and 0.004, respectively. Considering the presence of noise when performing measurements in industrial environments, the simulation result is satisfactory.
Using the model made it possible to obtain a more uniform representation of the criteria for fuel oil consumption ( G f l u e l ) and fuel oil temperature ( T f l u e l ) in the space of criteria before transferring them to the NSGA-II algorithm, which builds the Pareto front. The Pareto front graph, which is obtained by maximizing the components of the criteria vector m a x G f l u e l , T f l u e l , taking into account the restrictions that are imposed on statistically independent components c ~ L c ~ c ~ U , is shown in Figure 4a. The value of the hyperparameter is indicated by a point on the Pareto front.
Using this value, we can determine what the control parameters of the fuel oil preparation system should be to reach the target mode specified on the plane [ G f l u e l , T f l u e l ] . If you set all available operating modes of the system as target modes, then you can check the algorithm, and the verification will be complete. The second neural network, INN, is trained to map any point in the space [ G f l u e l , T f l u e l ] to the principal components of the control parameters. Based on the test results, the relative γ = 2 σ ( c o u t c ) max ( c ) was calculated to determine the principal components corresponding to each mode with a data volume of 15,000. Figure 4a,b show the values of the first and the second principal component of the control parameters obtained using I A N N , their actual values, and relative error. The graphs show that the use of only one component allows for you to get 80% closer to the specified mode; the use of the second component allows for you to get even closer to the optimal mode.
Figure 5 shows a polar diagram of the principal components of control parameters, the values of which make it possible to determine the optimal mode for the fuel oil preparation system corresponding to the selected point on the Pareto frontier shown in Figure 4a. Based on the dynamic principal components, it is possible to determine recommended sequences changing the values of control parameters when necessary to achieve the optimal value of the criteria. Figure 6 shows the recommended physical values of the control parameters recovered from the DPCA. The parameter values shown in Figure 5 and Figure 6 are normalized in such a way that the maximum parameter value is equal to one.
The control parameters listed in Figure 6 characterize the operation of a fuel oil pumping station (FOPS), a fuel oil supply warehouse (FOSW), and a thermal power plant (TPP).
For a system at the top level of the hierarchy, the required value on the criteria plane is specified as a hyperparameter.

3.3. Hierarchical Optimization

As an example of a multicriteria optimization problem, consider a boiler equipment control system that produces superheated steam for production shops and for heating needs. A simplified block diagram is presented in Figure 7. The control object is the steam boiler equipment, which consists of the boiler itself (1), fuel oil preparation subsystems (2), water preparation (3), and air (4). As a result of the production process, superheated steam is generated, which is supplied to the production workshops for consumers (5).
The solution to the problem was divided into several hierarchically organized stages, which are shown in a simplified form in Figure 8, where three dots show blocks that perform similar functions (calculating the Pareto front for air preparation, operating mode for water preparation, and control parameters for air preparation). At the top level of the hierarchy, general goals were set for the boiler equipment control system: efficiency. and system performance. Specific target values for steam boiler efficiency and steam boiler productivity were selected on the Pareto frontier. Then, in accordance with the developed algorithm, using a pre-trained INN for the boiler system, all control parameters were determined that could help to achieve the specified target operating mode of the boiler equipment as a top-level system.
The resulting set of control parameters of the top-level system includes the target parameters of production subsystems of the next hierarchy level. For example, the target parameters of the fuel oil preparation subsystem (2) are fuel oil flow and temperature; the target parameters of the water preparation subsystem (3) are water temperature and flow. The target parameters of the air preparation subsystem (4) are temperature and oxygen percentage. The target values of the subsystem parameters are a projection of the goals of the upper level of the hierarchy onto the Pareto front of lower-level systems.

4. Discussion

With an increase in the number of criteria, the optimization of production control processes makes it possible to create and maintain sustainability not only during production, but also in the production ecosystem. Multicriteria optimization allows for us to achieve better production results while saving resources and reducing the harmful impact on nature and humans. Such production ecosystems include metallurgical production complexes that use natural resources, produce, and energy resources, and strictly control the impact of production on the environment [26].
The parameters of all production processes are documented, and their current and historical data should be used to optimize production. For example, for just one fuel oil treatment system, over the course of a year, there are more than 30 million process records for 32 parameters that have autocorrelation and cross-correlation relationships. The steam boiler system has more than 60 million records for 62 parameter changes. Currently, production process control has several goals, including increasing productivity, increasing efficiency, and reducing harmful emissions. Control optimization is multicriteria, based on an analysis of multivariate correlated statistical processes.
Multicriteria optimization is performed using an evolutionary algorithm, instead of a simpler search for non-dominated solutions, which allows for you to include all possible operating modes of production equipment in the analysis, including those modes that were not implemented during the period of the accumulation of measurement data. This algorithm uses a neural network model of a controlled system, which considers all permissible operating modes of controlled systems and many restrictions on input parameters. The periodic updating of data and retraining/retraining of the model allows for you to maintain its relevance.
The presence of strong inter-correlation dependencies between parameters makes it difficult to write restrictions on parameter values when implementing the NSGA-II algorithm, which leads to information redundancy and makes it difficult to solve the problem of finding control actions that correspond to the Pareto boundary. Therefore, the use of dynamic principal components of control parameters will avoid unnecessary complexity. For example, for a fuel oil preparation system, instead of 29 control parameters, it was found to be possible to use 6 of their principal components.
The reverse transition to the control parameters allows for us to provide recommendations on how to adjust the physical parameters to achieve the Pareto optimal regime. In this case, it is necessary to monitor the stability of the system [33].
The method we proposed was applied to optimize the operation of boiler equipment at a metallurgical enterprise. As a result, with the same productivity of boiler equipment, we were able to increase the average monthly efficiency value from 86.5% to 88.5%.

5. Conclusions

The multicriteria optimization of production processes at a large enterprise is divided into several levels. The top level represents the overall goals for the entire system; the choice of a hyperparameter for the optimization algorithm (the desired point on the Pareto frontier) for such a system is a subjective decision. At a lower level, individual subsystems are connected through variables to higher-level systems and coordinate a point on the Pareto frontier with the higher-level system.

Author Contributions

Conceptualization, D.A. and V.S.; methodology, G.M.; software, G.M.; validation, G.M., V.S. and D.A.; formal analysis, G.M.; investigation, G.M.; resources, D.A.; data curation, V.S.; writing—original draft preparation, G.M.; writing—review and editing, G.M.; visualization, V.S.; supervision, D.A.; project administration, V.S.; funding acquisition, D.A. All authors have read and agreed to the published version of the manuscript.

Funding

The research is partially funded by the Ministry of Science and Higher Education of the Russian Federation as part of the World-class Research Center program: Advanced Digital Technologies (contract № 075-15-2022-311 dated 20 April 2022).

Data Availability Statement

Data are contained within the article.

Acknowledgments

The Laboratory of Intelligent Control Systems, staff for technical support during discussions and provision of resources.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhu, Z.; Yang, Y.; Wang, D.; Tian, X.; Chen, L.; Sun, X.; Cai, Y. Deep multi-layer perceptron-based evolutionary algorithm for dynamic multiobjective optimization. Complex. Intell. Syst. 2022, 8, 5249–5264. [Google Scholar] [CrossRef]
  2. Jin, Y.; Sendhoff, B. Pareto-Based Multiobjective Machine Learning: An Overview and Case Studies. IEEE Trans. Syst. Man. Cybern.–Part. C Appl. Rev. 2008, 38, 397–415. [Google Scholar]
  3. Rai, R.; Sahu, C.K. Driven by Data or Derived Through Physics. A Review of Hybrid Physics Guided Machine Learning Techniques With Cyber-Physical System (CPS) Focus. IEEE Access 2020, 8, 71050–71073. [Google Scholar] [CrossRef]
  4. Suzuki, N.; Okamoto, T.; Koakutsu, S. A Pareto Optimal Solution Visualization Method Using an Improved Growing Hierarchical Self-Organizing Maps Based on the Batch Learning. J. Adv. Comput. Intell. Intell. Inform. 2016, 20, 691–703. [Google Scholar] [CrossRef]
  5. Axell, K.B.; Umnov, E.A.; Umnov, A.E. Optimization of the shape of the Pareto set in multicriterial programming problems. Math. Proc. Mipt 2017, 9, 4. [Google Scholar]
  6. Nasiri, M.M.; Khaleghi, A.; Govindan, K.; Bozorgi-Amiri, A. Sustainable hierarchical multi-modal hub network design problem: Bi-objective formulations and solution algorithms. Oper. Res. Int. J. 2023, 23, 35. [Google Scholar] [CrossRef]
  7. Voronin, A.N.; Savchenko, A.S. A Systematic Approach to Multiobjective Optimization. Cybern. Syst. Anal. 2020, 56, 1000–1011. [Google Scholar] [CrossRef]
  8. Zhang, F.; Zhang, Y.; Xue, Y. Design and Application of Hierarchical Multi-objective Predictive Control for Continuous Flow Stirred Tank Reactor. Int. J. Control Autom. Syst. 2022, 20, 1500–1508. [Google Scholar] [CrossRef]
  9. Gebken, B.; Peitz, S.; Dellnitz, M. On the hierarchical structure of Pareto critical sets. J. Glob. Optim. 2019, 73, 891–913. [Google Scholar] [CrossRef]
  10. Berkooz, G.; Holmes, P.; Lumley, J.L. The proper orthogonal decomposition in the analysis of turbulent flows. Annu. Rev. Inc. Wayback Mach. Annu. Rev. Fluid. Mech. 1993, 25, 539–575. [Google Scholar] [CrossRef]
  11. Tu, C.; D’Odorico, P.; Suweis, S. Dimensionality reduction of complex dynamical systems. Sci. Direct 2021, 24, 22. [Google Scholar] [CrossRef]
  12. Shang, L.; Yan, Z.; Qiu, A.; Li, F.; Zhou, X. Efficient recursive kernel principal component analysis for nonlinear time-varying processes monitoring. In Proceedings of the 2019 Chinese Control And Decision Conference (CCDC), Nanchang, China, 3–5 June 2019; pp. 3057–3062. [Google Scholar] [CrossRef]
  13. Gorgoglione, A.; Castro, A.; Iacobellis, V.; Gioia, A. A Comparison of Linear and Non-Linear Machine Learning Techniques (PCA and SOM) for Characterizing Urban Nutrient Runoff. Sustainability 2021, 13, 2054. [Google Scholar] [CrossRef]
  14. Cerda-Flores, S.C.; Rojas-Punzo, A.A.; Nápoles-Rivera, F. Applications of Multi-Objective Optimization to Industrial Processes: A Literature Review. Processes 2022, 10, 133. [Google Scholar] [CrossRef]
  15. Biegel, T.; Helm, P.; Jourdan, N.; Metternich, J. SSMSPC: Self-supervised multivariate statistical in-process control in discrete manufacturing processes. J. Intell. Manuf. 2023. [Google Scholar] [CrossRef]
  16. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multi-objective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef]
  17. Asilian Bidgoli, A.; Rahnamayan, S.; Erdem, B.; Erdem, Z.; Ibrahim, A.; Deb, K.; Grami, A. Machine learning-based framework to cover optimal Pareto-front in many-objective optimization. Complex Intell. Syst. 2022, 8, 5287–5308. [Google Scholar] [CrossRef]
  18. Kulkarni, A.; Kohns, M.; Bortz, M.; Küfer, K.H.; Hasse, H. Regularities of Pareto sets in low-dimensional practical multi-criteria optimisation problems: Analysis, explanation, and exploitation. Optim. Eng. 2023, 24, 1611–1632. [Google Scholar] [CrossRef]
  19. Rato, T.J.; Reis, M.S. Defining the structure of DPCA models and its impact on process monitoring and prediction activities. Chemom. Intell. Lab. Syst. 2013, 125, 74–86. [Google Scholar] [CrossRef]
  20. Box, G.E.P.; Jenkins, G.M. Time Series Analysis: Forecasting and Control, Revised Edition; Holden Day: San Francisco, CA, USA, 1976. [Google Scholar]
  21. Liu, C.; Bai, J.; Wu, F. Fault Diagnosis Using Dynamic Principal Component Analysis and GA Feature Selection Modeling for Industrial Processes. Processes 2022, 10, 2570. [Google Scholar] [CrossRef]
  22. Pilario, K.E.; Shafiee, M.; Cao, Y.; Lao, L.; Yang, S.H. A Review of Kernel Methods for Feature Extraction in Nonlinear Process Monitoring. Processes 2020, 8, 24. [Google Scholar] [CrossRef]
  23. Cao, H.; Wang, G.; Sun, J. DcPCA: A Deep Learning Model for Contrastive Analytics of Nonlinear Data. In Proceedings of the 2022 China Automation Congress (CAC), Xiamen, China, 25–27 November 2022; pp. 6915–6919. [Google Scholar] [CrossRef]
  24. Billard, L.; Douzal-Chouakria, A.; Samadi, S.Y. Exploring Dynamic Structures in Matrix-Valued Time Series via Principal Component Analysis. Axioms 2023, 12, 570. [Google Scholar] [CrossRef]
  25. Zhu, L.; Li, H.; Feng, Y. Research on big data mining based on improved parallel collaborative filtering algorithm. Cluster Comput. 2019, 22 (Suppl. 2), 3595–3604. [Google Scholar] [CrossRef]
  26. Pinto, T.; Rocha, T.; Reis, A.; Vale, Z. Clustering-Based Filtering of Big Data to Improve Forecasting Effectiveness and Efficiency. In Multimedia Communications, Services and Security. MCSS 2022. Communications in Computer and Information Science; Dziech, A., Mees, W., Niemiec, M., Eds.; Springer: Cham, Switzerland, 2022; Volume 1689. [Google Scholar] [CrossRef]
  27. Furht, B.; Villanustre, F. Introduction to Big Data. In Big Data Technologies and Applications; Springer: Cham, Switzerland, 2016. [Google Scholar] [CrossRef]
  28. Meguetta, Z.E.; Conrard, B.; Bayart, M. Multi-criteria design optimization of control system instrumentation using Principal Component Analysis (PCA) and structural modeling approaches. Int. J. Eng. Adv. Technol. 2014, 4, hal-01413045. [Google Scholar]
  29. Esmaeilzadeh, A.; Deal, B.; Yousefi-Koma, A.; Zakerzadeh, M.R. How Multi-Criterion Optimized Control Methods Improve Effectiveness of Multi-Zone Building Heating System Upgrading. Energies 2022, 15, 8675. [Google Scholar] [CrossRef]
  30. Ebrahimi, M.; Ahmadi, M.A.; Khalife, E. Multi-criteria evaluation, and dynamic modeling of combining thermal photovoltaic and thermoelectric generators to extend electricity generation at night. J. Clean. Prod. 2022, 344, 131107. [Google Scholar] [CrossRef]
  31. Oluyisola, O.E.; Bhalla, S.; Sgarbossa, F.; Strandhagen, J.O. Designing and developing smart production planning and control systems in the industry 4.0 era: A methodology and case study. J. Intell. Manuf. 2022, 33, 311–332. [Google Scholar] [CrossRef]
  32. Shkodyrev, V.P.; Khokhlovskiy, V.; Oleinikov, V. Building a Digital Twin for Local Heating Housing Services. In Cyber-Physical Systems and Control II. CPS&C 2021. Lecture Notes in Networks and Systems; Arseniev, D.G., Aouf, N., Eds.; Springer: Cham, Switzerland, 2023; Volume 460. [Google Scholar] [CrossRef]
  33. Graef, J.R.; Tunç, C.; Tunç, O. Stability of time-delay systems via the Razumikhin method. Bol. Soc. Mat. Mex. 2022, 28, 26. [Google Scholar] [CrossRef]
Figure 1. Structure of a recurrent multilayer perceptron.
Figure 1. Structure of a recurrent multilayer perceptron.
Processes 11 03329 g001
Figure 2. Data mapping: (a)—mapping from data space to criteria space, (b)—mapping from criteria space to control parameter space.
Figure 2. Data mapping: (a)—mapping from data space to criteria space, (b)—mapping from criteria space to control parameter space.
Processes 11 03329 g002
Figure 3. Modeling system performance criteria using ANN, (a) simulation of the fuel oil consumption of a steam boiler based on data obtained over 250 h, (b) simulation of steam boiler fuel oil temperature based on data obtained over 250 h.
Figure 3. Modeling system performance criteria using ANN, (a) simulation of the fuel oil consumption of a steam boiler based on data obtained over 250 h, (b) simulation of steam boiler fuel oil temperature based on data obtained over 250 h.
Processes 11 03329 g003
Figure 4. Modeling of control actions using A N N (a) Pareto boundary; (b) first principal component of the control parameters obtained using I A N N ; (c) second principal component of the control parameters obtained using I A N N .
Figure 4. Modeling of control actions using A N N (a) Pareto boundary; (b) first principal component of the control parameters obtained using I A N N ; (c) second principal component of the control parameters obtained using I A N N .
Processes 11 03329 g004
Figure 5. Recommended values of DPCA control parameters.
Figure 5. Recommended values of DPCA control parameters.
Processes 11 03329 g005
Figure 6. Recommended values of control parameters.
Figure 6. Recommended values of control parameters.
Processes 11 03329 g006
Figure 7. Simplified block diagram for steam production workshops and heating.
Figure 7. Simplified block diagram for steam production workshops and heating.
Processes 11 03329 g007
Figure 8. Two-level hierarchical optimization scheme for a steam boiler.
Figure 8. Two-level hierarchical optimization scheme for a steam boiler.
Processes 11 03329 g008
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Arsenyev, D.; Malykhina, G.; Shkodyrev, V. Industrial Process Control Using DPCA and Hierarchical Pareto Optimization. Processes 2023, 11, 3329. https://doi.org/10.3390/pr11123329

AMA Style

Arsenyev D, Malykhina G, Shkodyrev V. Industrial Process Control Using DPCA and Hierarchical Pareto Optimization. Processes. 2023; 11(12):3329. https://doi.org/10.3390/pr11123329

Chicago/Turabian Style

Arsenyev, Dmitriy, Galina Malykhina, and Viacheslav Shkodyrev. 2023. "Industrial Process Control Using DPCA and Hierarchical Pareto Optimization" Processes 11, no. 12: 3329. https://doi.org/10.3390/pr11123329

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop