Next Article in Journal
A Real-Time and Privacy-Preserving Facial Expression Recognition System Using an AI-Powered Microcontroller
Previous Article in Journal
Real-Time Deep Learning Framework for Accurate Speed Estimation of Surrounding Vehicles in Autonomous Driving
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved BP Neural Network Algorithm for Predicting Structural Parameters of Mirrors

1
College of Optical and Electrical Engineering, Changchun University of Science and Technology, Changchun 130022, China
2
Zhongshan Research Institute, Changchun University of Science and Technology, Zhongshan 528437, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(14), 2789; https://doi.org/10.3390/electronics13142789
Submission received: 14 June 2024 / Revised: 11 July 2024 / Accepted: 14 July 2024 / Published: 16 July 2024
(This article belongs to the Section Computer Science & Engineering)

Abstract

:
Addressing the nonlinear correlations between input variables and output responses, in addition to the time-consuming nature of finite element analysis in mirror design, this study introduces an enhanced back-propagation (BP) neural network (BR-TLDBO-BPNN) employing Bayesian regularization and an optimized dung beetle algorithm. This novel approach facilitates rapid and efficient parameter estimations, significantly reducing the computational overhead. Utilizing an integrated analysis platform, the study obtained training and test samples, and the BR-TLDBO-BPNN model is used to predict the reflector’s mass and root mean square (RMS). The optimization mathematical model is built, and the nonlinear planning function ( f min c o n ) is utilized to solve the problem and find an ideal set of structural parameters. The outcomes demonstrate that the prediction model is accurate enough to predict the mirror characteristics to optimize the mirror structural parameters. Empirical validation demonstrates that the proposed model achieves an over 99% accuracy in predicting mirror characteristics against finite element simulations. As a result, the BR-TLDBO-BPNN algorithm studied in this article not only broadens the application scope of neural networks, but also provides a new practical technique for engineering design.

1. Introduction

With the continuous advancement in technology, optical mirrors are increasingly being used in fields such as astronomy, laser technology, and precision measurement [1]. However, due to the complex structure and meticulous manufacturing processes of mirrors, accurately predicting their structural parameters has become a significant research challenge [2,3,4]. Traditional methods for structural optimization include topology, shape, and size optimization [5,6,7,8,9]. Shape and size optimization, commonly known as parametric optimization, can be limited by the designer’s experience and lack of insight. Topology optimization seeks to determine the optimal layout of a structure under given constraints and load conditions. However, it can require extensive iterative processes and computational resources, making it challenging to implement. Therefore, new optimization methods must be introduced to address these limitations to improve efficiency and accuracy.
The BP neural network, as a typical artificial neural network model, is widely used in data prediction and pattern recognition due to its powerful self-learning and self-adaptation capabilities [10,11,12,13]. However, traditional BP neural networks are prone to getting trapped in local minima during training and have a slow convergence rate, which affects their prediction performance [14,15,16,17]. Additionally, the highly nonlinear relationships between input and output variables in structural optimization require robust numerical models to solve these problems accurately [18,19]. Despite the progress in neural-network-based prediction methods, there is still a notable research gap in addressing the limitations of traditional BP neural networks, particularly their susceptibility to local minima and slow convergence rates. This study aims to fill this gap by proposing a novel back-propagation neural network model that integrates advanced optimization algorithms, thereby guiding the optimization design of mirrors.
In this study, an improved dung beetle algorithm (DBO) and Bayesian regularization (BR) were combined into the BP neural network, to create the BR-TLDBO-BPNN model. The DBO algorithm is utilized to optimize the initial weights and thresholds of the neural network, enhancing the BP network ability to escape local minima and accelerate convergence. BR is incorporated to prevent overfitting, ensuring the model generalizes well to unseen data [20]. The inputs to the neural network include the thickness of the center, the substrate, the fascia, and the outermost wall thickness of the mirror, while the outputs are the weight of the mirror and the RMS accuracy of the mirror’s faceted shape under gravity.
An integrated analysis platform is used to obtain training and test samples, facilitating the construction of an optimized mathematical model of the mirror. The BR-TLDBO-BPNN model quickly solves for the optimal structure of the mirror, offering a novel approach and direction for the structural design of optical mirrors. This improved algorithm significantly enhances the prediction accuracy and efficiency, providing crucial technical support for mirror manufacturing and quality control. By addressing the gap in intelligent structural design, this study provides a new solution for future optimization designs of similar structures.

2. Basic Theory

2.1. BP Neural Network

The BP neural network is a multi-layer feed-forward model trained according to the error back-propagation algorithm with a strong nonlinear mapping capability. The structure of the BP neural network consists of three parts, namely, the input layer, the hidden layer, and the output layer, as shown in Figure 1.
The neural network training process consists of forward propagation and BP. In the forward propagation phase, the input data reaches the output layer from the input layer through the hidden layer and produces the output value [21]. The calculation produces an error if the output value does not match the actual value. In the BP stage, the error is propagated backward from the output layer to the input layer, and the weights and biases of the neural network are adjusted layer by layer during the process [22]. After repeated propagation, the network training stops when the minimum error requirement or the maximum number of training times is met. The goal of the BP neural network is to continuously reduce the error between the output value and the actual value through the learning process, thus improving prediction accuracy [23].

2.2. BR Algorithm

Since traditional BP neural networks will fit the sampling error contained in the training samples during the training process, it will cause the network overfitting phenomenon; i.e., it exhibits a minor error in the training set and a more significant error in the test set [24]. The BR algorithm is a Bayesian-theory-based regularization method mainly applied to neural networks and other machine-learning models to improve their generalization ability and prevent overfitting [25]. The core of the BR algorithm lies in transforming the traditional objective of minimizing the error function to maximizing the posterior probability while considering the probability distribution of the model weights. This study uses the BR algorithm to control the model complexity and constrain the parameters to be optimized to improve the model’s generalization ability.
In the traditional BP neural network training process, the error function is usually a mean square error function (MSE), denoted as:
E m s e = 1 n i = 1 n ( y i - t i ) 2
where y i denotes the desired output value, t i denotes the actual output value, and n denotes the number of samples. The BR algorithm constrains the model complexity by adding a regularization term to the error function. The regularization error function E BR is denoted as:
E B R = β n i = 1 n ( y i t i ) 2 + α n i = 1 n w 2
The first term is the prediction error, the second term is the regularization penalty of the weights, and α and β are the regularization coefficients.
In the Bayesian framework, the BR algorithm considers the merit of data fitting and the model complexity [26]. It maximizes the posterior probability of the model by adjusting the regularization parameters, α and β . The regularization parameters, α and β , are the regularization coefficients. If α > β , the training error is appropriately reduced to prevent overfitting; if α < β , the scale of weights is appropriately limited to reduce the effective network parameters. A larger α will increase the weight penalty, simplifying the model and reducing the risk of overfitting. Nevertheless, it may underfit the data and reduce the performance of the model. A smaller α will lessen the weight penalty, making the model more complex and able to better fit the training data, but increasing the risk of overfitting. A larger β will increase the weight of the data error term, making the model fit the training data more accurately, but it may lead to overfitting, especially on noisy datasets. A smaller β will reduce the weight of the data error term, allowing the model to have a certain degree of error, thereby reducing overfitting, but may lead to the underfitting of the model. From the above, it can be seen that the BR algorithm is an iterative process, each iteration of which optimizes the regular coefficients α and β . This iterative process ensures that the network training error is as small as possible to make the network’s effective weights as small as possible, to reduce the size of the network, improve the network’s generalization ability, and make up for the shortcomings of the BP neural network that is prone to overfitting.

2.3. Improving the Dung Beetle Optimizer

2.3.1. Dung Beetle Optimizer

DBO is a population intelligence optimization strategy inspired by the biological behavior of dung beetles, which centers on simulating the natural behaviors of dung beetles, such as rolling, foraging, stealing, and reproducing. To form an effective search mechanism [27]. This algorithm shows unique advantages in optimizing the parameters of neural networks, especially in optimizing the BPNN’s initial weights and threshold settings. With the DBO algorithm, the initial parameters of the BPNN can be set more precisely, thus improving the efficiency and accuracy of network training [28,29]. When dealing with complex data patterns and nonlinear problems, the DBO-assisted BPNN shows a better performance and generalization ability, which is significant to deep learning and artificial intelligence research [30].
The specific process of the DBO algorithm is as follows: firstly, the population initialization is performed—i.e., a set of initial solutions is randomly generated, and its fitness is calculated; then, it enters the iterative updating phase—in each generation, a set of individuals is selected as the parent, a mutation operation is performed on the parent to generate a new set of solutions, and then a crossover operation is performed to generate the next-generation population; then, the fitness of the solutions of the new-generation population is evaluated and compared with that of the previous generation; and, when the preset termination conditions are satisfied, the iterative process is ended and it results in the optimal solution.

2.3.2. Improved Dung Beetle Optimizer

The DBO optimization algorithm is excellent in optimality-seeking ability and convergence speed. However, it needs to be improved regarding the imbalance between global exploration and local exploitation, the ease of falling into local optimal solutions, and the need for global exploration ability. This study proposes an improved dung beetle optimizer (TLDBO), which is mainly improved in the following three aspects:
  • Logistic-tent chaotic mapping initializes the population and maintains diversity, enhancing the algorithm’s optimality-seeking ability;
  • The dynamic bipartite group convergence factor with an adaptive inertia threshold balances the algorithm’s local and global optimization-seeking ability;
  • The Levy flight mechanism is used to optimize the foraging behavior and improve the global search ability.
To understand these aspects better, a brief introduction is discussed regarding them.
  • Logistic-Tent Chaotic Mapping
Since the size of each dimension of dung beetle individuals is randomly generated in the initialization stage, this leads to the initial solution being prone to aggregation, and having poor coverage in the solution space and low variability among individuals. At the same time, chaotic mapping initializes the population and can effectively improve the problem. Tent chaotic mapping, logistic chaotic mapping, and circle chaotic mapping are more common. The logistic-tent chaotic mapping adopted in this study integrates the complex, chaotic dynamics of logistics and the faster iteration speed, more autocorrelation, and applicability to many sequences of the tent chaotic system. The mathematical expression of logistic–tent chaotic mapping is:
x n + 1 = r x n ( 1 x n ) + ( 4 r ) 2 x n mod 1 , i f   x n < 0.5 r x n ( 1 x n ) + ( 4 r ) ( 1 x n ) 2 mod 1 , i f   x n 0.5
where r is the control parameter, r ( 0 , 1 ) , and x n is the position information of the dung beetle.
Figure 2 shows the initialization of the logistic–tent chaotic mapping population. The improved logistic–tent chaotic mapping chaotic values are more evenly distributed. Therefore, the improved logistic–tent chaotic mapping is used to initialize the population to enhance the diversity of the population, which, in turn, enhances the algorithm’s ability to find the best.
  • Dynamic Bipartite Group Convergence Factors and Adaptive Inertia Thresholds
In the DBO, as the convergence factor R decreases linearly, the area of the dung beetle breeding area and dung beetle foraging area decreases gradually, which makes it challenging to co-ordinate the global search ability with the local search ability. Therefore, this study proposes a dynamic community convergence factor and a dynamic community strategy suitable for the dung beetle optimizer.
During the iterative process of dung beetle reproduction and dung beetle foraging populations, the Euclidean distances between the current dung beetle position x i t and the optimal dung beetle position x g b e s t and the worst dung beetle position x g w o r s t are calculated as D g b e s t and D g w o r s t , respectively. A threshold ( D g b e s t D g w o r s t / 3 ) is set to classify dung beetles in this range as better subpopulations, while the others are categorized as worse. According to the characteristics of the two subgroups, this study adopts different nonlinear convergence factors, R 1 and R 2 , as, respectively:
R 1 = 2 4.95 ln 1 + ( t / T ) 0.5 e 1
R 2 = 2 4.95 ln 1 + ( t / T ) 2 e 1
This study uses different convergence factors for the better subpopulation with the higher adaptation and the worse subpopulation with the lower adaptation. The nonlinear convergence factor R 1 of the better subpopulation decreases rapidly at the beginning of the iteration to promote the dung beetle in approaching the optimal solution quickly. It decreases slowly at the end of the iteration to prevent local optimization. The convergence factor R 2 of the worst subgroup decreases gradually at the beginning of the iteration to expand the dung beetle’s search range. It decreases rapidly at the end of the iteration to achieve convergence. Figure 3 shows the change in the convergence factor after the improvement.
Inspired by the idea of inertia weights in the particle swarm optimization (PSO), an adaptive inertia weighting strategy is added to the dung beetle daisy ball position update and the foraging small dung beetle position update to reduce the probability of the algorithm falling into a local optimum and the adaptive inertia weights are calculated as follows:
w ( t ) = w min + ( w max w min ) ( 1 t t max ) δ
where w max is the maximum value of the inertia weights, w min is the minimum value of the inertia weights, t max is the maximum number of iterations, and δ is a positive accurate number parameter that controls the decay rate. When δ > 1 , the inertia weights decline slower at the beginning of the iteration, which helps the algorithm to perform a broader global search, while the rate of decline is accelerated at the later stages of the iteration, which helps the algorithm to perform a more focused local search. When δ < 1 , the inertia weights fall faster at the beginning of the iteration, allowing the algorithm to focus more on local search early on, and fall slower later in the iteration, helping the algorithm to maintain the diversity of the search, thus avoiding falling into a local optimum solution too early.
The dung beetle daisy ball position update formula after introducing adaptive inertia weights is:
B i ( t + 1 ) = w ( t ) X * + b 1 × ( B i ( t ) L b * ) + b 2 × ( B i ( t ) U b * )
where B i ( t ) is the position of the i th dung beetle at the t th iteration, b 1 and b 2 are the random variables of 1 × D , and D is the optimization dimension.
The position update formula of the foraging dung beetle after introducing adaptive inertia weights is:
x i ( t + 1 ) = w ( t ) x i ( t ) + C 1 × ( x i ( t ) L b * ) + C 2 × ( x i ( t ) U b * )
where x i ( t ) is the position of the i th daisy at the t th iteration, C 1 is a random number that follows a normal distribution, and C 2 is a random number belonging to (0, 1).
  • Levy flight mechanism
The Levy flight mechanism is a stochastic search strategy commonly used in optimization algorithms that mimics the natural behavior of certain animals, such as the migratory paths of birds, which exhibit a specific type of random walk with step sizes that obey a Levy distribution. As shown in Figure 4, compared to Gaussian and Cauchy distributions, the Levy distribution is heavy-tailed, converges faster, and provides extensive global search capabilities in optimization. The formula for generating the Levy flight random walk is:
L e v y ( d ) = u r 2 1 / β
where μ ( 0 , σ μ 2 ) , ν ( 0 , σ ν 2 ) . σ μ and σ ν are:
σ μ = Γ ( 1 + β ) sin ( π β / 2 ) Γ ( 1 + λ ) / 2 β 2 ( β 1 ) / 2 1 / β , σ ν = 1
where Γ is the gamma function, β is a constant, and β 0 , 2 .

2.3.3. Performance Evaluation of Improved Dung Beetle Optimizer

To verify the performance of the improved algorithm, the improved TLDBO algorithm in this study is compared with gray wolf optimization (GWO), northern goshawk optimization (NGO), the whale optimization algorithm (WOA), and DBO in terms of the test functions. The main purpose of optimizing algorithm testing functions is to evaluate and validate the performance and applicability of the optimization algorithm. These test functions are usually mathematical functions designed to simulate practical optimization problems and have known global optimal solutions or known optimal solution properties. By conducting experiments on these test functions, we can understand the performance of optimization algorithms in different situations, and discover their advantages, disadvantages, and applicability. Six benchmark test functions were selected for the experiment, and the dimensions and search ranges of the single-peak and multi-peak benchmark test functions are shown in Table 1, where f 1 , f 2 , and f 3 are single-peak benchmark test functions and f 4 , f 5 , and f 6 are multi-peak benchmark test functions. The number of populations of each algorithm is set to 30, and the number of iterations is 500. In order to reduce the chance of the experiment and increase the persuasive power of the experimental results, each algorithm is run independently 50 times on each of the six benchmark test functions, and the optimal value, average value, and standard deviation of the function are calculated. Figure 5 shows the three-dimensional view of the single-peak and multi-peak functions.
The convergence curves of different optimization algorithms on the six benchmark test functions are shown in Figure 6. From the convergence curves and the experimental results, it can be visualized that the improved TLDBO algorithm performs better in finding optimization in different test functions. Compared with the other four algorithms, TLDBO has a better solution accuracy and faster convergence speed in single-peak test functions; TLDBO has a better ability to jump out of the local optimum in multi-peak test functions. The improvement strategy proposed above is effective, and the solution effect and robustness of TLDBO are better and more stable than the other algorithms. It also has a solid ability to find the optimum.

3. BPNN Prediction Model Construction

3.1. Integration Analysis

The structural form of the mirror directly affects its structural rigidity, thermal performance, and optical surface face type accuracy. In this study, the mirror’s diameter is 600 mm, the back of the mirror adopts the back-opening structure, and the shape of the aperture is a triangular aperture, which is a better structural form of craftsmanship and has good stability. Figure 7 is the finite element model of the mirror; according to the finite element analysis results, the structural parameters affecting the precision and stability of the optical surface pattern include the center thickness of the mirror body, the thickness of the substrate, the thickness of the ribbed plate, and the thickness of the outermost wall, which determine the quality of the mirror and the mirror surface shape error.

3.2. Training Sample Confirmation

Using the Latin hypercube sampling method to sample the input random variables, 400 sets of input variable data were obtained. Based on 400 sets of input and output data, we generate 320 sets of training samples and 80 sets of testing samples. Then, we determine the structure of the neural network, set prediction errors, and establish a prediction model using an improved DBO algorithm, BR algorithm, and BPNN method. We normalize and preprocess the training and testing sample data. The number of hidden neurons in the M model is 9, and the number of hidden neurons in the improved dung beetle algorithm is 40.
The inputs to the neural network are the thickness of the overall center, the substrate, the fascia, and the outermost wall. The outputs of the neural network are the weight of the reflector and the RMS accuracy of the mirror’s face shape under gravity. A neural network consisting of one hidden layer and one output layer is used to build a prediction model for the structural parameters of the reflector. As shown in Figure 8, x1, x2, x3, and x4 represent the thickness of the center, the substrate, the fascia, and the outermost wall of the reflector, respectively, and M and RMS represent the weight of the reflector and the RMS of the mirror face shape accuracy under gravity, respectively.
Because of the high cost of manually modifying finite element model parameters and finite element analysis, as well as the need for different tools and software, the Isight integration platform is one of the most mature platforms in the field of optomechanical integration optimization, which can easily and effectively integrate multidisciplinary software (such as optical error calculation and finite element analysis) with automated processes. Firstly, parameterizing the 3D model through UG and using coarse mesh partitioning can effectively save computational time, while completing attribute settings and constraint definitions to generate a finite element model. Subsequently, utilizing the recording of macro programs and generating loop scripts to achieve automated processes, the application of integrated analysis platforms can greatly improve efficiency. We substitute input variable data into the finite element model for modal analysis, and obtain 400 sets of output responses.

3.3. Establishment of BR-TLDBO-BP Prediction Model

Although BP neural networks have a strong nonlinear mapping ability, they often fall into local optimal solutions, making the prediction less accurate. The BR-TLDBO-BPNN model utilizes the global optimality-searching ability of the DBO algorithm to search for the optimal initial weights and thresholds of the network in its solution space. The position of the dung beetle on each dimension is the solution seeking to establish the mapping relationship between the dimensions of the dung beetle and the weights and thresholds. The BP neural network is trained and examined using the initial weights and thresholds optimized by the TLDBO method and Figure 9 shows the specific flowchart of the BR-TLDBO-BPNN model.
The BR-TLDBO-BPNN model proposed in this study consists of four main parts, as follows:
  • The acquisition of samples: The application of the integrated analysis platform can significantly improve efficiency. The input variable data are substituted into the finite element model for simulation and analysis, and 400 output variable data are obtained.
  • The TLDBO model: Three improvement strategies are proposed in this study to aim at the problems of the original DBO. First, the logistic–tent chaotic mapping initializes the population to maintain diversity, enhancing the algorithm’s ability to find the optimal. Second, the dynamic bipartite group convergence factor with an adaptive inertia threshold balances the algorithm’s local and global optimization-seeking ability. Finally, the Levy flight mechanism is used to optimize the foraging behavior and improve the global search ability.
  • The BR-TLDBO-BPNN model: TLDBO-BPNN utilizes TLDBO to obtain the initial weights w and initial thresholds b of the BPNN, and TLDBO keeps updating the initial weights w and initial thresholds b until it finds the global best position, i.e., the optimal solution. The BR algorithm is used to control the model complexity and constrain the parameters to be optimized to improve the model’s generalization ability.
  • Objective optimization: First, we determine the objective function, then set the constraints and parameter weights, and, finally, use TLDBO to optimize the objective to obtain the best optimization result.

4. Results and Analysis

4.1. Validation of the BR-TLDBO-BPNN Prediction Model

4.1.1. Evaluation Indicators for the Model

In order to measure the prediction effect of the model, the mean absolute error (MAE), mean square error (MSE), root mean square error (RMSE), and mean absolute percentage error (MAPE) are used as the evaluation indices, and the formulae are calculated as follows, respectively:
E M A E = 1 n i = 1 n y i y i E M S E = 1 n i = 1 n ( y i y i ) 2 E R M S E = 1 n i = 1 n ( y i y i ) 2 E M A P E = 1 n s a m p l e s i = 0 n y i y i y i

4.1.2. Analysis of Cross-Validation Results

Table 2 shows the comparison results of the cross-validation results of the two models, and each evaluation index of the model is taken as the average of the K prediction results. As shown in Table 2, compared with the BP neural network, the prediction value of BR-TLDBO-BPNN is closer to the actual value. The mean absolute error ( E M A E ), mean square error ( E M S E ), root mean square error ( E R M S E ), and mean absolute percentage error ( E M A P E ) are reduced by 39.9%, 50.8%, 29.1%, and 31.27%, respectively. The above results show that the BR-TLDBO-BPNN model effectively overcomes the limitation whereby BP neural networks are prone to fall into a local optimization and improves the prediction effect of the model in dealing with complex nonlinear problems, especially in terms of accuracy and stability. This advancement provides a more reliable and efficient computational tool for the optimal design of reflector structures.

4.1.3. Analysis of Forecast Results

The regression analysis plots of BR-TLDBO-BPNN after training are shown in Figure 10, respectively. The fitting curves were obtained by fitting the test data to the predicted values, Y representing the predicted values, T representing the test data, and the dashed line indicating the ideal conditions where the predicted values and test data are equal. The regression analysis plot reflects the fit between the predicted and actual values of the data, and the R-values of all regression lines are very close to 1. The fitting accuracies of the training set, validation set, test set, and the total dataset are all above 99%. It shows an excellent agreement between the actual and predicted values, which ensures that the improved neural network model proposed in this study has a good generalization ability.
Figure 11 shows the comparison between the predicted results and the actual values of the two network models, BPNN and BR-TLDBO-BPNN, and Figure 11a,c show the actual values of each test sample, as well as the predicted values of the two models. It can be observed that the BR-TLDBO-BPNN model predicts closer to the actual values on most of the samples, with a more compact distribution of points showing a minor variance, suggesting that its prediction stability and accuracy may be higher. Figure 11b,d, on the other hand, are scatter plots showing the prediction errors of the two models. It can be seen that the BR-TLDBO-BPNN model has a minor error on most of the test samples, and the error points are mainly concentrated around zero, which indicates its higher prediction accuracy.
The experiment was carried out on a computer with an Intel i7 14600KF processor, Nvidia RTX 3080 graphics card, 32 GB RAM, and MATLAB 2019a. The time of calculating a test sample by ANSYS is about 300 s, while the method proposed in this paper only needs 0.13 s, which shows that the BR-TLDBO-BPNN model has the significant advantages of its running time and high efficiency. In general, the combination of Bayesian regularization and improved dung beetle optimization algorithms helps to improve the generalization ability of the model and avoid overfitting; the BR-TLDBO-BPNN model is superior to the standard BPNN model in prediction performance.
Overall, the BR-TLDBO-BPNN model outperforms the standard BPNN model in terms of predictive performance due to the combined application of Bayesian regularization and improved dung beetle optimization algorithms, which helps to improve the generalization of the model and avoid overfitting.
To analyze the relationship between the influence of each input parameter on the output parameter, the Pearson correlation coefficient method was used to quantify the linear relationship between the inputs and outputs for the established prediction model, and the degree of influence between the two was assessed by a multiple regression analysis. Figure 12 shows the calculated results of the relationship between input and output parameters.
As can be seen in Figure 12a, Y1 (mass) has a significant positive correlation with X1 (center thickness) and X3 (tendon thickness). At the same time, Y2 (RMS) shows a significant negative correlation with the two input parameters, X1 and X3. In particular, the correlation coefficient between X3 and Y1 is as high as 0.81, indicating that, when the tendon thickness increases, the mass also increases, while the correlation coefficient between X3 and Y2 is −0.41, suggesting that an increase in the tendon thickness may lead to a decrease in the RMS. From Figure 12b, it is further known that X3 has a significant positive effect on Y1 with a coefficient of 2.01; on the contrary, X3 has a negative correlation with RMS with a coefficient of −1.1403, which is consistent with the results of the thermogram. Combining the analysis of multiple regression coefficient plots and thermograms, it can be seen that the rib thickness (X3) has the highest sensitivity to mass (Y1) and RMS (Y2) relative to other variables, and the influence of rib thickness on the performance of mirrors should be prioritized in the design process, which is an essential guideline for the optimization of parameter adjustments in the design and manufacturing process.

4.2. Optimization of Reflector Parameters Based on BR-TLDBO-BPNN Model

For a long time, structural design and optimization are often carried out based on the designer’s experience. Considering the optimization scheme’s accuracy and efficiency, this study approximates the relationship between the structural parameters and performance parameters of the reflector through the BR-TLDBO-BPNN model. It uses it to construct the objective function of the mathematical model for reflector optimization and obtains a better combination of structural parameters of the reflector after solving.

4.2.1. Parameter Optimization Process

  • Design variables
Since the objective function in the optimization mathematical model will be used in the neural network model, the optimization design variables must be consistent with the input samples of the neural network data set. x 1 (the center thickness), x 2 (the substrate thickness), x 3 (the fascia thickness), and x 4 (the outermost wall thickness) are selected as the design variables of the optimization mathematical model, which can be expressed as follows:
x = ( x 1 , x 2 , x 3 , x 4 )
  • Objective function
The mirror studied in this study’s face shape accuracy RMS value under the action of gravity load comprehensively reflects the mirror’s stiffness and accuracy retention ability, and the quality affects its processing cost. In this study, the weighted combination of mass and RMS value defines the optimization variables, and the specific formula is:
F ( x ) = w 1 y 1 ( x ) + w 2 y 2 ( x )
f ( x ) = F ( x )
where w 1 and w 2 are the weight, here taken as w 1 = w 2 = 0.5 ; y 1 ( x ) is the mass; and y 2 ( x ) is the RMS.
  • Constraints
The constraints of the reflector optimization mathematical model are as follows:
60   mm     x 1   90   mm       1   mm     x 2   10   mm       1   mm     x 3   10   mm       5   mm     x 4   15   mm
In summary, the optimization problem in this study is to minimize the F ( x ) of the reflector by adjusting the design variables while satisfying the relevant constraints. Therefore, the optimization mathematical model of the mirror is as follows:
f i n d                       x = ( x 1 , x 2 , x 3 , x 4 ) min                   f ( x ) = F ( x )       s . t .             60   mm     x 1   90   mm                             1   mm     x 2   10   mm                           1   mm     x 3   10   mm                           5   mm     x 4   15   mm  

4.2.2. Reflector Structure Optimization Results

The mathematical model of the control rod component optimization is solved using the f min c o n function, one of the optimization toolbox functions that come with MATLAB for solving multidimensional objective optimization problems with constraints. The function is called in the form:
x = f min   c o n = ( f u n , x 0 , A , b , A e q , b e q , l b , u b , n o n l c o n , o p t i o n s )
where x is the optimal solution of optimization; f u n is the objective function of minimization; x 0 is the initial value of optimization variables; A and b are the constraints of linear inequality, which satisfies A x b ; A e q and b e q are the constraints of linear equation, which satisfies A e q = b e q ; l b and u b are the upper and lower bounds of variables; n o n l c o n is the nonlinear constraints; and o p t i o n s is the optimization options, including the number of iterations, the function value, and the algorithm selection.
To verify the effectiveness of the BR-TLDBO-BPNN model in practical application, this algorithm is compared with the Adam algorithm and RMSPROP algorithm which are commonly used in opto-mechanical structure optimization. According to the optimized results of BR-TLDBO-BPNN, the reflector is fabricated and assembled into the optical system, and the reflector is located in the optical axis horizontal state (working state); a 4D dynamic interferometer is used to measure the surface profile accuracy of the mirror, and the results are shown in Figure 13. The BR-TLDBO-BPNN model, Adam algorithm, RMSProp algorithm and experimental mirror parameters x 1 , x 2 , x 3 , x 4 , and corresponding Mass and RMS data are shown in Table 3.
As shown in Figure 14, there are some errors between the three models and the real data measured by the experiment. Among them, the BR-TLDBO-BPNN model used in this paper is the closest to the experimental data after the optimization of the mirror, showing a low level of error. In contrast, the other two models have large deviations from the measured values, which shows that their prediction accuracy is not as good as the BR-TLDBO-BPNN model. Therefore, it can be concluded that the BR-TLDBO-BPNN model used in this paper has significant advantages in the mirror optimization problem, which provides an important reference for future opto-mechanical system design.

5. Conclusions

Aiming at the nonlinear characteristics between the input variables and the output response during the optimization and design process of the mirror structure and the high time cost of the finite element analysis, the BP neural network, BR algorithm, and the improved dung beetle algorithm are applied to establish the output characteristic prediction model (BR-TLDBO-BPNN) of the mirror, and the optimization process of the mirror structure is described in detail. The main conclusions of this study are as follows:
  • The improved DBO algorithm is used to introduce the optimized initial weights and thresholds in BPNN, which effectively improves the prediction performance of BPNN.
  • The BR-TLDBO-BPNN model incorporates the BR algorithm into the BPNN, which enables the BPNN to avoid overfitting and improves the generalization ability of the BPNN.
  • In order to overcome the limitations of the original DBO algorithm, this study proposes the TLDBO algorithm. Logistic-tent chaotic mapping initializes the population and maintains its diversity, enhancing the algorithm’s optimization-seeking ability. Secondly, the algorithm’s local and global optimization ability is balanced by using the dynamic bipartite group convergence factor and adaptive inertia threshold. Again, the Levy flight mechanism is used to optimize the foraging behavior and improve the global search capability. Finally, four other meta-heuristic algorithms are used to compare with the TLDBO algorithm on six benchmark functions. The results show that TLDBO exhibits a good performance for both single-peak and multi-peak functions, verifying the effectiveness of the improved strategy and its competitiveness with other meta-heuristic algorithms.
  • The BR-TLDBO-BPNN model can fit the nonlinear function between the input variables and the output response in the optimal design of mirrors with a prediction error of 99.9%, which provides a high prediction accuracy and reduces a large amount of finite element computation time.
The BR-TLDBO-BPNN model, i.e., Bayesian regularization with the improved dung beetle optimizer for back-propagation neural networks, proposed in this study, is intended for applications in the optimal design of reflector structures. The model exhibits a high prediction accuracy, effectively avoiding overfitting and saving time. The method is able to quickly and efficiently solve for the optimal structural design through predictive modeling, thus ensuring the superiority and functionality of the design prior to manufacturing. In addition, the method is able to adapt to different design requirements and conditions by adjusting the algorithmic parameters, which suggests that it is capable of solving more complex optimization problems in practical industrial applications, such as lens assembly, grating positioning, and the manufacturing design of other precision optical components. Given its excellent performance, the BR-TLDBO-BPNN model was expected to play a vital role in performing the multiphysics field analysis of complex optomechanical structures. In the future, this technique is expected to provide a generalized solution for optical design.

Author Contributions

Conceptualization, K.X. and J.W.; methodology, K.X., J.W. and Y.C.; software, K.X., Y.C. and H.W.; validation, K.X., Y.C., J.W. and H.W.; formal analysis, K.X., Y.C. and H.W.; investigation, K.X., Y.C., J.W. and H.W.; resources, K.X., Y.C. and H.W.; data curation, K.X. and Y.C.; writing—original draft preparation, K.X.; writing—review and editing, K.X., J.W., Y.C. and H.W.; visualization, K.X.; supervision, K.X. and J.W.; project administration, K.X. and J.W.; funding acquisition, J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Jilin Provincial Science and Technology Development Program Key R&D Projects, grant number 20210203156SF, and the Balanced lifting systems, grant number 1Q-2022-010-BX-07-07.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon reasonable request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Rao, B.V.; Kiran, M.; Venkateswaran, R.; Sriram, K.; Narayanamurthy, C. Convex optical freeforms using fringe Zernike overlay approach for two-mirror and three-mirror telescopes for space applications. Opt. Commun. 2023, 541, 129533. [Google Scholar]
  2. Jiang, C.; Chen, Z.; Wang, X.; Zhang, Y.; Cheng, S. Design of compact off-axis triple mirror optical system. Opto-Electron. Eng. 2024, 50, 230231-1–230231-11. [Google Scholar]
  3. Peng, X.; Zhai, D.; Shi, F.; Tian, Y.; Song, C.; Tie, G.; Shen, Y.; Qiao, S.; Shen, X.; Zhang, W.; et al. Design of optical polarization system for defect detection on highly reflective surfaces. Infrared Laser Eng. 2023, 52, 20220863-1–20220863-14. [Google Scholar]
  4. Wang, Z.; Zhang, J.; Wang, J.; He, X.; Fu, L.; Tian, F.; Liu, X.; Zhao, Y. A Back Propagation neural network based optimizing model of space-based large mirror structure. Optik 2019, 179, 780–786. [Google Scholar] [CrossRef]
  5. Ribeiro, T.P.; Bernardo, L.F.A.; Andrade, J.M.A. Topology optimisation in structural steel design for additive manufacturing. Appl. Sci. 2021, 11, 2112. [Google Scholar] [CrossRef]
  6. Frangedaki, E.; Sardone, L.; Marano, G.C.; Lagaros, N.D. Optimisation-driven design in the architectural, engineering and construction industry. Proc. Inst. Civ. Eng.-Struct. Build. 2022, 176, 998–1009. [Google Scholar] [CrossRef]
  7. Kambampati, S.; Gray, J.S.; Kim, H.A. Level set topology optimization of structures under stress and temperature constraints. Comput. Struct. 2020, 235, 106265. [Google Scholar] [CrossRef]
  8. Han, Y.; Xu, B.; Wang, Q.; Liu, Y.; Duan, Z. Topology optimization of material nonlinear continuum structures under stress constraints. Comput. Methods Appl. Mech. Eng. 2021, 378, 113731. [Google Scholar] [CrossRef]
  9. Assran, M.; Aytekin, A.; Feyzmahdavian, H.R.; Johansson, M.; Rabbat, M.G. Advances in asynchronous parallel and distributed optimization. Proc. IEEE 2020, 108, 2013–2031. [Google Scholar] [CrossRef]
  10. van de Schoot, R.; Depaoli, S.; King, R.; Kramer, B.; Märtens, K.; Tadesse, M.G.; Vannucci, M.; Gelman, A.; Veen, D.; Willemsen, J.; et al. Bayesian statistics and modelling. Nat. Rev. Methods Primers 2021, 1, 1. [Google Scholar] [CrossRef]
  11. Ding, S.; Li, R.; Tao, Z. A novel adaptive discrete grey model with time-varying parameters for long-term photovoltaic power generation forecasting. Energy Convers. Manag. 2021, 227, 113644. [Google Scholar] [CrossRef]
  12. Kassaymeh, S.; Al-Laham, M.; Al-Betar, M.A.; Alweshah, M.; Abdullah, S.; Makhadmeh, S.N. Backpropagation Neural Network optimization and software defect estimation modelling using a hybrid Salp Swarm optimizer-based Simulated Annealing Algorithm. Knowl.-Based Syst. 2022, 244, 108511. [Google Scholar] [CrossRef]
  13. Han, K.; Wang, Y. A review of artificial neural network techniques for environmental issues prediction. J. Therm. Anal. Calorim. 2021, 145, 2191–2207. [Google Scholar] [CrossRef]
  14. Abdolrasol, M.G.M.; Hussain, S.M.S.; Ustun, T.S.; Sarker, M.R.; Hannan, M.A.; Mohamed, R.; Ali, J.A.; Mekhilef, S.; Milad, A. Artificial neural networks based optimization techniques: A review. Electronics 2021, 10, 2689. [Google Scholar] [CrossRef]
  15. Zhang, Y.-G.; Tang, J.; Liao, R.-P.; Zhang, M.-F.; Zhang, Y.; Wang, X.-M.; Su, Z.-Y. Application of an enhanced BP neural network model with water cycle algorithm on landslide prediction. Stoch. Environ. Res. Risk Assess. 2021, 35, 1273–1291. [Google Scholar] [CrossRef]
  16. Yang, H.; Li, X.; Qiang, W.; Zhao, Y.; Zhang, W.; Tang, C. A network traffic forecasting method based on SA optimized ARIMA-BP neural network. Comput. Netw. 2021, 193, 108102. [Google Scholar] [CrossRef]
  17. Liu, J.; Zhan, C.; Wang, H.; Zhang, X.; Liang, X.; Zheng, S.; Meng, Z.; Zhou, G. Developing a hybrid algorithm based on an equilibrium optimizer and an improved backpropagation neural network for fault warning. Processes 2023, 11, 1813. [Google Scholar] [CrossRef]
  18. Zhang, X.; Sun, L.; Qi, L. Bayesian Regularization Algorithm Based Recurrent Neural Network Method and NSGA-II for the Optimal Design of the Reflector. Machines 2022, 10, 63. [Google Scholar] [CrossRef]
  19. Ma, W.; Liu, Z.; Kudyshev, Z.A.; Boltasseva, A.; Cai, W.; Liu, Y. Deep learning for the design of photonic structures. Nat. Photonics 2021, 15, 77–90. [Google Scholar] [CrossRef]
  20. Zhang, R.; Zhu, Y.; Liu, Z.; Feng, G.; Diao, P.; Wang, H.; Fu, S.; Lv, S.; Zhang, C. A Back Propagation Neural Network Model for Postharvest Blueberry Shelf-Life Prediction Based on Feature Selection and Dung Beetle Optimizer. Agriculture 2023, 13, 1784. [Google Scholar] [CrossRef]
  21. Wright, L.G.; Onodera, T.; Stein, M.M.; Wang, T.; Schachter, D.T.; Hu, Z.; McMahon, P.L. Deep physical neural networks trained with backpropagation. Nature 2022, 601, 549–555. [Google Scholar] [CrossRef] [PubMed]
  22. Panda, S.; Panda, G. Fast and improved backpropagation learning of multi-layer artificial neural network using adaptive activation function. Expert Syst. 2020, 37, e12555. [Google Scholar] [CrossRef]
  23. Xiao, M.; Liao, Y.; Bartos, P.; Filip, M.; Geng, G.; Jiang, Z. Fault diagnosis of rolling bearing based on back propagation neural network optimized by cuckoo search algorithm. Multimed. Tools Appl. 2022, 81, 1567–1587. [Google Scholar] [CrossRef]
  24. Yang, B.; Li, D.; Zeng, C.; Chen, Y.; Guo, Z.; Wang, J.; Shu, H.; Yu, T.; Zhu, J. Parameter extraction of PEMFC via Bayesian regularization neural network based meta-heuristic algorithms. Energy 2021, 228, 120592. [Google Scholar] [CrossRef]
  25. Wali, S.A.; Tyagi, A. Comparative study of advance smart strain approximation method using levenberg-marquardt and bayesian regularization backpropagation algorithm. Mater. Today Proc. 2020, 21 Pt 2, 1380–1395. [Google Scholar] [CrossRef]
  26. Nguyen, D.H.; Le, X.H.; Anh, D.T.; Kim, S.H.; Bae, D.H. Hourly streamflow forecasting using a Bayesian additive regression tree model hybridized with a genetic algorithm. J. Hydrol. 2022, 606, 127445. [Google Scholar] [CrossRef]
  27. Xue, J.; Shen, B. Dung beetle optimizer: A new meta-heuristic algorithm for global optimization. J. Supercomput. 2023, 79, 7305–7336. [Google Scholar] [CrossRef]
  28. Chang, Z.; Luo, J.; Zhang, Y.; Teng, Z. A Mixed Strategy Improved Dung Beetle Optimization Algorithm and Its Application. PREPRINT (Version 1). Available online: https://www.researchsquare.com/article/rs-2988123/v1 (accessed on 1 June 2023).
  29. Shi, Y.; He, T.; Zhong, J.; Mei, X.; Li, Y.; Li, M.; Zhang, W.; Ji, D.; Su, L.; Lu, T.; et al. Classification and rapid non-destructive quality evaluation of different processed products of Cyperus rotundus based on near-infrared spectroscopy combined with deep learning. Talanta 2024, 268, 125266. [Google Scholar] [CrossRef] [PubMed]
  30. Zheng, J.; Lu, C.; Hao, C.; Chen, D.; Guo, D. Improving the generalization ability of deep neural networks for cross-domain visual recognition. IEEE Trans. Cogn. Dev. Syst. 2020, 13, 607–620. [Google Scholar] [CrossRef]
Figure 1. BP network structure.
Figure 1. BP network structure.
Electronics 13 02789 g001
Figure 2. Logistic–tent chaotic mapping population initialization: (a) scatter plot and (b) histogram of frequency distribution.
Figure 2. Logistic–tent chaotic mapping population initialization: (a) scatter plot and (b) histogram of frequency distribution.
Electronics 13 02789 g002
Figure 3. Comparison of different convergence factors.
Figure 3. Comparison of different convergence factors.
Electronics 13 02789 g003
Figure 4. Levy, Gaussian, and Cauchy distributions.
Figure 4. Levy, Gaussian, and Cauchy distributions.
Electronics 13 02789 g004
Figure 5. Three-dimensional view of single-peak and multi-peak functions.
Figure 5. Three-dimensional view of single-peak and multi-peak functions.
Electronics 13 02789 g005
Figure 6. Convergence curves for single- and multi-peak functions.
Figure 6. Convergence curves for single- and multi-peak functions.
Electronics 13 02789 g006
Figure 7. Mirror finite element analysis.
Figure 7. Mirror finite element analysis.
Electronics 13 02789 g007
Figure 8. Mirror structure model.
Figure 8. Mirror structure model.
Electronics 13 02789 g008
Figure 9. BR-TLDBO-BPNN model flowchart.
Figure 9. BR-TLDBO-BPNN model flowchart.
Electronics 13 02789 g009
Figure 10. Network regression analysis of BR-TLDBO-BPNN model.
Figure 10. Network regression analysis of BR-TLDBO-BPNN model.
Electronics 13 02789 g010
Figure 11. Predicted vs. actual values for different benchmark models: (a) comparison of predicted and true values of mass, (b) comparison of mass errors of different models, (c) comparison of predicted and true values of RMS, and (d) comparison of RMS errors of different models.
Figure 11. Predicted vs. actual values for different benchmark models: (a) comparison of predicted and true values of mass, (b) comparison of mass errors of different models, (c) comparison of predicted and true values of RMS, and (d) comparison of RMS errors of different models.
Electronics 13 02789 g011
Figure 12. Relationship between the influence of input and output parameters: (a) heat map and (b) multiple regression coefficient plots.
Figure 12. Relationship between the influence of input and output parameters: (a) heat map and (b) multiple regression coefficient plots.
Electronics 13 02789 g012
Figure 13. Experimental testing: (a) physical drawing of the reflector and (b) interferometer inspection of the accuracy of the reflector face type.
Figure 13. Experimental testing: (a) physical drawing of the reflector and (b) interferometer inspection of the accuracy of the reflector face type.
Electronics 13 02789 g013
Figure 14. Comparison of errors of different models.
Figure 14. Comparison of errors of different models.
Electronics 13 02789 g014
Table 1. Single-peak and multi-peak benchmark test functions.
Table 1. Single-peak and multi-peak benchmark test functions.
Function CategoriesFormula ExpressionDimensionSearch ScopeOptimum Value
Single-peak function f 1 ( x ) = i = 1 n x i 2 30[−100,100]0
f 2 ( x ) = i = 1 n | x i | + i = 1 n | x i | 30[−10,10]0
f 3 ( x ) = i = 1 n ( j = 1 i x j ) 2 30[−100,100]0
Multi-peak function f 4 ( x ) = i = 1 n x i sin | x i | 30[−500,500]0
f 5 ( x ) = i = 1 n x i 2 10 cos ( 2 π x i + 10 ) 30[−5.12,5.12]0
f 6 ( x ) = 1 4000 i = 1 n x i 2 i = 1 n cos ( x i i ) + 1 30[−600,600]0
Table 2. Comparison of cross-validation results.
Table 2. Comparison of cross-validation results.
Evaluation IndicatorsBP Predictive ModelBR-TLDBO-BPNN Predictive Modeling
E M A E 0.1710.118
E M S E 0.0630.031
E R M S E 0.2510.178
E M A P E 0.7930.545
Table 3. Comparison results of optimization algorithms.
Table 3. Comparison results of optimization algorithms.
Algorithm x 1 x 2 x 3 x 4 MassRMS
BR-TLDBO-BPNN75 mm2 mm2 mm6 mm15.55 kg16.41 nm
Adam81 mm3 mm6 mm7 mm18.93 kg19.17 nm
RMSProp78 mm4 mm3 mm8 mm17.09 kg20.38 nm
Experimental results75 mm2 mm2 mm6 mm15.13 kg15.19 nm
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xue, K.; Wang, J.; Chen, Y.; Wang, H. Improved BP Neural Network Algorithm for Predicting Structural Parameters of Mirrors. Electronics 2024, 13, 2789. https://doi.org/10.3390/electronics13142789

AMA Style

Xue K, Wang J, Chen Y, Wang H. Improved BP Neural Network Algorithm for Predicting Structural Parameters of Mirrors. Electronics. 2024; 13(14):2789. https://doi.org/10.3390/electronics13142789

Chicago/Turabian Style

Xue, Kejuan, Jinsong Wang, Yuan Chen, and Hao Wang. 2024. "Improved BP Neural Network Algorithm for Predicting Structural Parameters of Mirrors" Electronics 13, no. 14: 2789. https://doi.org/10.3390/electronics13142789

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop