1. Introduction
Modern engineering structures such as aircraft, high-rise buildings, and new bridges operate in complex environments and are subjected to various external loads, leading to potential structural damage and failure [
1]. In particular, aircraft require precise load assessments to balance safety and design costs. Inadequate load assessments can compromise safety margins, causing premature failures, while overly conservative assessments increase design costs due to excess weight [
2]. Accurate load information is essential for ensuring structural safety and minimizing damage during normal operations. External load information can be obtained in two main ways: direct measurement [
2] and indirect measurement. Direct measurement involves using multiple sensors to measure load information directly when installation is feasible. However, due to harsh conditions, functional requirements, and spatial or temporal constraints, direct measurement is often impractical. Therefore, the common approach in engineering is to indirectly determine structural loads by measuring more accessible structural responses, such as strain, displacement, velocity, and acceleration. This method involves back-calculating external excitations from known structural responses and system characteristics, which is a mechanical inverse problem in structural dynamics [
3].
Load identification, originating in aviation in the 1970s [
4], remains in its early stages despite extensive research. Typically, load identification involves three steps: establishing a mechanical system model, measuring structural responses, and selecting an identification method. This process spans multiple disciplines, including modeling, structural vibration analysis, and computational inverse methods. Recently, neural networks [
2,
5,
6,
7,
8,
9] have gained attention for rapid load identification, addressing the inefficiencies of traditional methods. Neural networks do not require theoretical expressions between vibration responses and loads. With sufficient data, the network learns the relationship between structural response and load. This is particularly useful for complex nonlinear systems. In 2000, Sofyan and Trivailo [
10] used back propagation neural networks (BPNN) to identify aerodynamic loads on thin plate structures and developed general-purpose software for multi-type dynamic load identification problems. In 2006, Trivailo [
11] introduced Elman networks to solve the problem of identifying buffeting loads and maneuvering loads on aircraft tail wings, and the identification results showed that Elman networks have good identification performance for aerodynamic load identification problems. In 2015, Samagassi et al. [
12] used wavelet vector machines to identify multiple impact loads on linear elastic structures and achieved the identification of multiple impact loads on a simply supported beam model under a single response. In 2018, Cooper and Dimaio [
13] achieved the identification of static loads on large wing ribs of an aircraft using feedforward neural networks (FNNs). In 2019, Chen et al. [
14] used deep neural networks (DNNs) to achieve impact load identification, where they used the damage deformation characteristic parameters of a hemispherical shell structure and various characteristic parameters of impact loads as inputs and outputs of the DNN network for load identification. In 2022, Yang [
8] proposed a dynamic load identification method based on deep dilated convolutional neural networks (DCNN). This method can be used as a filter in dynamic load identification due to its strong noise immunity in convolutional layers. To address the issue of rapid fatigue life depletion caused by airframe flutter, Candon [
15] focused on a single-reference aircraft under a multi-input single-output (MISO) load monitoring scenario. Using strain as the input, the goal was to predict the representative bending and torsional dynamics as well as the quasi-static load spectra on the aircraft wings during transonic flutter maneuvers. A suite of machine learning models was consolidated, including linear regression models, traditional artificial neural networks, and deep learning strategies. In 2024, Liu [
7] proposed a new hybrid model-data-driven interval structure dynamic load identification framework, which seamlessly combines finite element modeling with machine learning techniques. When facing challenges such as limited training data, significant noise interference, and non-zero initial conditions, this method can also improve the accuracy, robustness, and generalization of dynamic load identification.
Traditional neural network methods can accurately establish a mapping bridge between structural responses and real loads, and while the mapping accuracy may be high, it is not credible when faced with data that deviate from true values. Due to the point-to-point input–output mapping in traditional neural network architectures, slight variations in the structure itself or minor disturbances in measurements can cause significant deviations in the performance of a well-trained traditional neural network, potentially rendering it completely ineffective. Dispersed data might all be correct but not entirely accurate. The fixed architecture of traditional neural networks can become confusing and ineffective when faced with the variability of a population of structures, as it struggles to adequately respond to and represent such diversity. In such cases, INNs [
16,
17,
18,
19] can effectively solve the modeling problem of uncertain data. INNs, based on the combination of interval analysis and traditional point-value neural networks, use interval numbers to characterize subjective and objective uncertainties, use traditional point-value neural networks to complete physical process modeling, and achieve the modeling and quantification of imprecise data in a clever combination. By using new modeling methods such as an INN in system design, it is possible to effectively avoid model structural errors and other requirements. Some scholars [
16,
20] have studied interval prediction models for neural network prediction, but the existing studies are all based on point-value neural networks to achieve interval value prediction, which is essentially still a statistical regression. Point-value neural network-based interval prediction is difficult to provide a comprehensible explanation during training, and the uncertainty of system inputs and parameters is difficult or unable to be quantified. INNs are a generalization of interval prediction models and can provide a comprehensive prediction result, not just a point value. In 1991, Ishibuchi and others [
21] proposed the backpropagation neural network (BPNN) model with interval input and point-value network parameters for the first time. They also provided forward calculation and backward learning algorithms based on interval analysis theory. Since then, INNs have begun to enter the academic research field. In 1998, Beheshti [
19] et al. defined the concept of INN for the first time and divided the training problem of INNs into two categories: directly obtaining numerical solutions and solving nonlinear optimization problems. In 2000, Garczarczyk [
22] proposed the gradient descent algorithm for the 4-layer INN with interval weights and thresholds. Yao [
23] proposed an improved INN based on the Widrow–Hoff learning rule in 2004 and combined it with the time-delay neural network to construct a dynamic INN model. In 2009, Campi, Calafiore, and Garatti et al. [
24] developed an interval prediction model for supervised learning, which can perform interval predictions with guaranteed accuracy. In 2019, Sadeghi et al. [
25] proposed a neural network backpropagation algorithm with interval predictions, which uses a small batch of random gradient descent algorithms to train and quantify the uncertainty of the neural network, thereby achieving constant-width interval predictions. In 2022, Saeed [
26] proposed two new interval prediction frameworks using independent recurrent neural networks. This method employs Gaussian functions centered around the predictions of point forecast models and the estimation errors of error forecast models to estimate the prediction intervals. The average coverage width index improved by 43% and 12%, respectively, compared to traditional models and models based on Long Short-Term Memory (LSTM). Shao [
27] introduced a new approach to obtain reliable predictions from the perspective of pattern classification. A novel hybrid framework was established, composed of nested LSTM, Multi-Head Self-Attention (MHSA) mechanisms, CNNs, and a feature space identification method, aimed at robust interval forecasting.
However, subsequent research on INNs has not developed or flourished as expected. One reason is the difficulty in expanding the application scenarios of INN in practical use, which significantly diminishes the motivation for their research. Nonetheless, as structural design increasingly emphasizes robustness alongside precision, the powerful capability of INN to handle uncertain data positions them as a potential breakthrough technology. In this paper, the significant uncertainty characteristics of load identification in a population of structures align well with the capabilities of INN, yet this area remains largely unexplored. In order to fill this gap, this study proposed a study on the INN method for uncertain load identification problems, which takes into account these potential deviations and quantifies the uncertainties involved, making the data model more robust. Even in the face of significant uncertainties within a population of structures, these methods can still grasp the main contradictions, establish interval mapping relationships, and address the credibility issues of the data and identification results.
The main contributions and innovations of this paper are summarized as follows:
(1) The INN framework is implemented for load identification in a population of structures for the first time;
(2) A global optimization algorithm is proposed to solve the problem of interval weight and threshold in INN;
(3) To reasonably evaluate the accuracy and credibility of the identification results, an improved loss function metric CMSC combining interval coverage rate and interval width is established.
2. Problem Description
In the aerospace industry, static and quasi-static concentrated loads are common types of loads encountered. During aircraft design and validation, static and quasi-static load tests are indispensable for simulating the loads that aircraft may encounter under normal flight or extreme conditions. During landing, the landing gear and aircraft structure are subjected to impact loads. Although these loads occur over a short duration, their nature is more quasi-static because they are predictable and repetitive. When weapons, fuel tanks, or other equipment are mounted under the wings or on the fuselage of an aircraft, the additional loads generated also constitute concentrated loads. Additionally, while parked, aircraft are subject to wind forces. Although these forces are dynamically changing, they can be approximated as quasi-static loads in conditions of low or slowly changing wind speeds.
The study of these loads is crucial for the design, safety, and performance of aircraft. By understanding and analyzing these static and quasi-static loads, engineers can better design aircraft structures to ensure their safe operation in all anticipated operating environments. The mathematical description of static concentrated load identification is shown in Equation (1).
In this equation, represents the concentrated load, represents the measured response, and represents the physical parameters of the structural system. It is evident that the mapping relationship between structural response and load bearing is influenced by the parameters of the structural system.
Structures of similar design are considered as a population, where parameter
variations in geometry, materials, and boundary conditions cause variable responses under identical loads, appearing as scatter points (
Figure 1). Additionally, experimental measurements introduce errors.
Identifying the real load from scattered experimental responses under uncertain conditions is a significant and challenging problem. Load identification methods under uncertainty aim to deduce the range of potential loads from the input scatter of experimental responses, ensuring the real load falls within this identified range. By employing a data-driven approach, neural networks offer an alternative to traditional mechanistic models. These networks, which mimic the human brain’s structure and information processing, are capable of self-learning, organizing, and adapting. Unlike mechanistic models, neural network-based models simplify the structure and bypass the need to understand the physical significance, directly mapping the relationships between sensor responses and load data.
This paper introduces an INN approach for structural load identification that leverages its ability to quantify uncertainty. The INN model uses dispersed measurement data from similar structures to provide a range of potential load values, accommodating uncertainties in geometry, materials, and measurement errors. This model does not make assumptions about the data by enhancing its generalization, robustness, or credibility. The neural network parameters are set as interval values, allowing the output to also be intervals based on interval arithmetic. This INN framework offers reliable prediction intervals even with limited data samples, effectively addressing load identification for a population of structures, as follows:
In the INN model for load identification, represents the identified load interval, represents the measured responses, and represents the physical interval parameters of the structural system, which correspond to the interval weights and interval thresholds within the INN. The learning process of the network parameters corresponds to the updating process of the uncertain physical model. This substitution allows the INN to adaptively refine its estimates based on the variability and uncertainty inherent in the input data, thereby improving the credibility and accuracy of the load identification under uncertain conditions.
4. Interval Back Propagation Neural Network Architecture for Load Identification
To ensure the generality and adaptability of the proposed INN architecture in practical applications, this paper does not introduce additional interval quantification methods. Instead, it maintains the original point-value form of the structure response–load data pairs during load identification. The load intervals identified by the interval back propagation neural networks (IBPNN) should contain the real load point values. The structure and learning algorithm of the IBPNN are as follows, with
Figure 3 illustrating a typical three-layer IBPNN.
In this context,
represents the input measured response data,
and
denote the interval weight values, and
and
represent the interval threshold values. The output
describes the interval identified load. The subscripts
,
, and
indicate the indices corresponding to the respective layers, where
. At this stage, the network’s weights and thresholds are set as interval values, while the input data to the network are in point-value form. The output of the network, as per Equation (3), is generated accordingly.
In this setting,
and
represent the upper and lower bounds of the interval parameter, respectively. The output corresponding to the hidden layer is denoted as
. The upper and lower bounds of its hidden layer output are
, respectively, as Equation (4).
By selecting the unipolar Sigmoid function
as the activation function, where the range is set as
, it follows that
. Combining Equations (3) and (4) reveals the specific form of the interval-identified load
, whose lower and upper bound can be expressed as
5. Solving INN: Loss Function Construction and Interval Parameter Optimization
In addition to the architecture of INN, the effectiveness of load identification is also influenced by the construction of the loss function and the optimization of network parameters. This chapter integrates the strengths of traditional INN and point value neural networks, proposing an improved loss function indicator. A global optimization algorithm is employed for training the network parameters, ultimately achieving efficient and credible interval load identification results.
5.1. Loss Function Based on Quality Assessment Metrics for Predictive Intervals
Traditional neural networks typically use metrics such as Mean Squared Error (MSE) as the loss function for network training. The original loss function can be constructed as follows:
Here,
represents the sample size and
denotes the number of network outputs.
represents the real load in the samples. However, for INN, the identified results are interval values while the real outputs are point values. Therefore, when constructing the loss function for INN, the metrics need to be adapted to account for the interval nature of the predictions. The loss function Interval Mean Squared Error (IMSE) for the IBPNN is defined as follows:
In this context, and represent the lower and upper bounds of the identified load, respectively. Similarly, when the real load is in point value form, we have .
However, modifying the traditional neural network loss function into an interval form and directly applying it to INN is still inappropriate. The loss function based on IMSE does not reflect the characteristics of the interval-enveloped target values, as illustrated in
Figure 4. In this figure, the upper bound of the load identification interval is at a distance of
from the real load sample
, while the lower bound is at a distance of
from
. In two different interval scenarios, both
and
are consistent, resulting in identical IMSE function values. However, the coverage of the intervals differs significantly. Clearly, scenario 1 is more desirable.
The researchers [
28] introduce the Coverage Width Criterion (CWC) as an evaluation metric for the quality of prediction intervals, used to construct a loss function for INN learning. CWC is related to two other metrics, Prediction Interval Coverage Probability (PICP) and the mean prediction interval width (MPIW). PICP is the most commonly used metric for assessing interval quality. It describes the extent to which the target value is contained within the prediction interval and is the primary indicator for evaluating the quality of prediction intervals. The PICP value directly reflects the accuracy of the prediction interval and represents the percentage of sample target values included in the prediction interval.
Mathematically, PICP is defined as
where
is the number of samples and
and
represent the lower and upper bounds of the
i-th prediction interval. If the
i-th prediction interval encloses the corresponding sample value, then
; otherwise,
as shown in
Figure 5. The PICP value of the identified interval is closely related to the interval width. When using the PICP metric to evaluate the quality of the prediction interval, it is evident that high-quality intervals can be easily obtained by simply widening the interval. However, overly wide prediction intervals have limited practical value and may not meet engineering requirements.
Therefore, in addition to using PICP as an interval quality evaluation metric, it is necessary to introduce a metric related to interval width, namely MPIW, to characterize the quality of the prediction interval. Its mathematical definition is as follows:
Here, represents the prediction interval width for the i-th sample. The MPIW can provide sensitive information about the prediction interval’s responsiveness to changes in the real target value. Specifically, under the same PICP, a smaller MPIW indicates a higher quality prediction interval.
From Equations (8) and (9), it is evident that an interval with a high PICP value and a small MPIW value represents a high-quality prediction interval. The metrics PICP and MPIW can describe and measure the quality of the prediction interval from different perspectives. However, increasing PICP typically results in a wider MPIW, while narrowing MPIW often leads to a lower PICP. Our goal is to construct a prediction interval with high coverage (larger PICP) and sufficient narrowness (smaller MPIW), ensuring its ability to quantify and express data errors effectively. Therefore, when using prediction interval quality metrics to construct the INN loss function, these two metrics should be integrated into a comprehensive measure. Based on the above analysis, the INN loss function integrating both quality evaluation metrics [
28] is constructed as follows:
An analysis of Equation (10) reveals that when the PICP is less than the predefined credibility level , the exponential term will exceed 1, causing the CWC to increase sharply with the PICP. In this scenario, PICP serves as the primary evaluation metric for prediction intervals, aligning with the priority of considering the weight percentage of PICP. Conversely, when the PICP exceeds , the curve of the exponential term gradually flattens as the PICP value increases, leading to a gradual decrease in its weight and MPIW becomes the primary optimization target. represents the weight amplification coefficient for the PICP.
However, the current construction of CWC still has issues. When the upper and lower bounds of the interval are the same, the MPIW remains constant at 0, resulting in a loss function of 0 and rendering the training ineffective. Even if MPIW is not 0, there is no clear rule for selecting intervals with the same interval width, which does not align with practical physical situations. In real-world engineering problems, the sample points are densely distributed around the real value and sparsely distributed far from it. Therefore, this paper proposes an improved coverage and mean square criterion (CMSC), where the MPIW in the formula is replaced by IMSE. This modification aims to have the real load values as close as possible to the median of the identification intervals while fulfilling the requirements of interval coverage rate and interval width. The improved form of CMSC is presented below, as follows:
During the INN learning process, if the obtained PICP value is lower than the set threshold value , the resulting prediction intervals may be misleading, emphasizing the optimization of the prediction interval coverage rate. Conversely, when the obtained PICP reaches the specified level, the INN based on the CMSC criterion optimizes IMSE to reduce the uncertainty in the network’s interval outputs.
5.2. Optimization of Parameters in INN
The fundamental concept of INN involves generating interval output results through neural networks. For interval prediction, algorithms like the LUBE (Lower Upper Bound Estimation) algorithm [
28] construct a neural network model with dual output neurons. In this model, the two outputs of the neural network represent the upper and lower bounds of the prediction interval, thereby forming the output interval. The INN model proposed in this chapter achieves interval results by utilizing interval weights and thresholds within the network.
Figure 6 illustrates the distinctions between these two methods. Compared to the LUBE algorithm, the INN architecture used in this study offers greater fitting flexibility due to the interval form of weights and thresholds, even with the same number of layers and neurons.
In this chapter, the CMSC is used as the objective function for the IBPNN model. Considering extensive interval calculations and the abrupt nature of CMSC, gradient-based algorithms are not particularly suitable for optimizing interval weights and thresholds. The GA originates from computer simulations of biological systems. It is a stochastic global search and optimization method that mimics the evolutionary mechanisms of nature, drawing inspiration from Darwin’s theory of evolution and Mendel’s principles of genetics. Essentially, GA is an efficient parallel global search method that can automatically acquire and accumulate knowledge about the search space during the search process and adaptively control the search process to find the optimal solution.
Given the advantages of GA in discovering optimal solutions, avoiding local optima, and ensuring convergence, it is chosen for network optimization. Let the interval weight matrix be and the threshold vector be . The optimization goal for the INN is to determine the optimal network such that the value of its loss function CMSC is minimized.
The steps for the genetic optimization of optimal interval network parameters are as follows:
(1) Encode the interval network parameters to be optimized into chromosomes, and set the loss function CMSC of the training data as the fitness function. (2) Evaluate the fitness of each chromosome corresponding to an individual. (3) Following the principle that the higher the fitness, the greater the probability of selection, select two individuals from the population as parents. (4) Extract the chromosomes of the parents and perform crossover to produce offspring. (5) Mutate the chromosomes of the offspring. (6) Repeat steps 2, 3, and 4 until the optimal population is produced, thereby obtaining the optimized network parameters and ultimately achieving high-quality prediction intervals.
Figure 7 illustrates the overall flowchart of the INN model for centralized load identification based on the CMSC function. The detailed steps are described as follows.
The structural response is selected and load data are measured under different loading conditions from multiple individuals of the same structural population as the data source for the INN model. Partition the data into training and testing sets.
Determine the input and output dimensions of the INN model based on the dimensionality of the load identification problem to be solved. Set algorithm-related parameters such as the maximum number of iterations and population size for GA. Initialize the interval weights and thresholds of the INN model. Set the credibility level and amplification factor in the CMSC-based network loss function.
Determine the optimal number of neurons in the hidden layer using empirical formulas. Set the fitness function of the GA to be the normalized CMSC index of the INN prediction interval. Once the GA meets the termination conditions, obtain the optimal interval weights and thresholds of the INN model, complete the training, and output the optimal network.
Based on the optimized IBPNN model obtained in step 3, input the structural response data from the test set. Predict and identify the load intervals. Evaluate the identification performance using PICP, IMSE, and CMSC metrics.
7. Conclusions
In engineering practice, similar structures exhibit uncertainties in material properties and geometric dimensions. Additionally, observational errors introduced during experimental measurements lead to dispersed responses among these structures under the same load, making accurate load identification challenging. Traditional point-value-based neural networks, constrained by a fixed data structure, struggle to address these uncertainties.
Therefore, this paper introduces the interval neural network method, which internalizes the weights and thresholds in the network, enhancing its robustness in dealing with model and data uncertainties while maintaining recognition accuracy. Furthermore, the paper presents an improved CMSC loss function metric. This function is shown to balance the interval coverage rate of the real load and the interval width, thereby making the recognized interval median closer to the real load, outperforming traditional point-value neural network loss functions like MSE and conventional interval neural network loss functions like CWC. Considering the characteristics of interval operations and the step-like nature of the loss function, a genetic algorithm is employed for the global optimization of interval network parameters. This ultimately achieves credible interval load identification under uncertain influences.
In summary, the paper, through numerical and experimental case studies, achieves uncertainty modeling and quantification in load identification problems, with an interval coverage rate greater than 95%. This demonstrates the method’s accuracy, credibility, and strong interference resistance when studying the performance of structural population, thereby expanding its application prospects in engineering.