1. Introduction
Containerization has provided a favorable environment for international trade, and specialization and technological advances have simplified international shipping [
1]. The China containerized freight index (CCFI) was released by the Shanghai Shipping Exchange in 1998 to objectively reflect changes in Chinese container freight rates. As the number of shipping containers continues to increase, the containerized freight index in China has a significant impact on the world economy [
2]. In recent years, an increasing number of scholars have focused on predicting the containerized freight index [
3]. Forecasting models that can accurately and stably predict the containerized freight index are becoming increasingly important and have considerable economic significance [
4].
A traditional econometric model is a theoretical structure that describes the relationships between relevant economic variables. Jeon et al. [
5] employed system dynamics to reflect the supply-side and demand-side drivers of the market and proposed a one-step prediction method by utilizing the cyclical system dynamics approach. Koyuncu and TAVACIOĞLU [
6] applied Holt–Winter smoothing and the Seasonal Autoregressive Integrated Moving Average (SARIMA) to forecast the Shanghai Containerized Freight Index (SCFI), which revealed that the SARIMA model has excellent prediction performance in short-term forecasting. Kawasaki et al. [
7] discussed the application of the SARIMA and Vector Autoregression (VAR) models in predicting container trade volume. Hirata and Matsuda [
8] used the SARIMA and Long Short-Term Memory (LSTM) methods to predict SCFI, which demonstrated that the performance of the SARIMA model is better than that of LSTM in predicting cargo volumes along the western and eastern routes of Japan. However, limited data for building models and unrealistic assumptions regarding future conditions have led to the poor performance of traditional economic models in short-term forecasting.
Artificial intelligence models are established based on mathematical algorithms to create prediction models that can use specific algorithms to analyze data and learn from data patterns to generate models. In recent years, deep learning models have been widely applied to time-series forecasting. Shankar et al. [
9] used LSTM to predict container throughput. Kim and Choi [
10] used Gate Recurrent Units (GRUs) and LSTM to construct a prediction model for CCFI, which illustrated the improvement of the artificial intelligence models compared to the Autoregressive Integrated Moving Average (ARIMA) in terms of prediction accuracy. Dasari and Bhukya [
11] analyzed the capability of Convolutional Neural Networks (CNNs) and CNN-LSTM in automatic feature extraction, which demonstrated that CNNs have a unique advantage in feature extraction. Khaksar Manshad et al. [
12] discovered a novel time-series link prediction method based on evolutionary computation and irregular cellular learning automata. Swathi et al. [
13] observed the satisfactory performance of LSTM model-based sentiment analysis in predicting time-series data. Xiao et al. [
14] used LSTM and ensemble learning technology to predict the China Coastal Bulk Coal Freight Index. Liang et al. [
15] transformed time-series data into network structures, selected relevant features with the Maximum Relevance, Minimum Redundancy (mRMR) method and applied Support Vector Machine (SVM), Deep Neural Network (DNN), and LSTM models to improve prediction accuracy. However, training artificial intelligence models requires a substantial amount of data for learning, and the quality and quantity of the data have a significant impact on performance.
Numerous models have been employed in the prediction field, and decomposition ensemble prediction model is one of the most commonly used models. Because the containerized freight index has strong nonlinearity and volatility, some scholars have proposed the use of data decomposition technology to decompose complex time series, which has the potential to enhance both the accuracy and stability of predictions. There are many data decomposition methods, such as wavelet packet decomposition [
16], wavelet soft-threshold denoising [
17], Ensemble Empirical Mode Decomposition (EEMD) [
18], Complementary Ensemble Empirical Mode Decomposition (CEEMD) [
19], Improved Complete Ensemble Empirical Mode Decomposition (ICEEMD) [
20], Variational Mode Decomposition (VMD) [
21], and Discrete Fourier Transform [
22]. Specifically, Bai et al. [
23] decomposed raw stock price-based Multivariate Empirical Mode Decomposition (MEMD) and the Neighborhood Rough Set (NRS). Rezaei et al. [
24] discussed Empirical Mode Decomposition (EMD) and CEEMD performance, and Liu et al. [
25] analyzed the results of Interval Discrete Wavelet Transform (IDWT), Interval Empirical Mode Decomposition (IEMD), and Interval VMD (IVMD) under different models. Nguyen and Phan [
26] applied EEMD to decompose time-series data to achieve excellent performance. Yang et al. [
27] introduced VMD and Improved Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (ICEEMDAN) to the prediction model, which illustrated that data preprocessing technology has the benefits of denoising. Jaseena and Kovoor [
28] applied the discrete wavelet transform algorithm to decompose raw data, which proved that Discrete Wavelet Transformation (DWT) is effective in enhancing the prediction model. Li et al. [
29] considered the robustness of the decomposition algorithm to improve the accuracy of the prediction model. Da Silva et al. [
30] presented an ensemble learning model that incorporates variable mode decomposition and singular spectrum analysis. Fang [
31] applied the EMD method to a backpropagation (BP) neural network and ARIMA and used EMD-BP and EMD-ARIMA models to predict the containerized freight index. Chen et al. [
32] decomposed the CCFI into multiple subseries (IMF) using the EMD method and predicted each IMF using the gray method and theAutoregressive Moving Average (ARMA) method separately, exploring the benefits of gray wave forecasting methods using periodic fluctuations.
There have also been many research achievements in predictors and ensemble schemes. For example, Tian and Chen [
33] used a reinforced LSTM model to predict decomposed modes, and a modified sparrow search algorithm was applied to optimize the hyperparameters of an LSTM model. Zhao et al. [
34] proposed a hybrid deep learning model based on CNN, VMD, and GRUs and validated the model using wind power. Niu et al. [
35] proposed a novel decomposition ensemble model and compared its accuracy with that of simple Recurrent Neural Network (RNNs), GRUs, and LSTM, which indicated that the EEMD-RNN model can improve accuracy for time-series forecasting. Altan et al. [
36] used LSTM to predict subseries, and prediction results were integrated using Multi-objective Gray Wolf Optimization (MOGWO), which indicated that the ensemble scheme is able to capture the inherent features of the time series and achieve more accurate prediction results than individual prediction models. Joseph et al. [
37] predicted subseries using a bi-directional LSTM network. Luo and Zhang [
38] demonstrated that Bi-directional LSTM (BiLSTM) can analyze the intrinsic characteristics of longer temporal-frequency information, and the attention mechanism can assign more reasonable weights to parameters, which can effectively enhance the accuracy of prediction results. Yang et al. [
39] proved that recursive empirical mode decomposition and LSTM perform well in time-series data. Yu et al. [
40] presented a new model based on LSTM and a time-series cross-correlation network, which achieved excellent prediction results in a wind power field. Yang et al. [
41] employed Extreme Learning Machine (ELM) series models to establish a novel forecasting model and conducted experimental simulations, showing that the ensemble model can achieve excellent performance. Wang et al. [
42] determined VMD parameters using Kullback–Leibler divergence and an applied LSTM-based attention module to predict stock prices. Hao et al. [
43] introduced an advanced prediction model for air pollutant levels that integrates time-series reconstruction, sub-model simulation, weight optimization, and combination techniques to improve both prediction accuracy and stability in environmental management.
By examining the pertinent literature on prediction models and the containerized freight index, these studies show that the decomposition ensemble model has many advantages in containerized freight index prediction, and an increasing number of scholars are focusing on it. However, many problems still exist in forecasting the Containerized Freight Index (CFI). This subsection summarizes some of the limitations of CFI forecasting.
From the above literature analysis, most of the studies directly use the same type of predictor without focusing on the diversity of the prediction models. In theory, the diversity of predictors could improve the robustness of the prediction model so that the model can be stabilized and continuously operated. Hence, this paper suggests introducing a novel model library to bolster the stability of the prediction model, building upon existing frameworks.
Most previous studies selected the current optimal predictors for each subseries, but these predictors do not guarantee that the prediction results are optimal on a global level. Thus, this paper introduces the concept of valid predictors to enhance the overall predictive capability of the model.
To address the gaps identified above, this paper introduces a novel intelligent prediction model. This model is designed to enhance the prediction performance of the CFI by adopting a new approach to adaptive model selection for subseries. The proposed model comprises the following four key modules: adaptive data preprocessing, model library, adaptive model selection, and multi-objective ensemble. First, an adaptive data preprocessing module is introduced based on Sequential VMD (SVMD) technology to mitigate the impact of fluctuations and noise in the original data series. Next, a new model library module is constructed, which consists of four prediction models, namely Outlier Robust Extreme Learning Machine (ORELM), BP, the Group Method of Data Handling (GMDH), and the Adaptive-network-based Fuzzy Inference System (ANFIS), to offer a diversity of predictors for one subseries. These four predictors have good nonlinear fitting abilities and can ensure the prediction accuracy of the proposed model. This study introduces an adaptive model selection module that can efficiently select the appropriate predictor to improve integral prediction effectiveness. Finally, the multi-objective artificial vulture optimization algorithm is introduced into the multi-objective ensemble module, which enhances the accuracy and stability of the prediction model.
In this study, four experiments and one discussion are designed and conducted on four different datasets (Southeast Asia Freight Index, Dalian Containerized Freight Index, SCFI, and CCFI) to verify the superiority of the proposed model. At the same time, eleven benchmark models are built to confirm the superiority of this prediction model. The experimental results indicate that this model significantly outperforms the comparison models across all datasets, with MAPE values of 0.545154, 0.572040, 0.536753, and 0.114937%. This demonstrates the effectiveness of the model selection module in adapting to different datasets. Therefore, the proposed prediction model can serve as a powerful tool for the forecasting of container freight indices. The contributions and innovations of the proposed novel intelligent prediction model are summarized as follows:
Developing a creative prediction model for CFI. Although many studies have applied the decomposition ensemble framework to predict the CFI, the processing of decomposed subseries is inadequate and needs further improvement. Therefore, to fill existing gaps, this study innovatively proposes a prediction model for CFI that incorporates the following four modules: adaptive data preprocessing, model library, adaptive model selection, and multi-objective ensemble. In addition, note that to achieve effective integral prediction, the number of predictors is not limited by applying the adaptive selection strategy.
Devising an innovative adaptive data preprocessing module. It is clear from the analysis of previous studies that VMD is widely used in the decomposition ensemble model, but this study innovatively applies SVMD technology to the decomposition ensemble model for the CFI. Compared with VMD technology, SVMD can adaptively decompose the raw data and determine the number of decomposition layers (K), which can effectively ensure the objectivity of the decomposition process.
Establishing a model library to predict the decomposed subseries. Unlike most previous studies, which did not focus on the variety of predictors for subseries prediction, a new model library module is constructed, incorporating the following four models: ORELM, BP, GMDH, and ANFIS. It can not only greatly improve the diversity of predictors but also effectively advance the nonlinear fitting ability of the prediction model.
Introducing the idea of model validity to adaptive model selection. The adaptive model selection module is designed to obtain valid predictors for each subseries in this study. In this module, Least Absolute Shrinkage and Selection Operator (LASSO) technology is introduced to conduct adaptive model selection, selecting the valid predictors for the subseries from the perspective of advancing the effectiveness of integral forecasting. In contrast to scholars who choose the best predictor for one subseries locally, the number of predictors is not restricted, which means that for one subseries, multiple models may be selected, and it is common that no model is chosen. Thus, adaptive model selection chooses predictors that can effectively enhance the prediction performance from the perspective of the prediction model rather than using the greedy idea to select the optimal current predictor.
Designing the multi-objective ensemble module based on the Multi-objective Artificial Vulture Optimization Algorithm (MOAVOA). The multi-objective ensemble is proposed to integrate the prediction values of subseries, which can concurrently fulfill the requirements of both prediction accuracy and stability. By introducing archive grid and leader selection mechanisms, the search ability of MOAVOA is greatly boosted, which can effectively advance the prediction accuracy and robustness of the prediction model.
The remainder of this paper is organized as follows:
Section 2 describes the preliminary methods.
Section 3 introduces the proposed model.
Section 4 presents the empirical results and verifies the contribution of this study.
Section 5 discusses the results, and
Section 6 summarizes the whole paper.
3. Proposed Method
This study establishes a new intelligent prediction model-based adaptive selection strategy for CFI, incorporating the following four steps: (1) adaptive data preprocessing, (2) construction of the model library and subseries prediction, (3) adaptive model selection, and (4) multi-objective ensemble (
Figure 1).
Step 1: Adaptive data preprocessing.
SVMD has many achievements in many fields and is not affected by the number of decomposed modes. Therefore, in this study, SVMD was selected to decompose the CFI. Multiple modes can be obtained from the raw data through adaptive data preprocessing module, which can significantly reduce the impact of data noise on the prediction accuracy.
Step 2: Construction of the model library and subseries prediction.
Owing to the strong nonlinearity of the containerized freight index, it is challenging for the single-predictor model to predict CFI trends; thus, the model library module is introduced into the prediction model. The model library contains the following four predictors: ORELM, BP, GMDH, and ANFIS. These predictors have a strong nonlinear fitting ability, and each predictor has unique characteristics, which gives the proposed model a high prediction accuracy.
Step 3: Adaptive model selection.
Because different models have different prediction abilities for different subseries, adaptive model selection module is essential. The LASSO feature selection algorithm achieves feature dimensionality reduction by introducing an L1 norm penalty term. In this study, the LASSO feature selection algorithm is used in adaptive model selection, which can effectively extract key impact factors and select the valid predictors for each subseries.
Step 4: Multi-objective ensemble.
The multi-objective ensemble is an important step in the prediction model that combines the prediction values of multiple subseries into a final prediction result. As different subseries prediction values may have different proportions in the final result, weights are added to different subseries prediction values. Because the MOAVOA has a strong ability to combine predictors, it is used to estimate the predictor weights in this study.
4. Empirical Results and Analysis
This section presents the empirical results of the proposed model.
Section 4.1 introduces the data selection and performance metrics for the prediction results.
Section 4.2 describes the experimental design used to demonstrate the experimental objectives. In
Section 4.3, the results of the adaptive model selection module are discussed and analyzed. In
Section 4.4, an experiment is designed to compare a single artificial neural network with the proposed model. In
Section 4.5, an experiment is designed to compare the equal-weight method with the proposed model. In
Section 4.6, an experiment is designed to compare different intelligent ensemble methods with the proposed model. These experiments were performed using MATLAB R2022b on a Windows 10 Intel Core i5-10210U CPU.
4.1. Data Description and Performance Metrics
The Southeast Asia Freight Index (SEAFI), Dalian Containerized Freight Index (DCFI), SCFI, and CCFI were selected to verify the model proposed in this study. These container freight indices have important applications in measuring the level of container transportation markets worldwide. The Southeast Asia CFI provides key insights into shipping trends in Southeast Asia and aids businesses and policymakers in making informed logistics and trade decisions. The Dalian CFI reflects shipping activities in Northeast China, helping businesses manage costs and optimize supply chains in this strategic area. The Shanghai CFI monitors shipping status throughout Shanghai and serves as a crucial indicator of global shipping trends and economic health. The China CFI offers a comprehensive overview of China’s container shipping market, covering multiple ports and helping businesses engaged in international trade with China forecast shipping expenses and adjust logistics strategies. The data are derived from
https://x.qianzhan.com/xdata/ (accessed on 8 March 2023). The SEAFI dataset has 288 samples, the DCFI dataset has 288 samples, the SCFI dataset has 480 samples, and the CCFI dataset has 960 samples. The datasets were partitioned into three segments, namely training, validation, and testing sets, with an allocation ratio of 12:3:1. The data features and segmentation of all datasets are shown in
Figure 2. Training samples from different datasets were used to train the prediction model, and the empirical results were compared to confirm the advantages of the proposed model. Specifically, SEAFI had 216 training samples, DCFI had 216 training samples, SCFI, had 360, and CCFI had 720 training samples.
Table 1 presents the descriptive statistics for the four datasets.
Table 1 shows that the descriptive statistical values of distinct datasets exhibit variations; therefore, different datasets have different characteristics, which verifies that the proposed model has strong robustness.
MAE, RMSE, MAPE, IA, and TIC were chosen to assess model performance because they each offer unique insights and are widely used in predictions [
41]. MAE measures the average error magnitude in the same units as the data, making it easy to understand. RMSE emphasizes larger errors by squaring them, which is useful when large errors require more attention. MAPE provides the error as a percentage, making it easy to compare different datasets. IA measures both the size and direction of errors, with values ranging from 0 to 1, indicating agreement between the observed and predicted values. TIC compares forecasting performances across different scales and balances different error types, which is particularly helpful in economic forecasting. In addition, the stability of the prediction model can be determined by assessing the standard deviation of the prediction results [
52]. A smaller standard deviation suggests that the model exhibits greater robustness and that the prediction results are more reliable. These metrics provide a comprehensive evaluation of model accuracy and error characteristics. The mathematical equations for these metrics are listed in
Table 2.
4.2. Experimental Design
Eleven comparative models were constructed to demonstrate the predictive performance of the proposed model. Four comparative experiments were conducted to verify the superiority of the proposed model. Experiment I aimed to introduce the selection results of the adaptive selection module on different datasets, proving the necessity of the adaptive data preprocessing module in the proposed prediction model. In Experiment II, four sub-predictors (ORELM, BP, GMDH, and ANFIS) were used as comparison models to prove the necessity and effectiveness of the adaptive data preprocessing and multi-objective ensemble modules, proving that the proposed model exhibits superior predictive capabilities compared to a single neural network model. To further demonstrate the superiority of the multi-objective ensemble, Experiment III was designed, and the proposed model was compared with an equal-weight ensemble model, which can prove that the proposed model exhibits superior predictive capabilities compared to an equal-weight ensemble model. In Experiment IV, we compared the proposed model with other multi-objective ensemble methods, proving that the proposed model exhibits superior predictive capabilities compared to other multi-objective ensemble models.
4.3. Experiment I: Results of the Adaptive Model Selection Module
A comprehensive assessment of the results of the adaptive model selection process across different datasets was conducted. The model library, comprising ORELM, BP, GMDH, and ANFIS, showed a robust nonlinear fitting capability. By employing consistent hyperparameters across all the models, the influence of model-specific parameters was minimized, ensuring a fair comparison. The necessity of selecting models individually for each dataset stems from their distinct predictive capabilities. This is further exemplified by the varying contributions of the different dataset modes to the final prediction results, necessitating adaptive model selection for optimal performance. The outcomes of the adaptive model selection module for different datasets are shown in
Table 3.
In the SEAFI dataset, ORELM and ANFIS are effective for mode 1, while ORELM, BP, and ANFIS are suited for mode 2. Interestingly, mode 3 requires no model, indicating that its exclusion may enhance forecasting accuracy. For the DCFI dataset, ORELM performed well for mode 1. BP and ANFIS were preferred for mode 2, and ORELM and ANFIS were suitable for mode 3. The results on the SCFI dataset demonstrate ORELM’s efficacy for mode 1, that of BP and ANFIS for mode 2, and that of a combination of ORELM and GMDH for mode 3. Finally, in the CCFI dataset, ORELM, BP, and ANFIS were effective for mode 1, and the same combination worked well for mode 2, with GMDH being the choice for mode 3. Owing to the use of adaptive model selection techniques, the proposed model achieved better predictive results than the comparison models. These results underscore the importance of adaptive model selection tailored to each dataset’s characteristics, demonstrating that different models and their combinations contribute variably to the prediction accuracy, depending on the data modes.
4.4. Experiment II: Comparison of Single Artificial Neural Network Model and Proposed Model
In this experiment, the comparative outcomes of a single artificial neural network and the proposed model were analyzed. The selected single artificial neural network models were ORELM, BP, GMDH, and ANFIS. The comparative outcomes between the single artificial neural network and the proposed model are displayed in
Table 4. Visualizations of both the single artificial neural network and the proposed model are presented in
Figure 3.
The experimental results compare the performance of single artificial neural network models (ORELM, BP, GMDH, and ANFIS) and those of the proposed model in predicting containerized freight indices across multiple datasets. The proposed model consistently outperforms the single models, demonstrating significantly lower MAE, RMSE, MAPE, and TIC values. For instance, for the SEAFI dataset, the proposed model achieved an MAE of 27.635876, an RMSE of 32.146729, and an MAPE of 0.545154%. The BP model, which performed the worst, recorded an MAE of 151.781720, an RMSE of 178.337512, and an MAPE of 3.015159%. This stark contrast highlights the superior predictive accuracy and reduced error rates of the proposed model compared with single neural network models. For the other datasets, the proposed model continued to demonstrate superior performance. The DCFI dataset achieved an MAE of 5.122103, an RMSE of 6.506787, and an MAPE of 0.572040%, whereas the ANFIS model performed poorly, with an MAE of 59.555643, an RMSE of 71.965793, and an MAPE of 6.865999%. Similarly, for the SCFI dataset, the proposed model had an MAE of 4.469254, an RMSE of 5.254923, and an MAPE of 0.536753%, significantly outperforming the ANFIS model’s MAE of 34.211761), RMSE of 39.326185, and MAPE of 4.063924%. For the CCFI dataset, the proposed model shows outstanding performance, with an MAE of 0.995431, RMSE of 1.278544, and MAPE of 0.114937%, compared with the BP model’s MAE of 10.592830, RMSE of 13.355790, and MAPE of 1.198346%.
These results indicate that the proposed model not only offers better accuracy but also greater reliability across various datasets, making it a superior choice for predicting containerized freight indices. Owing to the use of adaptive data preprocessing techniques, the proposed model achieved better predictive results than the comparison models. This consistent superiority across multiple performance metrics and datasets underscores the robustness and effectiveness of the proposed model for handling complex predictive tasks in freight index forecasting.
4.5. Experiment III: Comparison of Equal-Weight Method and the Proposed Model
In this section, we present a comparative experiment to assess the equal-weight strategy against the proposed model. The following comparison models were chosen: ADP-ORELM-EW, ADP-BP-EW, ADP-GMDH-EW, and ADP-ANFIS-EW, where ADP is an adaptive data preprocessing operation using SVMD and EW is the equal-weight method. All models used the same hyperparameters to mitigate the impact of the model parameters on the prediction outcomes. The comparative outcomes between the simple ensemble model and the proposed model are presented in
Table 5, and a visual presentation of the comparative outcomes is presented in
Figure 4.
The proposed model demonstrates superior prediction performance on the SEAFI, DCFI, SCFI, and CCFI datasets, with MAE values of 27.635876, 5.122103, 4.469254, and 0.995431, respectively. In comparison, ADP-BP-EW records the highest MAE of 113.873704 for SEAFI, ADP-ANFIS-EW has the maximum MAE loss of 10.902471 for DCFI, and ADP-GMDH-EW exhibits the poorest MAE values of 6.653787 and 1.877313 for SCFI and CCFI, respectively. In terms of RMSE, the proposed model achieves 32.146729, 6.506787, 5.254923, and 1.278544 for the SEAFI, DCFI, SCFI, and CCFI datasets, respectively. In contrast, ADP-GMDH-EW shows the poorest RMSE values, reaching 245.850597 for SEAFI, 7.532368 for SCFI, and 2.411816 for CCFI. Additionally, ADP-ANFIS-EW experienced a maximum RMSE loss of 13.981185 for DCFI. For MAPE, the proposed model attained values of 0.545154%, 0.572040%, 0.536753%, and 0.114937% for SEAFI, DCFI, SCFI, and CCFI, respectively. Comparatively, ADP-BP-EW achieved the worst MAPE of 2.219730% on SEAFI, ADP-ANFIS-EW recorded the maximum MAPE loss of 1.224123% on DCFI, and ADP-GMDH-EW showed the poorest MAPE values of 0.784850% and 0.215563% on SCFI and CCFI, respectively.
The experimental results highlight the benefits and advantages of the proposed model, particularly in its ability to significantly outperform alternative models across multiple performance indicators such as MAE, RMSE, and MAPE. The strength of the proposed model lies in its adaptive model selection module and multi-objective ensemble module, which collectively enhance its predictive accuracy and robustness. These components enable the model to select the most suitable models dynamically and optimally combine them, leading to more precise and reliable predictions. This adaptability is crucial for handling the varying characteristics and complexities of different datasets. Owing to the use of multi-objective ensemble techniques, the proposed model achieved better predictive results than the comparison models. Moreover, compared with traditional equal-weight ensemble models, the proposed model shows a significant improvement. This further demonstrates the superiority of multi-objective ensemble methods, providing a more nuanced and effective approach for predictive modeling. Its superior performance across diverse datasets underscores the versatility and potential of the model for broad applicability to various predictive tasks.
4.6. Experiment IV: Comparison of Different Intelligent Ensemble Method and the Proposed Model
The multi-objective ensemble module plays a vital role in consolidating the outputs of multiple predictors to produce the final prediction value. In this section, a comparison experiment designed to explore the influence of various multi-objective ensemble schemes on the final prediction value is presented. ADP-LASSO-MOGWO, ADP-LASSO-MODA, and ADP-LASSO-MOPSO were selected for comparison with the proposed model, and the adaptive model selection module was established based on LASSO technology. All models used the same hyperparameters to mitigate the impact of the model parameters on the prediction outcomes.
Because each intelligent optimization scheme has different ensemble capabilities for different datasets, it is necessary to analyze the different ensemble schemes separately for each dataset. The performance of each intelligent optimization scheme is shown in
Table 6. A visualization of each intelligent optimization scheme is shown in
Figure 5.
In the SEAFI dataset, the proposed model excels, with an MAE of 27.635876, an RMSE of 32.146729, and an MAPE of 0.545154%. In contrast, ADP-LASSO-MOGWO performed poorly, with MAE, RMSE, and MAPE values of 29.484178, 34.217756, and 0.583031%, respectively. ADP-LASSO-MODA showed the worst RMSE of 34.409024. For the DCFI dataset, the proposed model demonstrated strong performance, achieving an MAE of 5.122103, RMSE of 6.506787, and MAPE of 0.572040%. Conversely, ADP-LASSO-MODA underperformed, with corresponding values of 5.525988, 7.271350, and 0.612448%, while ADP-LASSO-MOGWO had the highest MAE and MAPE values of 5.568783 and 0.632117%, respectively. On the SCFI dataset, the proposed model again shows superior results, with an MAE of 4.469254, RMSE of 5.254923, and MAPE of 0.536753%. Meanwhile, ADP-LASSO-MODA lags behind, recording corresponding values of 4.868545, 5.737525, and 0.588101%. Finally, on the CCFI dataset, the proposed model achieved outstanding performance, with an MAE of 0.995431, RMSE of 1.278544, and MAPE of 0.114937%. Conversely, ADP-LASSO-MODA yielded less favorable results, with corresponding values of 1.142011, 1.440428, and 0.131289%.
The experimental results indicate that the proposed model significantly outperforms its counterparts in predictive accuracy across multiple datasets. This performance is underscored by lower MAE, RMSE, MAPE, IA, and TIC values, suggesting that the model delivers more precise and reliable forecasts. The experimental results prove that the proposed model exhibits superior predictive capabilities compared to other multi-objective ensemble models. Owing to the use of multi-objective ensemble techniques based on the MOAVOA, the proposed model achieved better predictive results than the comparison models. The benefits of the proposed model include enhanced predictive accuracy, which can lead to better decision making and resource allocation. The robustness and consistency of the model across various datasets also highlight its adaptability and potential applicability to a wide range of predictive tasks. In addition, the model effectively enhanced the overall predictive performance, indicating its ability to efficiently handle complex data patterns.
5. Further Discussion
This section discusses the necessity of the adaptive model selection and multi-objective ensemble modules in the proposed model. To further verify the contribution of each module to the overall model, Equation (
35) is used to analyze the improvement percentage of point predictions [
41]. Here,
and
are the statistical indicator values for the comparison and proposed models, respectively. The larger the
value, the better the performance of the proposed prediction model compared to that of the comparison model.
Table 7 lists the improvement percentages of the proposed model across the four datasets.
Comparison of the proposed model wit EW models ADP-ORELM-EW, ADP-BP-EW, ADP-GMDH-EW, and ADP-ANFIS-EW shows that the values are all positive, which indicates that the proposed model outperforms the other models with EW. Specifically, the values of the proposed model and ADP-GMDH-EW models for RMSE and TIC are as high as 0.869243 and 0.868656, respectively, on the SEAFI dataset. Therefore, the adaptive model selection module in the proposed model is more significant for the CFI. According to a comparison of the proposed model with models employing other intelligent optimization algorithms such as ADP-LASSO-MOGWO, ADP-LASSO-MODA, and ADP-LASSO-MOPSO, the values are all positive, which demonstrates that the performance of the proposed model surpasses that of the models using other intelligent optimization algorithms. The difference in the value between the proposed model and models using other intelligent optimization algorithms is relatively small for most indicators and datasets. However, the values of the proposed model and ADP-LASSO-MODA models on RMSE and TIC indicators are as high as 0.105147 and 0.105402, respectively, on the DCFI. Hence, the multi-objective ensemble module in the proposed model is more effective than other intelligent optimization algorithms for the CFI.
Comparative analysis shows that the proposed model consistently outperforms other models, thanks to its effective adaptive model selection and multi-objective ensemble modules. These components significantly improve the forecasting performance of the model, making it superior to models with equal-weight ensembles and those using other intelligent optimization algorithms.