Next Article in Journal
Spatial Modeling of Auto Insurance Loss Metrics to Uncover Impact of COVID-19 Pandemic
Previous Article in Journal
Aggregation and Pruning for Continuous Incremental Multi-Task Inference
Previous Article in Special Issue
Heterogeneous Spillover Networks and Spatial–Temporal Dynamics of Systemic Risk Transmission: Evidence from G20 Financial Risk Stress Index
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale TsMixer: A Novel Time-Series Architecture for Predicting A-Share Stock Index Futures

1
School of Computer Science and Engineering, Faculty of Innovation Engineering, Macau University of Science and Technology, Macau 999078, China
2
Macau Institute of Systems Engineering, Faculty of Innovation Engineering, Macau University of Science and Technology, Macau 999078, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(9), 1415; https://doi.org/10.3390/math13091415
Submission received: 22 March 2025 / Revised: 19 April 2025 / Accepted: 21 April 2025 / Published: 25 April 2025
(This article belongs to the Special Issue Machine Learning Methods and Mathematical Modeling with Applications)

Abstract

:
With the advancement of deep learning, its application in financial market forecasting has become a research hotspot. This paper proposes an innovative Multi-Scale TsMixer model for predicting stock index futures in the A-share market, covering SSE50, CSI300, and CSI500. By integrating Multi-Scale time-series features across the short, medium, and long term, the model effectively captures market fluctuations and trends. Moreover, since stock index futures reflect the collective movement of their constituent stocks, we introduce a novel approach: predicting individual constituent stocks and merging their forecasts using three fusion strategies (average fusion, weighted fusion, and weighted decay fusion). Experimental results demonstrate that the weighted decay fusion method significantly improves the prediction accuracy and stability, validating the effectiveness of Multi-Scale TsMixer.

1. Introduction

Futures contracts are agreements to buy or sell a commodity at a future date at a price that is agreed upon today. They are important hedging tools for investors [1]. Stock index futures, derived from index values, are highly effective in minimizing investors’ exposure to market volatility and addressing systemic risk. This is especially important in the Chinese market, where the difficulty of short-selling individual stocks heightens the significance of the short-selling mechanism in the futures market as a primary risk management strategy [2]. Acknowledging the significance of fostering financial market growth and employing effective risk management tools, China officially launched its first financial futures product, the Shanghai Shenzhen 300 (CSI300) index futures, in 2010 [3]. Subsequently, the Shanghai Stock Exchange 50 (SSE50) index futures and the China Securities 500 (CSI500) index futures were also launched. The launch of these stock index futures has facilitated an increase in trading volume and market liquidity, while emphasizing the importance of anticipating changes in the stock index futures market [4,5]. However, existing research indicates that the market does not fully align with the efficient market hypothesis [6,7]. As a result, predicting stock index futures prices is not only theoretically possible but also practical and valuable in real-world applications [8].
Long-term stock index futures prediction is a popular research area. Previous mainstream methods primarily rely on traditional econometric models [9] such as Logistic Regression (LOGIT) [10,11], autoregressive integrated moving average (ARIMA) [12,13], and generalized autoregressive conditional heteroscedasticity (GARCH) [14,15]. In addition, Zhe Lin [16] proposed an empirical analysis of the SSE Composite Index using GARCH-type models to examine its volatility, finding that the exponential GARCH (EGARCH) model outperforms others and offering suggestions for improving China’s securities market stability. However, traditional econometric models struggle to effectively address the nonlinear relationships in financial data, resulting in lower prediction accuracy [17].
To address the limitations of traditional models, machine learning and deep learning have been increasingly applied in financial time-series forecasting, due to its ability to accurately capture complex nonlinear patterns in the data [18,19,20,21]. For example, Huang et al. investigated the predictability of financial movement direction. They used support vector machine (SVM) to forecast the weekly movement of the NIKKEI 225 index. The study demonstrated that SVM outperforms other methods [22]. Illa et al. proposed a stock price prediction methodology using random forest and SVM algorithms to forecast market trends, demonstrating that both models perform well, with SVM yielding the highest accuracy for time-series data [23]. Furthermore, Yang et al. focused on predicting stock market trends by analyzing emotions through news articles using natural language processing techniques and machine learning algorithms, finding that gradient boosting decision tree (GBDT) outperforms Adaboost, XGboost, decision tree, and logistic regression in accuracy [24].
Predicting stock index futures is essentially a time-series forecasting problem, and deep learning-based time-series models have become one of the hottest research topics in recent years [25]. For instance, Waswani et al. proposed the Transformer model, which replaced traditional recurrent networks with a self-attention mechanism, thus significantly improving parallelization and performance in sequence modeling tasks such as machine translation [26]. Building upon this, Zhou et al. developed Informer, a Transformer-based model tailored for long sequence time-series forecasting. Informer leveraged ProbSparse self-attention and a generative decoder to efficiently manage long-range dependencies, further enhancing prediction accuracy [27]. While these Transformer-based models showed strong performance, Zeng et al. questioned their effectiveness for long-term forecasting and presented DLinear, a simple linear model that outperformed complex Transformer models by more effectively capturing temporal relations with a single-layer architecture, suggesting that simpler approaches may be more suitable in some cases [28]. In contrast, Nie et al. introduced PatchTST, which enhanced Transformer models for multivariate time-series forecasting by employing a patching strategy to segment time series, thereby reducing computation and memory usage while improving the model’s ability to capture long-term dependencies [29]. To address the complexity of temporal variations in time-series data, Wang et al. proposed TimeMixer, a multiscale mixing architecture that decomposes time series into seasonal and trend components, enabling more accurate predictions for both short-term and long-term forecasting [30]. Extending the idea of temporal variation modeling, Wu et al. developed TimesNet, which transformed time series into 2D tensors to capture both intra- and inter-period variations, achieving state-of-the-art performance in a range of time-series analysis tasks [31]. Moreover, Chen et al. presented TSMixer, an all-MLP architecture that efficiently captured both temporal and cross-variate dependencies using a novel mixing operation along both the time and feature dimensions, further contributing to advancements in time-series forecasting [32]. Due to its advantages, we will adopt this method.
Time-series forecasting, particularly in the financial domain, often faces the challenge of handling complex market volatility and the need for information across multiple time scales [33]. Trading decisions in stock index futures are influenced not only by short-term fluctuations but also by long-term trends, such as weekly or monthly movements. However, most existing forecasting models primarily focus on short-term trends, neglecting cross-time scale information, which limits their effectiveness in addressing the complex dynamics of financial markets.
Moreover, many existing models rely on the historical data of stock index futures for overall modeling and forecasting [34]. This approach overlooks the fact that stock indices are composed of multiple weighted component stocks, resulting in models that capture only the overall trend of the index without accurately measuring the contribution of individual component stocks to index volatility. Due to the aggregation nature of stock indices, different component stocks may exhibit distinctly varying price changes and volatility characteristics at different times, which undermines the model’s precision and accuracy.
To address these issues, this paper proposes three improved methods. First, based on the TSMixer model, this paper introduces a Multi-Scale module to better capture feature variations across different time granularities, such as 20-day and 60-day periods. This multi-path fusion strategy enables the model to retain daily details while effectively integrating weekly and monthly information, enhancing the accuracy and robustness of stock index futures forecasting. Second, considering that stock index futures are composed of multiple component stocks, this paper proposes using the historical data of all component stocks for training, and then inferring future trends of the stock index futures by weighting the forecast results of the individual component stocks during the prediction phase. This approach not only captures the unique characteristics of each component stock but also estimates their contributions to the overall index volatility, significantly enhancing the interpretability of the model. Third, to further explore the differentiated contributions of each component stock to the overall index in the prediction phase, three different weighted fusion methods are designed. These methods are average-based fusion, weighted-based fusion, and weighted-based decay fusion. In our experiments, the weighted-based decay fusion method considers both the weighted information of component stocks and the future expectation of the index at the same time and achieves the optimal predicted results.
The main contributions of this paper are as follows: (1) A deep learning-based Multi-Scale TsMixer model is proposed, which integrates a Multi-Scale time-series data processing strategy. This model performs feature fusion across multiple time scales (short term, medium term, and long term) to more comprehensively capture the short-term fluctuations and long-term trends in stock index futures history. (2) A stock index futures trend forecasting method combining independent component stock predictions with weighted fusion is proposed. (3) Three different weighted fusion strategies (average-based fusion, weighted-based fusion, and weighted-based decay fusion) are designed and systematically evaluated for their forecasting performance. Experimental results demonstrate that the weighted-based decay fusion method significantly outperforms the other methods in terms of prediction accuracy and stability.

2. Methodology

In this work, we propose a method for predicting stock index futures price movements, comprising the three main components: (1) building baseline models and proposing a more advanced network architecture, (2) adopting an individual training strategy for each constituent stock, and (3) applying a weighted fusion of the prediction outputs. The integration of these three components yields superior predictive performance. This study addresses two distinct tasks: a regression task predicting the percentage change in stock index futures prices and a binary classification task determining whether prices will increase or decrease.

2.1. Baselines and Proposed Model

We employ various advanced time-series models with different network architectures as baselines, including Transformer [26], Informer [27], DLinear [28], PatchTST [29], TimeMixer [30], TimesNet [31], and TSMixer [32], to address both regression and classification tasks in the experiments. To explore the effect of varying sequence lengths on model prediction accuracy, we configure the input data sequences for each model to lengths of 20, 60, and 80 during the baseline setup phase, thereby capturing different time scales. By comparing these sequence lengths, we then evaluate the models’ predictive performance under varying amounts of historical information.
Among these baseline models, TsMixer is implemented based on a pure multilayer perceptron (MLP) architecture. Pure MLP architectures are frequently utilized as baseline models for various complex networks [35]. The primary advantages of pure MLPs include their simplicity, ease of implementation, parallelizability, and versatility. Additionally, when modeling high-dimensional or structured inputs, pure MLP architectures do not depend on specific structural priors, such as convolutional layers or self-attention mechanisms, yet they consistently deliver a stable performance. Consequently, we select TsMixer, which is founded on a pure MLP architecture, to innovate and enhance the network architecture in this work.
Typically, trading decisions for stock index futures require reference to longer timescales, such as weekly or monthly trends. To better capture feature variations across different temporal granularities (e.g., 5-day and 20-day), we incorporate a Multi-Scale module into TsMixer to enhance predictive accuracy. As illustrated in Figure 1, we refer to this enhanced architecture as the Multi-Scale TsMixer. Specifically, we apply average pooling to the daily data to obtain weekly feature sequences at a resolution of 5 and monthly feature sequences at a resolution of 20, ensuring the exclusion of weekends when forming weekly or monthly intervals.
Next, each weekly and monthly feature sequence passes through a small MLP for independent feature extraction, after which we concatenate these features with the original daily Tsmixer outputs. Consequently, we extend Tsmixer to embed Multi-Scale (daily, weekly, and monthly) information within the temporal dimension, thereby increasing the learnable input features. Finally, we project and fuse these Multi-Scale features to generate the final prediction. This multi-path fusion strategy not only preserves finer daily-level details but also integrates weekly and monthly information, aligning with the need to consider multiple time horizons in financial markets. As a result, the model achieves higher accuracy and robustness in capturing both short-term fluctuations and medium-to long-term trends in stock index futures.

2.2. Individual Training Strategy

In Section 2.1, we describe the network architecture of our proposed Multi-Scale TsMixer. In this section, we introduce a novel training strategy for the proposed model. In the field of stock index futures prediction, many previous studies directly utilize the historical data of stock index futures for overall modeling and prediction [19,36]. These approaches often overlook the fact that an index is composed of multiple constituent stocks with varying weights. As a result, the models are limited to capturing only the overall trend and are unable to accurately assess the differential contributions of individual constituent stocks to index fluctuations.
Due to the aggregative nature of indices, constituent stocks can exhibit significantly varying magnitudes of price changes and volatility characteristics across different periods. While stock index futures generally capture the overall trend, those approaches tend to obscure or attenuate the unique characteristics of individual stocks, thereby diminishing the precision and accuracy of the predictive models.
To address the aforementioned limitations, we utilize the historical data of all constituent stocks for training and predict stock index futures trends by weighting the predicted outcomes of individual constituent stocks during the prediction phase (detailed in Section 2.3). For example, when the model is trained exclusively on the index futures’ historical data, it primarily learns a composite index curve. This approach hinders the accurate capture of each constituent stock’s unique characteristics and weight variations. Conversely, by conducting parallel training on the constituent stocks, the model can learn each stock’s distinct features and aggregate their predicted results using their respective weights to predict the overall index movement. This individual training strategy enables the explicit estimation of each stock’s contribution to the index, thereby enhancing the interpretability and precision of the prediction results.
On the left side of Figure 2, we present the individual training strategy for constituent stocks. Taking the SSE50 Index as an example in Figure 2, we decompose the explicit features of stock index futures into 50 latent features, each representing an individual constituent stock. These latent features are then utilized as inputs to the time-series model, thereby capturing the unique characteristics of each constituent stock.

2.3. Fusion Method for Prediction

In the model prediction phase, to further investigate the differential contributions of individual constituent stocks to the overall index (as shown in the right side of Figure 2), we propose three different fusion methods. By weighting and integrating the prediction outcomes of each constituent stock, these methods effectively enhance the accuracy of predicting the rise and fall of stock index futures. The three fusion methods include average fusion, weighted fusion, and weighted decay fusion, are shown in Algorithm 1.
Algorithm 1 Predicting stock index futures using Multi-Scale TsMixer and fusion methods.
1:
Input: Historical data of constituent stocks S = { S 1 , S 2 , , S n } , trained model M , fusion method f fusion , optional weights ω , optional decay factors α , current time step t.
2:
Output: Predicted index futures value Y ^ S , t + 1 .
3:
Initialization: Y ^ S , t + 1 0
Step 1: Generate Individual Stock Predictions
4:
for each stock S i S  do
5:
     y ^ S i , t + 1 M ( S i , t )            ▹ Predict next-day value
6:
end for
Step 2: Apply Fusion Method
7:
if f fusion = “average” then
8:
     Y ^ S , t + 1 1 n i = 1 n y ^ S i , t + 1            ▹ Simple average fusion
9:
else if f fusion = “weighted” then
10:
     Y ^ S , t + 1 i = 1 n ω i y ^ S i , t + 1              ▹ Weighted fusion
11:
else if f fusion = “weighted decay” then
12:
    for each stock S i S  do
13:
         y S i , decay k = 1 n α k y ^ S i , t + k            ▹ Apply decay factor
14:
    end for
15:
     Y ^ S , t + 1 i = 1 n ω i y S i , decay            ▹ Decay-weighted fusion
16:
end if
17:
return Y ^ S , t + 1

2.3.1. Average-Based Fusion Method

Building on the independent predictions of each constituent stock, we first adopt the common average-based fusion method. Specifically, we compute the simple mean of the predicted expected values for all constituent stocks. For multiple constituent stocks S = { S 1 , S 2 , , S n } , the model’s prediction for each stock on the next trading day can be expressed as:
y ^ S i , t + 1 = M ( S i , t )
where y ^ S i , t + 1 denotes the predicted outcome for stock S i at time t + 1 , and M ( S i , t ) represents the trained model that produces the next-day forecast based on information up to time t. We then combine these individual predictions by taking the simple mean across all stocks in to derive the average-based forecast for the index. The futures prediction obtained by the average-based fusion method can be formulated as follows:
Y ^ S , t + 1 = 1 n i = 1 n y ^ S i , t + 1
where n represents the number of constituent stocks.
However, because the average-based fusion method disregards differences in the constituent stocks’ weights within the index, it fails to capture the more pronounced influence of heavily weighted constituents on the overall index. As a result, fluctuations in these key constituents may be diluted or overlooked.

2.3.2. Weighted-Based Fusion Method

To address the average-based fusion method’s inability to account for the distinct influence of each constituent stock, we propose a weighted-based fusion method. This approach assigns weights to each stock’s predicted outcome based on factors such as market capitalization and trading volume, and then aggregates these weighted predictions to produce the overall stock index futures forecast. By highlighting stocks with higher contributions to the index, it more accurately captures the structural relationships within the market. The weighted fusion formula is given by:
Y ^ S , t + 1 = i = 1 n ω i y ^ S i , t + 1
where ω i denotes the weighting coefficient of stock S i and i = 1 n ω i = 1 .
However, there is a certain limitation in the weighted-based fusion method. Because an expectation gap exists between the spot market and the futures market, this method establishes only a relatively static weight mapping between each constituent stock and the index, and does not account for the potential dynamic linkage that may arise between stock index futures and the spot market over time.

2.3.3. Weighted-Based Decay Fusion Method

To better reflect the potential expectation gap and asynchrony between futures and spot markets, we further propose a weighted-based decay fusion method. In this approach, we incorporate the future price fluctuations of each constituent stock, and the corresponding weighting formula is given as follows:
α k = 0 . 5 k , if k = 1 , 2 , , ( n 2 ) 0 . 5 n 1 , if k = ( n 1 ) , n
where n represents the number of total future trading day and k = 1 n α k = 1 .
By using the weighted-based decay fusion method, the model more effectively captures short-term continuity or leading signals during training, highlighting the dynamic interactions between constituent stocks and futures in the merged outcomes. This approach not only retains the benefits of the weighted-based fusion method, but also incorporates a decaying mechanism to integrate short-to medium-term future information, thereby providing a more nuanced understanding of each constituent stock’s influence on the overall index. Besides reflecting the dynamic relationships between constituent stocks and futures, it also enhances the synchronization and accuracy of predictions.

3. Experiments

3.1. Dataset

In this experiment, the dataset is sourced from the SSE50, CSI300, and CSI500 stock index futures and their corresponding constituent stock data. The historical composition of constituent stocks for each stock index future is obtained through the Tonghuashun platform. As is well known, the composition of constituent stocks in an index is updated over time.
The dataset covers daily trading data from 1 January 2016 to 31 October 2024. For each stock index future, there are 16 input features, including opening price, closing price, highest price, lowest price, trading volume, and others. Additionally, for each constituent stock within the stock index futures, the data include not only the same features as the index futures but also additional characteristics such as industry categories and market capitalization, which are used as model inputs. A detailed description of the input features and calculation methods is presented in Table A1 of Appendix A.
We divide the dataset chronologically into three parts: the training set (from 1 January 2016 to 31 December 2020), the validation set (from 1 January 2020 to 31 December 2021), and the test set (from 1 January 2022 to 31 October 2024). Since CSI 500 futures were introduced on 16 April 2015, and our experiment incorporates a maximum output sequence length of 80, we reserve a portion of the historical data for model inputs and begin the training set on 1 January 2016. Additionally, to more accurately evaluate the performance of different methods and models, we utilize as much data as possible for validation and testing. Therefore, we use two years of data for the validation set and about two years of data for the test set.

3.2. Experimental Setting

To ensure the fairness of our experiments, all models are compared under the same dataset and evaluation metrics, using a consistent random seed. The batch size is set to 128 in the training part. The initial learning rate is fixed at 0.0005, and parameter updates are performed via the Adam optimizer. The mean squared error (MSE) is adopted as the loss function, defined by the following formula:
MSE = 1 n i = 1 n ( y i y ^ i ) 2
where y i represents the true value and y ^ i denotes the predicted value for the ith observation. In order to flexibly adjust the learning rate, we adopt a dynamic decay strategy under which the learning rate is reduced to one-fifth of its former value if the validation loss remains unimproved for 15 consecutive epochs. We also employ early stopping to prevent overfitting, halting training whenever the validation performance does not improve for 25 consecutive epochs. These hyperparameter settings and techniques are uniformly applied to all baseline methods as well as our proposed Multi-Scale TsMixer model, thereby ensuring experimental consistency. Finally, we evaluate performance for each trained model on the SSE50, CSI300, and CSI500 stock index datasets.

3.3. Evaluation Criteria

In our regression task, we aim to predict the percentage increase or decrease in stock index futures prices. To evaluate model performance under this framework, we employ several metrics commonly used in regression, including mean absolute error (MAE), mean absolute percentage error (MAPE), and the coefficient of determination ( R 2 ). Their definitions are given by:
MAE = 1 n i = 1 n | y i y ^ i |
MAPE = 1 n i = 1 n | y i y ^ i y i |
R 2 = 1 i = 1 n y i y ^ i 2 i = 1 n y i y ¯ 2
where y i represents the true value, y ¯ represents the mean value, and y ^ i denotes the predicted value for the ith observation. Moreover, in practical stock index futures trading, we typically initiate long or short positions based on the model’s predicted trends. Because these regression metrics do not directly reflect profit or loss, we further treat the outputs as a binary classification. To achieve a more intuitive assessment of trading performance, we adopt precision, specificity, accuracy, and the Sharpe ratio:
Precision = TP TP + FP Specificity = TN TN + FP Accuracy = TP + TN TP + TN + FP + FN
where TP, FP, TN, and FN refer to the numbers of true positives, false positives, true negatives, and false negatives, respectively.

4. Experimental Results and Analysis

In this section, we comprehensively compare and analyze the performance of each model in both regression and binary classification tasks, employing multiple metrics (MAPE, MAE, R 2 , precision, specificity, and accuracy) across different sequence lengths and indices. To streamline the discussion, we provide the regression and classification results for the CSI500 stock index in Table 1 and Table 2, respectively, while supplementary findings for the CSI300 and SSE50 indices appear in Appendix B and Appendix C (Table A2, Table A3, Table A4 and Table A5).
This is a detailed explanation of the tables (see Table 1, Table 2, Table A2, Table A3, Table A4 and Table A5). In all the tables mentioned, upper arrows indicate that larger values are better; lower arrows vice versa. Green, red, and black represent the results that stand out for the 20, 60, and 80-day input sequences, respectively. Bold numbers indicate the best results, and double-underlined numbers indicate the suboptimal results.

4.1. Regression Results

In this regression task, we systematically evaluate the performance of each baseline model (TSMixer, Informer, DLinear, PatchTST, Transformer, TimesNet, TimeMixer) alongside the proposed Multi-Scale TsMixer under input sequences of 20, 60, and 80 days (see Table 1).
Generally, as the input sequence length increases, models capture more historical data, resulting in steady improvements in prediction performance. This trend is particularly evident with longer sequences (60 and 80 days). For instance, from Table 1, with DLinear, the MAE error is 55.147 for a 20-day sequence. However, when the sequence length increases to 60 and 80 days, the error decreases by 0.055 and 0.167, respectively.
It is noteworthy that the proposed Multi-Scale TsMixer demonstrates consistent stability across all input lengths, progressively leveraging the advantages of its Multi-Scale features as the input sequence length increases. For instance, at 20 and 60 days, the performance of Multi-Scale TsMixer and DLinear is relatively close, but at an 80-day input, Multi-Scale TsMixer obviously outperforms the other baseline models in terms of MAPE, MAE, and R 2 . This further validates the efficacy of the Multi-Scale (weekly, monthly) feature fusion strategy in capturing longer-term market trends and subtle fluctuations. These findings not only highlight the importance of utilizing multi-level information in long sequences, but also demonstrates that the proposed method can effectively accommodate varying time scales, offering greater applicability and robustness in predicting stock index futures price changes.
To comprehensively evaluate the performance of our model, we further compare its results across different indices. The experimental results for the CSI300 and SSE50 indices are detailed in Appendix B and Appendix C, respectively. The findings reveal that the proposed Multi-Scale TsMixer model consistently delivers the optimal performance across all these indices, demonstrating the robustness of our proposed method and its capability to yield accurate predictions across diverse market conditions.

4.2. Classification Results

Additionally, we further validate the model’s effectiveness in trading strategies through the binary classification task (up/down) and evaluate its accuracy during simulated trading using metrics such as precision, specificity, and accuracy. As illustrated in Table 2 and Figure 3, the performance trends of most models remain consistent. Interestingly, we find that improvements in regression metrics (MAE, MAPE) are often accompanied by higher classification accuracy. Our proposed model consistently achieves the highest prediction accuracy across various indices. Figure 3 presents a comparison of accuracy variations across different input sequence lengths for multiple models. In the classification task, our proposed model consistently outperforms all baseline models, regardless of the input sequence length.

4.3. Weighted Fusion Results

After validating the model with overall index data, we turn our attention to the internal structure of the index, performing more detailed predictions at the constituent stock level. Specifically, we independently model each constituent stock and apply different weighting fusion strategies during the prediction phase to aggregate multiple stock forecasts into a composite prediction for the overall index. This approach enables the explicit differentiation of the contribution that each constituent stock makes to the index’s movement. By integrating this method with the previously proposed Multi-Scale TsMixer network structure, we further improve the model’s ability to capture both short-term fluctuations and long-term trends in stock index futures.
In this phase, we use the best-performing Multi-Scale TsMixer with an 80-day sequence length from the previous subsection as the baseline. We then compare several fusion methods and their performance. In the experiments, we compare three fusion methods, including average-based fusion, weighted-based fusion, and weighted-based decay fusion methods. In Table 3, the experimental results show that the weighted-based decay fusion method delivers the best results across all evaluation metrics. This indicates the weighted-based decay fusion method’s superiority in capturing the dynamic relationships between constituent stocks and stock index futures.
The experimental results also show that the average-based fusion method, by failing to account for the varying importance of constituent stocks in the index composition, results in moderate accuracy. The weighted-based fusion enhances the alignment with the index structure by applying the weights based on market capitalization or liquidity, which improves predictive accuracy, but it overlooks the potential expectation gap between futures and spot markets, leading to the insufficient capture of dynamic relationships.
In contrast, the weighted-based decay fusion method achieves a more significant breakthrough. By incorporating a short-term decay factor into the training labels or prediction process, it more accurately reflects the expectation gap between futures and spot markets, resulting in superior performance across accuracy and other metrics compared to the other methods.
The further analysis of the performance of these fusion methods across various indices (CSI500, CSI300, and SSE50) shows that the weighted-based decay fusion method outperforms the others in all indices. For instance, in the CSI500 index, the accuracy (0.61848) and specificity (0.64136) of the weighted-based decay fusion method significantly exceed those of the average-based and weighted-based fusion methods. In Table 3, the weighted-based decay fusion method also shows notable improvements in accuracy and specificity on the CSI300 and SSE50 indices. Particularly for the SSE50 index, the weighted-based decay fusion method achieves an accuracy of 0.60018, which is approximately 4% higher than the accuracy of the average-based (0.57394) and weighted-based fusion methods (0.57998).
In conclusion, introducing the weighted decay mechanism significantly enhances the prediction synchronization and accuracy in the dynamic relationship between constituent stocks and stock index futures. The weighted-based decay fusion method demonstrates distinct advantages across different indices and fusion strategies, particularly in capturing the dynamic relationships and short-term fluctuations between constituent stocks and stock index futures, highlighting its broad potential for stock index futures prediction.
Finally, we backtest the performance of the Multi-Scale TsMixer in both independently predicting index futures and merging forecasts from individual components on the test set over the period from 1 January 2022 to 31 October 2024. In the return calculation, we incorporate both the predicted and actual signals. Specifically, let y ^ t denote the predicted signal on day t and y t denote the actual signal on day t (with 1 corresponding to an upward movement and 0 corresponding to a downward movement). We define the indicator function as
I ( t ) = 1 , if y ^ t = y t , 0 , if y ^ t y t .
Then, we introduce the conversion factor δ t , defined as
δ t = 2 I ( t ) 1 .
Thus, when the predicted signal matches the actual signal, δ t = 1 ; otherwise, δ t = 1 . The cumulative return over the entire backtesting period is calculated as
R total = t = 1 T 1 + r t 1 ,
where the net return on day t, r t , is defined as
r t = δ t · price t price t 1 price t 1 f ,
where price t denotes the closing price of the futures on day t, and f is the transaction fee, set at 0.000023 (i.e., 0.23 per 10,000). Figure 4 shows the return comparison between the single-future input and the constituent-stock weighted approach for three index futures with an 80-day feature length.
Table 4 demonstrates that our weighted decay fusion method consistently achieves superior returns across various stock index futures. In addition, the approach achieves a notably lower maximum drawdown and enhanced profitability compared to single-future based model, underscoring the remarkable efficacy and robustness of the proposed fusion strategy. Moreover, continuous iterations and optimizations of the fusion method lead to a steady increase in returns across the three stock index futures.

5. Conclusions and Future Works

This paper presents an innovative stock index futures prediction framework, Multi-Scale TsMixer. Experimental results demonstrate that the Multi-Scale TsMixer model excels in long time-series data, effectively capturing both short-term fluctuations and long-term trends. In addition, we present a deep learning-based approach for stock index futures prediction, integrating the independent modeling of constituent stocks with Multi-Scale time-series fusion. By designing three fusion methods, particularly introducing a weighted decay mechanism, the proposed approach significantly enhances both predictive accuracy and stability.
While the proposed method delivers promising results, there remains room for further optimization. We will provide a more comprehensive assessment of the model’s generalizability on major international indices such as the DJI30, S&P500, and FTSE100 in future work. Future research will focus on enhancing the model’s ability to handle market volatility, improving computational efficiency, and incorporating additional market factors to further strengthen predictive performance.

Author Contributions

Conceptualization, Z.P. and J.Y. (Jianqi Yan); Methodology, Z.P.; Supervision, X.L.; Software, Z.P.; Validation, B.Y.; Writing—original draft preparation, Z.P., J.Y. (Jianqi Yan) and J.Y. (Jin Yan); Writing—review and editing, X.L., Z.P., J.Y. (Jianqi Yan), J.Y. (Jin Yan) and B.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the joint research and Development fund of Wuyi University, Hong Kong and Macao (2019WGALH20), Macao Science and Technology Development Fund (No. 0013/2021/ITP).

Data Availability Statement

Data available upon request from the authors.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. List of feature names, corresponding equations, and descriptions for financial explanations of each equation.
Table A1. List of feature names, corresponding equations, and descriptions for financial explanations of each equation.
Feature NamesEquationsDescription
close_ratio close t close t 1 close t 1 The ratio of the change in the closing price from the previous day to the previous day’s closing price.
open_ratio open t close t 1 close t 1 The ratio of the change in the opening price from the previous day’s closing price.
high_ratio high t close t 1 close t 1 The ratio of the change in the highest price from the previous day’s closing price.
low_ratio low t close t 1 close t 1 The ratio of the change in the lowest price from the previous day’s closing price.
 swing_volatility_5_0 i = 1 5 price i price i 1 5  A measure of volatility over the last 5 periods, capturing swings in price.
volume_ratio volume t volume t 1 volume t 1 The ratio of the change in trading volume from the previous period’s volume.
ma5_ratio MA 5 t MA 5 t 1 close t 1 The change in the 5-period moving average from the previous day’s 5-period moving average, normalized by the previous day’s closing price.
ma20_ratio MA 20 t MA 20 t 1 close t 1 The change in the 20-period moving average from the previous day’s 20-period moving average, normalized by the previous day’s closing price.
ma60_ratio MA 60 t MA 60 t 1 close t 1 The change in the 60-period moving average from the previous day’s 60-period moving average, normalized by the previous day’s closing price.
  rsi_tt_3 100 100 1 + i = 1 3 gain i i = 1 3 loss i   The 3-period relative strength index (RSI), which measures overbought or oversold conditions over a short time frame.
  rsi_tt_6 100 100 1 + i = 1 6 gain i i = 1 6 loss i    The 6-period RSI, similar to the 3-period but calculated over a longer window.
  rsi_tt_12 100 100 1 + i = 1 12 gain i i = 1 12 loss i    The 12-period RSI, commonly used to evaluate market trends over a medium time span.
  rsi_tt_14 100 100 1 + i = 1 14 gain i i = 1 14 loss i    The 14-period RSI, the standard time period for calculating RSI in financial analysis.
day_of_week day t { 1 , 2 , 3 , 4 , 5 , 6 , 7 } The day of the week, represented numerically from Monday = 1 to Sunday = 7.
day_of_month day t { 1 , 2 , 3 , , 31 } The day of the month, ranging from 1 to 31, representing the specific day in a given month.
day_of_year day t { 1 , 2 , 3 , , 365 } The day of the year, ranging from 1 to 365, representing the position of the day within the year.

Appendix B

Table A2. Performance comparison of baseline models and Multi-Scale TsMixer on CSI300 regression task for different input sequences.
Table A2. Performance comparison of baseline models and Multi-Scale TsMixer on CSI300 regression task for different input sequences.
ModelDaysMAPE (↓)MAE (↓) R 2 (↑)
BaselineTSMixer200.852233.0073 0.9822 ̲ ̲
600.854233.07020.9822
800.846932.81420.9822
Informer200.851032.96100.9820
600.847932.84650.9821
80 0.8454 ̲ ̲ 32.7390 ̲ ̲ 0.9821
DLinear200.852232.76350.9823
600.854232.9006 0.9823 ̲ ̲
800.846932.8755 0.9824 ̲ ̲
PatchTST200.933436.05310.9805
601.027539.64570.9774
801.018939.27410.9775
Transformer200.851932.99210.9820
600.849132.89790.9821
800.847532.78970.9823
TimesNet200.8490 32.8955 ̲ ̲ 0.9821
60 0.8465 ̲ ̲ 32.7864 ̲ ̲ 0.9822
800.846432.77280.9823
TimeMixer20 0.8496 ̲ ̲ 32.90930.9820
600.850632.95790.9818
800.845832.75530.9823
Multi-Scale TsMixer (ours)200.850032.8990.9823
600.841132.5770.9824
800.830432.0240.9825
Table A3. Performance comparison of baseline models and Multi-Scale TsMixer on SSE50 regression task for different input sequences.
Table A3. Performance comparison of baseline models and Multi-Scale TsMixer on SSE50 regression task for different input sequences.
ModelDaysMAPE (↓)MAE (↓) R 2 (↑)
BaselineTSMixer200.846322.2496 0.9805 ̲ ̲
600.838422.0446 0.9805 ̲ ̲
800.838522.05750.9806
Informer200.849722.34430.9802
600.844822.22350.9804
800.842322.15870.9804
DLinear20 0.8371 ̲ ̲ 22.0097 ̲ ̲ 0.9806
60 0.8357 ̲ ̲ 21.9767 ̲ ̲ 0.9806
800.836822.00540.9806
PatchTST200.931424.44890.9781
601.037427.19000.9747
800.964725.30070.9768
Transformer200.841922.14720.9804
600.840122.10390.9803
800.836121.99480.9806
TimesNet200.839522.08320.9803
600.833721.9292 0.9805 ̲ ̲
80 0.8355 ̲ ̲ 21.9880 ̲ ̲ 0.9805 ̲ ̲
TimeMixer200.846322.26160.9804
600.843922.1849 0.9805 ̲ ̲
800.838222.0440 0.9805 ̲ ̲
Multi-Scale TsMixer (ours)200.835021.95610.9806
600.838022.03140.9806
800.831021.73110.9806

Appendix C

Table A4. Performance comparison of baseline models and Multi-Scale TsMixer on CSI300 classification task for different input sequences.
Table A4. Performance comparison of baseline models and Multi-Scale TsMixer on CSI300 classification task for different input sequences.
ModelDaysPrecision (↑)Specificity (↑)Accuracy (↑)
BaselinesTSMixer200.48710.55130.5283
600.49120.55500.5324
800.50830.55920.5409
Informer200.48390.56210.5262
600.48820.56580.5298
800.49050.57010.5323
DLinear200.4912 0.5705 ̲ ̲ 0.5343 ̲ ̲
600.50410.5759 0.5403 ̲ ̲
80 0.5128 ̲ ̲ 0.5783 0.5471 ̲ ̲
PatchTST200.44520.52840.4893
600.45910.54060.5061
800.48140.55790.5228
Transformer200.48780.56700.5281
600.4903 0.5789 ̲ ̲ 0.5345
800.49870.57220.5372
TimesNet200.49410.57230.5347
600.49610.57340.5389
800.5082 0.5709 ̲ ̲ 0.5451
TimeMixer20 0.4943 ̲ ̲ 0.56020.5318
60 0.5094 ̲ ̲ 0.57270.5382
800.51180.57530.5436
Multi-Scale TsMixer (ours)200.49560.56640.5303
600.51520.58210.5501
800.52190.59020.5548
Table A5. Performance comparison of baseline models and Multi-Scale TsMixer on SSE50 classification task for different input sequences.
Table A5. Performance comparison of baseline models and Multi-Scale TsMixer on SSE50 classification task for different input sequences.
ModelDaysPrecision (↑)Specificity (↑)Accuracy (↑)
BaselinesTSMixer200.48450.54870.5279
600.48930.55320.5328
800.50670.55860.5393
Informer200.48630.56120.5287
600.48760.56450.5311
800.48970.56870.5343
DLinear200.49760.55930.5312
600.50270.57420.5394
80 0.5104 ̲ ̲ 0.5738 0.5462 ̲ ̲
PatchTST200.44350.52630.4882
600.45760.53890.5053
800.48080.55670.5225
Transformer200.48610.56680.5229
600.49730.57150.5367
800.49510.57720.5384
TimesNet20 0.4949 ̲ ̲ 0.56590.5301
600.51470.58080.5483
800.50730.57810.5446
TimeMixer200.4905 0.5699 ̲ ̲ 0.5318 ̲ ̲
600.49340.57120.5335
800.50870.57190.5373
Multi-Scale TsMixer (ours)200.49240.57210.5362
60 0.5116 ̲ ̲ 0.5775 ̲ ̲ 0.5428 ̲ ̲
800.52090.58950.5531

References

  1. Gorton, G.; Rouwenhorst, K.G. Facts and fantasies about commodity futures. Financ. Anal. J. 2006, 62, 47–68. [Google Scholar] [CrossRef]
  2. Fang, W.; Zhang, S.; Xu, C. Improving prediction efficiency of Chinese stock index futures intraday price by VIX-Lasso-GRU Model. Expert Syst. Appl. 2024, 238, 121968. [Google Scholar] [CrossRef]
  3. Hu, G.X.; Wang, J. A review of China’s financial markets. Annu. Rev. Financ. Econ. 2022, 14, 465–507. [Google Scholar] [CrossRef]
  4. Wang, X.; Wang, X.; Li, B.; Bai, Z. The nonlinear characteristics of Chinese stock index futures yield volatility: Based on the high frequency data of CSI300 stock index futures. China Financ. Rev. Int. 2020, 10, 175–196. [Google Scholar] [CrossRef]
  5. Carpenter, J.N.; Lu, F.; Whitelaw, R.F. The real value of China’s stock market. J. Financ. Econ. 2021, 139, 679–696. [Google Scholar] [CrossRef]
  6. Chen, R.; Pan, B. Chinese stock index futures price fluctuation analysis and prediction based on complementary ensemble empirical mode decomposition. Math. Probl. Eng. 2016, 2016, 3791504. [Google Scholar] [CrossRef]
  7. Li, Y.; Shen, D.; Wang, P.; Zhang, W. Does intraday time-series momentum exist in Chinese stock index futures market? Financ. Res. Lett. 2020, 35, 101292. [Google Scholar] [CrossRef]
  8. Wang, C.P.; Lin, S.H.; Huang, H.H.; Wu, P.C. Using neural network for forecasting TXO price under different volatility models. Expert Syst. Appl. 2012, 39, 5025–5032. [Google Scholar] [CrossRef]
  9. Sezer, O.B.; Gudelek, M.U.; Ozbayoglu, A.M. Financial time series forecasting with deep learning: A systematic literature review: 2005–2019. Appl. Soft Comput. 2020, 90, 106181. [Google Scholar] [CrossRef]
  10. Zhou, F.; Zhang, Q.; Sornette, D.; Jiang, L. Cascading logistic regression onto gradient boosted decision trees for forecasting and trading stock indices. Appl. Soft Comput. 2019, 84, 105747. [Google Scholar] [CrossRef]
  11. Asghar, M.Z.; Rahman, F.; Kundi, F.M.; Ahmad, S. Development of stock market trend prediction system using multiple regression. Comput. Math. Organ. Theory 2019, 25, 271–301. [Google Scholar] [CrossRef]
  12. Junior, P.R.; Pamplona, E.D.O.; Salomon, F.L.R. ARIMA: An applied time series forecasting model for the Bovespa stock index. Appl. Math. 2014, 5, 3383. [Google Scholar] [CrossRef]
  13. Wang, C. Forecast on price of agricultural futures in China based on ARIMA model. Asian Agric. Res. 2016, 8, 9–16. [Google Scholar]
  14. Byun, S.J.; Cho, H. Forecasting carbon futures volatility using GARCH models with energy volatilities. Energy Econ. 2013, 40, 207–221. [Google Scholar] [CrossRef]
  15. Arnerić, J.; Poklepović, T. Nonlinear extension of asymmetric garch model within neural network framework. Comput. Sci. Inf. Technol. 2016, 6, 101–111. [Google Scholar]
  16. Lin, Z. Modelling and forecasting the stock market volatility of SSE Composite Index using GARCH models. Future Gener. Comput. Syst. 2018, 79, 960–972. [Google Scholar] [CrossRef]
  17. Tang, Y.; Song, Z.; Zhu, Y.; Yuan, H.; Hou, M.; Ji, J.; Tang, C.; Li, J. A survey on machine learning models for financial time series forecasting. Neurocomputing 2022, 512, 363–380. [Google Scholar] [CrossRef]
  18. Yan, H.; Ouyang, H. Financial time series prediction based on deep learning. Wirel. Pers. Commun. 2018, 102, 683–700. [Google Scholar] [CrossRef]
  19. Cao, J.; Li, Z.; Li, J. Financial time series forecasting model based on CEEMDAN and LSTM. Phys. A Stat. Mech. Its Appl. 2019, 519, 127–139. [Google Scholar] [CrossRef]
  20. Lazcano, A.; Herrera, P.J.; Monge, M. A combined model based on recurrent neural networks and graph convolutional networks for financial time series forecasting. Mathematics 2023, 11, 224. [Google Scholar] [CrossRef]
  21. Olorunnimbe, K.; Viktor, H. Ensemble of temporal Transformers for financial time series. J. Intell. Inf. Syst. 2024, 62, 1087–1111. [Google Scholar] [CrossRef]
  22. Huang, W.; Nakamori, Y.; Wang, S.Y. Forecasting stock market movement direction with support vector machine. Comput. Oper. Res. 2005, 32, 2513–2522. [Google Scholar] [CrossRef]
  23. Illa, P.K.; Parvathala, B.; Sharma, A.K. Stock price prediction methodology using random forest algorithm and support vector machine. Mater. Today Proc. 2022, 56, 1776–1782. [Google Scholar] [CrossRef]
  24. Yang, J.; Zhao, C.; Yu, H.; Chen, H. Use GBDT to predict the stock market. Procedia Comput. Sci. 2020, 174, 161–171. [Google Scholar] [CrossRef]
  25. Lei, M.; Wang, X. DWT-TimesNet: An Improved Model for Wheat Price Long-Time Series Forecasting. In Proceedings of the 2023 IEEE International Conference on Big Data (BigData), Sorrento, Italy, 15–18 December 2023; pp. 3431–3438. [Google Scholar]
  26. Waswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the NIPS, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  27. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 11106–11115. [Google Scholar]
  28. Zeng, A.; Chen, M.; Zhang, L.; Xu, Q. Are transformers effective for time series forecasting? In Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence and Thirty-Fifth Conference on Innovative Applications of Artificial Intelligence and Thirteenth Symposium on Educational Advances in Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; AAAI Press: Washington, DC, USA, 2023. [Google Scholar]
  29. Nie, Y.; Nguyen, N.H.; Sinthong, P.; Kalagnanam, J. A Time Series is Worth 64 Words: Long-term Forecasting with Transformers. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  30. Wang, S.; Wu, H.; Shi, X.; Hu, T.; Luo, H.; Ma, L.; Zhang, J.Y.; Zhou, J. TimeMixer: Decomposable Multiscale Mixing for Time Series Forecasting. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  31. Wu, H.; Hu, T.; Liu, Y.; Zhou, H.; Wang, J.; Long, M. TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  32. Chen, S.A.; Li, C.L.; Arik, S.O.; Yoder, N.C.; Pfister, T. TSMixer: An All-MLP Architecture for Time Series Forecasting. Trans. Mach. Learn. Res. arXiv 2023, arXiv:2303.06053. [Google Scholar]
  33. Jiang, W. Applications of deep learning in stock market prediction: Recent progress. Expert Syst. Appl. 2021, 184, 115537. [Google Scholar] [CrossRef]
  34. Bhandari, H.N.; Rimal, B.; Pokhrel, N.R.; Rimal, R.; Dahal, K.R.; Khatri, R.K. Predicting stock market index using LSTM. Mach. Learn. Appl. 2022, 9, 100320. [Google Scholar] [CrossRef]
  35. Tolstikhin, I.O.; Houlsby, N.; Kolesnikov, A.; Beyer, L.; Zhai, X.; Unterthiner, T.; Yung, J.; Steiner, A.; Keysers, D.; Uszkoreit, J.; et al. Mlp-mixer: An all-mlp architecture for vision. Adv. Neural Inf. Process. Syst. 2021, 34, 24261–24272. [Google Scholar]
  36. Wang, C.; Chen, Y.; Zhang, S.; Zhang, Q. Stock market index prediction using deep Transformer model. Expert Syst. Appl. 2022, 208, 118128. [Google Scholar] [CrossRef]
Figure 1. Illustration of our proposed Multi-Scale TsMixer network architecture.
Figure 1. Illustration of our proposed Multi-Scale TsMixer network architecture.
Mathematics 13 01415 g001
Figure 2. Workflow of predicting index futures with aggregating constituent stock’s predicted results.
Figure 2. Workflow of predicting index futures with aggregating constituent stock’s predicted results.
Mathematics 13 01415 g002
Figure 3. Comparison of the accuracy of different baseline models and our proposed model across various input lengths.
Figure 3. Comparison of the accuracy of different baseline models and our proposed model across various input lengths.
Mathematics 13 01415 g003
Figure 4. A comparative comparison of profits between the single-future and three fusion methods on three indices.
Figure 4. A comparative comparison of profits between the single-future and three fusion methods on three indices.
Mathematics 13 01415 g004
Table 1. Performance comparison of baseline models and Multi-Scale TsMixer on the CSI500 regression task for different input sequences (upper arrows indicate that larger values are better; lower arrows vice versa).
Table 1. Performance comparison of baseline models and Multi-Scale TsMixer on the CSI500 regression task for different input sequences (upper arrows indicate that larger values are better; lower arrows vice versa).
ModelDaysMAPE (↓)MAE (↓) R 2 (↑)
BaselinesTSMixer200.974255.2110 0.9802 ̲ ̲
600.973655.1976 0.9803 ̲ ̲
800.972855.1477 0.9803 ̲ ̲
Informer200.980455.60390.9801
600.994256.44210.9796
800.971455.12110.9802
DLinear20 0.9730 ̲ ̲ 55.14660.9803
600.971855.0922 0.9803 ̲ ̲
800.969754.9805 0.9803 ̲ ̲
PatchTST201.208569.29010.9738
601.188368.04900.9741
801.085061.75780.9778
Transformer200.982855.75720.9801
600.971855.15750.9802
800.974855.2661 0.9803 ̲ ̲
TimesNet200.988156.10150.9798
600.979355.50100.9802
80 0.9694 ̲ ̲ 54.9557 ̲ ̲ 0.9803 ̲ ̲
TimeMixer200.976855.3756 0.9802 ̲ ̲
600.973155.21060.9802
800.973655.2130 0.9803 ̲ ̲
Multi-Scale TsMixer (ours)200.9729 55.1629 ̲ ̲ 0.9803
60 0.9724 ̲ ̲ 55.1209 ̲ ̲ 0.9804
800.968254.90340.9804
Table 2. Performance comparison of baseline models and Multi-Scale TsMixer on CSI500 classification task for different input sequences (upper arrows indicate that larger values are better; lower arrows vice versa).
Table 2. Performance comparison of baseline models and Multi-Scale TsMixer on CSI500 classification task for different input sequences (upper arrows indicate that larger values are better; lower arrows vice versa).
ModelDaysPrecision (↑)Specificity (↑)Accuracy (↑)
BaselinesTSMixer200.48210.54650.5297
600.49060.55300.5336
800.51090.55210.5439
Informer200.48890.56370.5292
600.48500.56780.5234
800.49120.56980.5307
DLinear200.49300.5625 0.5336 ̲ ̲
600.5031 0.5773 ̲ ̲ 0.5424
80 0.5123 ̲ ̲ 0.5764 0.5497 ̲ ̲
PatchTST200.44480.52510.4868
600.45830.53790.5044
800.48250.55830.5234
Transformer200.48810.56560.5222
600.49800.56860.5395
800.48910.57910.5336
TimesNet200.49090.56740.5307
600.49390.57020.5338
800.5079 0.5804 ̲ ̲ 0.5468
TimeMixer20 0.4952 ̲ ̲ 0.5691 ̲ ̲ 0.5321
60 0.5121 ̲ ̲ 0.5763 0.5439 ̲ ̲
800.51000.57030.5395
Multi-Scale TsMixer (ours)200.49720.57070.5365
600.51380.57880.5482
800.52030.59180.5526
Table 3. Classification performance of different fusion methods in the Multi-Scale TsMixer on three indices with 80-day input lengths (bold indicates best performance; upper arrows indicate that larger values are better; lower arrows vice versa).
Table 3. Classification performance of different fusion methods in the Multi-Scale TsMixer on three indices with 80-day input lengths (bold indicates best performance; upper arrows indicate that larger values are better; lower arrows vice versa).
IndexFusion MethodPrecision (↑)Specificity (↑)Accuracy (↑)
CSI500Average-based Fusion0.54630.60910.5787
Weighted-based fusion0.54610.61560.5856
Weighted-based decay fusion0.58670.64140.6185
CSI300Average-based fusion0.53670.59890.5707
Weighted-based fusion0.54490.61260.5857
Weighted-based decay fusion0.57450.63260.6075
SSE50Average-based fusion0.53980.59590.5739
Weighted-based fusion0.54340.60950.5800
Weighted-based decay fusion0.56850.62870.6002
Table 4. Comparison of profitability and maximum drawdown for different index futures using various fusion methods (bold indicates best performance; upper arrows indicate that larger values are better; lower arrows vice versa).
Table 4. Comparison of profitability and maximum drawdown for different index futures using various fusion methods (bold indicates best performance; upper arrows indicate that larger values are better; lower arrows vice versa).
IndexMethodProfit (%) (↑)Maximum Drawdown (%) (↓)
SSE50Single-future55.5217.05
Average-based fusion107.3316.38
Weighted-based fusion144.0614.00
Weighted-based decay fusion257.1014.89
CSI300Single-future67.7319.21
Average-based fusion146.4713.99
Weighted-based fusion189.5513.26
Weighted-based decay fusion314.8115.74
CSI500Single-future139.1818.83
Average-based fusion210.0216.52
Weighted-based fusion260.5814.40
Weighted-based decay fusion388.7117.74
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pei, Z.; Yan, J.; Yan, J.; Yang, B.; Liu, X. Multi-Scale TsMixer: A Novel Time-Series Architecture for Predicting A-Share Stock Index Futures. Mathematics 2025, 13, 1415. https://doi.org/10.3390/math13091415

AMA Style

Pei Z, Yan J, Yan J, Yang B, Liu X. Multi-Scale TsMixer: A Novel Time-Series Architecture for Predicting A-Share Stock Index Futures. Mathematics. 2025; 13(9):1415. https://doi.org/10.3390/math13091415

Chicago/Turabian Style

Pei, Zhiyuan, Jianqi Yan, Jin Yan, Bailing Yang, and Xin Liu. 2025. "Multi-Scale TsMixer: A Novel Time-Series Architecture for Predicting A-Share Stock Index Futures" Mathematics 13, no. 9: 1415. https://doi.org/10.3390/math13091415

APA Style

Pei, Z., Yan, J., Yan, J., Yang, B., & Liu, X. (2025). Multi-Scale TsMixer: A Novel Time-Series Architecture for Predicting A-Share Stock Index Futures. Mathematics, 13(9), 1415. https://doi.org/10.3390/math13091415

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop