Next Article in Journal
Machine-Learned Codes from EHR Data Predict Hard Outcomes Better than Human-Assigned ICD Codes
Previous Article in Journal
Pattern Matching-Based Denoising for Images with Repeated Sub-Structures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ADTime: Adaptive Multivariate Time Series Forecasting Using LLMs

1
Institute of Computing Technology, University of Chinese Academy of Sciences, Beijing 100190, China
2
Research Institute of Petroleum Processing, SINOPEC, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Mach. Learn. Knowl. Extr. 2025, 7(2), 35; https://doi.org/10.3390/make7020035
Submission received: 19 February 2025 / Revised: 9 April 2025 / Accepted: 11 April 2025 / Published: 17 April 2025
(This article belongs to the Section Learning)

Abstract

:
Large language models (LLMs) have recently demonstrated notable performance, particularly in addressing the challenge of extensive data requirements when training traditional forecasting models. However, these methods encounter significant challenges when applied to high-dimensional and domain-specific datasets. These challenges primarily arise from inability to effectively model inter-variable dependencies and capture variable-specific characteristics, leading to suboptimal performance in complex forecasting scenarios. To address these limitations, we propose ADTime, an adaptive LLM-based approach for multivariate time series forecasting. ADTime employs advanced preprocessing techniques to identify latent relationships among key variables and temporal features. Additionally, it integrates temporal alignment mechanisms and prompt-based strategies to enhance the semantic understanding of forecasting tasks by LLMs. Experimental results show that ADTime outperforms state-of-the-art methods, reducing MSE by 9.5% and MAE by 6.1% on public datasets, and by 17.1% and 13.5% on domain-specific datasets. Furthermore, zero-shot experiments on real-world refinery datasets demonstrate that ADTime exhibits stronger generalization capabilities across various transfer scenarios. These findings highlight the potential of ADTime in advancing complex, domain-specific time series forecasting tasks.

1. Introduction

Time series forecasting is widely used in various domains, including weather prediction [1], health monitoring [2], energy management [3,4], and financial forecasting [5]. These methods are essential for analyzing temporal trends, patterns, and relationships among variables, providing critical data support for economic and infrastructure planning.
Benefiting from self-attention mechanisms, transformer-based approaches have leveraged strengths in modeling long-term dependencies for sequential data. They enhance models’ ability to capture temporal patterns by modifying the model architecture or employing processing techniques for time series data. However, the promising performance of transformer-based methods relies heavily on large and domain-specific pretraining datasets. Traditional forecasting approaches generally require domain-specific models, which demand significant expertise and have limited applicability across domains. In contrast, the development of pretrained large language models (LLMs) such as GPT [6,7] and LLaMA [8] offers a promising alternative. These models acquire general domain knowledge through pretraining on vast amounts of textual data, enabling their application across a wide range of downstream tasks [9]. Additionally, LLMs exhibit exceptional performance in few-shot and zero-shot learning, leading to their increasing adoption in time series forecasting even in data-scarce scenarios. Therefore, many studies have directly fed time series data into LLMs. Through techniques such as prompt engineering, LLMs can surpass domain-specific models with minimal or no reliance on downstream training data.
Real-world time series data often have high dimensionality and exhibit significant variations across different domains, posing several challenges for domain-specific time series forecasting with large models. First, real-world time series data often exhibit unique temporal features that cannot be effectively captured through a single fixed strategy. Existing approaches tend to apply the same strategy across all data dimensions; for instance, Tempo performs temporal decomposition uniformly across all variables, which may lead to information loss and interference. Second, existing LLM-based forecasting methods typically model each univariate time series independently using a channel-independent approach, which neglects inter-variable dependencies [10,11]. For example, in PatchTST, each model’s input token contains information from only a single channel.
To address these challenges, we propose an adaptive time series forecasting framework called ADTime, which is designed to efficiently leverage inter-variable dependencies and diverse temporal patterns in domain-specific time series analysis. To enable flexible time series feature learning, we design an adaptive time series processing module that dynamically adjusts processing strategies based on data features. After performing correlation-based joint modeling of variables and analyzing each channel’s periodicity using FFT, STL decomposition is applied to strongly periodic data. On the other hand, weakly periodic data retain their original sequences, ensuring effective feature extraction and adaptation to diverse temporal patterns. To enhance the ability of LLMs to understand temporal features, we integrate temporal text alignment and prompt engineering into the design of the embedding layers within the LLM. Extensive evaluations show that ADTime offers an effective solution for diverse domain-specific datasets and that it excels in both few-shot and zero-shot time series forecasting scenarios. The main contributions of this work are as follows:
  • We investigate the domain-specific correlations among multivariate data and the heterogeneous temporal characteristics across different channels. This leads to the development of an elastic forecasting method tailored to diverse temporal patterns, enhancing the capacity of LLMs to process multi-scale temporal information.
  • We propose an adaptive LLM-based framework called ADTime, which integrates two modules in order to process multivariate time series data. First, we apply channel clustering and temporal decomposition based on the distinct temporal features of each channel. This is followed by an embedding construction module that uses text alignment and adaptively designed prompts to improve the model’s ability to capture temporal patterns.
  • Experimental results on seven forecasting benchmarks show that our model outperforms existing methods. Specifically, ADTime reduces MSE by 9.5% and MAE by 6.1% on public datasets, while achieving reductions of 17.1% and 13.5% on real-world refinery datasets (the refinery dataset utilized in this study was provided by Sinopec, and contains sensor and product generation data from multiple refineries that are not publicly available due to privacy concerns). Additionally, through ablation studies and model analysis, we further investigate the key factors contributing to ADTime’s effectiveness.

2. Related Work

This paper focuses on the development of multivariate time series forecasting. In this section, we provide a comprehensive review of key developments in time series decomposition, channel strategies, and the application of transformer-based models, highlighting both their contributions and limitations.

2.1. Transformer-Based Time Series Models

In recent years, transformer-based models [12,13,14,15] have been widely applied to time series tasks. Informer [16] introduced the ProbSparse self-attention mechanism to distill input sequences, resulting in enhanced efficiency when processing long sequences. Autoformer [17] is an auto-correlation mechanism for uncovering dependencies between variables, while FEDformer [18] leverages a Fourier-enhanced block and Wavelet-enhanced block to capture key structures in time series data. PatchTST [10] turns time series into patches in order to capture comprehensive semantic information, while Pathformer [19] introduces a dual attention mechanism for enhanced transformer-based time series modeling capabilities. However, these approaches are not suitable for domain-specific time series prediction with small amounts of data.
Meanwhile, the development of pretrained large language models (LLMs) has provided new possibilities for time series analysis [20]. The robust text generation and pattern recognition capabilities of LLMs provide a foundation for their application to time series tasks. LLM-centered approaches primarily enhance model performance through two main strategies. One common approach is to fine-tune LLMs for specific downstream tasks [21]. For example, GPT4TS [22] employs a frozen pretrained transformer in which the self-attention and feedforward layers of the LLM remain frozen while fine-tuning is performed for time series tasks. While noteworthy, this method still requires training on task-specific data.
Another strategy is to modify the input format for LLMs, allowing them to be used directly without further training. Several studies [23] have adopted multimodal processing techniques that align time series features with LLM representations. For instance, TEST [24] applies contrastive learning to generate embeddings for time series tokens, while TIME-LLM [25] employs multihead attention to reprogram time series data into textual representations. Additionally, several works have leveraged prompts to improve performance on specific tasks. For example, TEMPO [26] utilizes soft prompt strategies to accommodate the dynamic nature of time series data. However, these approaches all apply the uniform embedding construction method across all variants, which fails to capture the unique characteristics of time series data in different domains. As a result, their performance in real-world temporal tasks remains limited.

2.2. Decomposition of Time Series

Time series decomposition is a classic method in time series analysis [18,26,27] that involves breaking down a time series into multiple independent components such as trend, seasonality, and residuals [28]. For instance, the trend component represents the long-term change in the time series, while the seasonality component captures shorter-term periodic fluctuations. N-BEATS [29] addresses decomposition by incorporating a neural network-based decomposition module to estimate local decomposition components, while Autoformer [17] applies traditional mean filtering to convolve the time series, then extracts the trend component while treating the remainder as the seasonality component. TimesNet [30] leverages Fourier transformation to decompose time series in the frequency domain, extracting multiple seasonal components with varying periodic lengths in order to better capture diverse periodic patterns.

2.3. Channel Strategies in Time Series Models

In multivariate time series forecasting, channel strategies significantly influence model performance. These strategies can be broadly classified into two categories, namely, channel-independent (CI) and channel-dependent (CD), based on how the inter-channel dependencies are modeled.
CI strategies treat multivariate time series as N independent univariate series, then predict each channel separately [10,11]. CI strategies excel at capturing the temporal dynamics of individual variables and may yield higher prediction accuracy, particularly when inter-channel similarities are low; however, they often overlook potential interactions between channels, which can limit their generality and robustness.
On the other hand, CD strategies treat all channel data as joint inputs while leveraging cross-channel information to model complex interdependencies [16,17,18,31]. TimesNet [30] uses CNNs to capture cross-dimensional dependencies, while Crossformer [32] employs a two-stage attention mechanism to effectively model both cross-time and cross-dimensional dependencies. In general, CD strategies aim to learn and represent intricate relationships among variables; however, they can be vulnerable to noise in time series data. Transformer-based models relying on CD strategies are particularly prone to overfitting noise, which can degrade their performance [33]. Therefore, designing an effective channel strategy is crucial for realizing improved forecasting accuracy [34,35].

3. Methodology

A multivariate time series can be represented as X = [ X 1 , . . . , X K ] R K × N , where K denotes the length of the historical data, N denotes the number of variables observed, and X t R N denotes the observation value at time t. Specifically, X t is composed of individual components x t i , where x t i denotes the observation value of the i-th channel at time t. Thus, and we can write X t = x t i , i = 1 , . . . , N .
The goal of multivariate time series forecasting is to learn a function that accurately predicts the time series data for the next H time steps based on the historical series of K time steps. In this context, the predicted values are denoted by X ^ t , which represents the forecasted observation vector at time t:
Y = [ X ^ t , . . . , X ^ t + H 1 ] = F ( X t K , . . . , X t 1 )
where Y is the sequence of predicted values over the next H time steps and F ( · ) is the forecasting function.
As depicted in Figure 1, our proposed framework comprises three core components: (1) multivariate time series processing, (2) temporal alignment mechanisms, and (3) adaptive prompting for LLMs. The framework begins with the temporal processing module, which first identifies inter-variable dependencies using the channel clustering module. It then evaluates the periodicity of each channel to determine whether STL decomposition is required. Following this, the aligned series are mapped to the textual semantic space. Ultimately, adaptive prompts are integrated to enhance the temporal understanding capabilities of LLMs, significantly improving their overall forecasting performance.
During the training process, only the parameters of the embedding layer and output projection layer need to be updated, while the backbone LLM remains frozen. This approach significantly enhances efficiency and reduces resource requirements compared to building or fine-tuning an entire time series model.

3.1. Multivariate Time Series Process

In order to address the inherent complexities and varied characteristics of multivariate domain data, we apply a set of strategies consisting of channel clustering and time series decomposition. These strategies can work together to improve the robustness and predictive performance of time series models, especially in the presence of complex dependencies and varying periodic behaviors across channels.
Channel Clustering. In multivariate time series analysis, the correlation between different dimensions is critical for predictive performance. When using domain-specific data, the interrelationships between features reflect underlying mechanisms and system parameters. Highly correlated dimensions often exhibit similar time series patterns, whereas low-correlation dimensions reveal more independent variation characteristics [35]. For channels with high similarity, the individual channel identity information contributes minimally to predictive power, as the model can derive comparable insights from other similar channels. By leveraging this channel similarity, reliance on specific identity information can be minimized, allowing the model to focus on shared features across channels for improved generalization and predictive performance.
To combine the strengths of both independent and dependent approaches, we propose a channel clustering strategy. As illustrated in Figure 2, the channel clustering module captures essential cross-channel dependencies by clustering channels based on their feature correlations. We use the distance correlation to measure interdependencies between channels and cluster time series when the distance correlation exceeds a predefined threshold T h c o r .
After channel clustering, the channels are grouped into C clusters. Each cluster i contains c i channels and can be denoted as X ˜ i R K × c i , { i = 1 , , C } . Channels with high similarity undergo joint embedding, enabling the model to exploit intrinsic connections and realize enhanced predictive accuracy. For other variants, we apply a channel-independent strategy that allows these channels pass through the projection layer independently, which reduces the influence of irrelevant data and improves the model’s generalization ability. The output of the channel clustering module is represented as X ˜ i R K , { i = 1 , , C } .
Time-Series Decomposition. The periodicity of time series can vary significantly across different channels. To address this, we apply the fast Fourier transform (FFT) to assess the periodicity of each channel. If the periodicity exceeds a predefined threshold F t h , then the data are decomposed into trend, seasonal, and residual components using STL decomposition [28]. These components are treated as separate features and passed into the embedding layer. For channels with weak periodicity, we utilize the original series directly in order to preserve its dynamic characteristics. This differentiated decomposition approach enables the model to better capture complex time series features, improving both the accuracy and robustness of multivariate time series forecasting.
For a strongly periodic time series X i R K with i dimensions and a given time step K, the original data can be decomposed into seasonal, trend, and residual components by additive decomposition:
X i = X T i + X S i + X R i
where X S i represents the periodicity of the time series, X T i represents the overall upward or downward trend or state of the time series within a given time step, and X R i is the remaining portion after removing the seasonal and trend components.
Patching. Unlike traditional LLMs, which process discrete word tokens, time series data require additional methods for extracting meaningful semantic information from individual time points. Here, we adopt the patching approach proposed by [10]. This approach significantly extends the historical time horizon while minimizing information redundancy and improving model efficiency.

3.2. Modal Alignment

Large language models operate on discrete tokens, while time series data are inherently continuous; thus, aligning time series data with natural language modalities is a crucial step in unlocking the full potential of these models for time series processing. As depicted in Figure 3, we adopt the reprogramming scheme introduced by [25], which transforms input time series into a natural language modality representation. This alignment enhances the backbone network’s ability to comprehend temporal patterns.
The scheme uses a multihead cross-attention layer to align the time series patch X P with the pretrained word embeddings E R V × D in the LLM backbone network, where V denotes the size of the vocabulary and D is the hidden dimension of the backbone model. To reduce the space required for reprogramming, we maintain a small set of textual prototypes, denoted as E R V × D , where V V . The prototypes are obtained using learnable linear projections.
Specifically, for each header h = { 1 , , H } , we define the query matrix Q h = X P W Q h , key matrix K h = E W K h , and value matrix V h = E W V h . The operation of reprogramming the time series patch in each attention head is defined as
Z h = Attention ( Q h , K h , V h ) = softmax ( Q h K h d h ) V h .

3.3. Adaptive Prompts for LLMs

To enhance the ability of large language models (LLMs) to understand prediction tasks and leverage domain knowledge encoded during pretraining [36], we design adaptive prompts that guide the LLM in performing precise time series forecasting. These prompts are tailored to the input series, and consist of the following three key components:
  • Dataset Context: Provides background information about the input time series, incorporating semantic features along with domain-specific knowledge such as chemical- or industry-specific insights.
  • Data Features: Includes statistical metrics (e.g., mean and standard deviation), temporal attributes (e.g., periodicity and trends), and correlations derived during preprocessing.
  • Task Instructions: These instructions explicitly guide the LLM to perform specific predictive tasks by leveraging the provided data features.
  • An example of such a prompt is shown in Figure 4, which illustrates how the prompt incorporates dataset context, data features, and task instructions for a specific time series forecasting problem.
The adaptive prompts are further customized based on the periodicity of the data, leveraging the temporal decomposition module described in Section 3.1. For strongly periodic data, decomposition into trend, seasonal, and residual components enables the LLM to process the embedded temporal features directly. In these cases, prompts are streamlined to avoid redundancy. Conversely, weakly periodic data, which are not decomposed, require more detailed prompts to help the LLM identify temporal patterns. This differentiation ensures optimal utilization of the LLM’s inference capabilities, enhancing both its temporal pattern recognition and overall prediction performance.
To fully leverage temporal patterns in the data, we concatenate the adaptive prompts with their corresponding components and input them to the pretrained LLM. For highly periodic channel data, the trend and seasonal components are combined to construct the final input representation x = x T x S x R , where ⊕ denotes the concatenation operation. For weakly periodic data, the original time series embedding x is directly input to the LLM without decomposition. This flexible input strategy effectively aligns with the varying characteristics of the data, enhancing the accuracy and robustness of time series forecasting.

4. Results

4.1. Experiment Setup

Dataset. We evaluated seven popular publicly available datasetsthat have been widely used for benchmarking long-term forecasting models: ETTh1, ETTh2, ETTm1, ETTm2, Weather, Electricity, and Traffic. Additionally, we introduced two real-world refinery datasets: the LIMS dataset, which records the variation in reactant concentrations during the refining process, and the PI dataset, which tracks changes in each generator sensor throughout the same process.
Baseline and Metrics. We selected a series of recent and representative baseline models, including the MLP-based model DLinear [11], the transformer-based models PatchTST [10], Informer [16], and Autoformer [17], and the recent LLM-based generalized temporal prediction models GPT4TS [22], TEMPO [26], and Time-LLM [25]. The evaluation metrics were the mean square error (MSE), mean absolute error (MAE), and relative standard error (RSE).
Experimental Setup. The input time series length T was set to 336 steps, and we considered four different prediction horizons of H { 96 , 192 , 336 , 720 } . To ensure fairness, all LLM-based methods used GPT-2 [9] as the backbone model. We trained the models using the AdamW optimizer and selected the model with the lowest average validation loss for testing. For fair evaluation, we used the optimal experimental configuration as provided in the official code to implement the base models as well as to adhere to TimesNet’s [30] experimental settings. All the experiments were implemented with PyTorch 2.3.1 on a single NVIDIA A100 40GB GPU, and each experiment was repeated at least three times using different random seeds.

4.2. Few-Shot Forecasting

LLMs have demonstrated strong performance in few-shot learning; therefore, we evaluated the few-shot capability of ADTime on both the public and refinery datasets. Following the classic experimental setup [22], we made predictions using 5% of the training data. The few-shot results for our model and the baselines on the public and domain-specific datasets are presented in Table 1 and Table 2.
The results show that ADTime outperforms all baseline models. After testing on seven public time series datasets, ADTime achieved the best performance in 22 out of 25 instances, outperforming the second-best model by 9.5%, 6.1%, and 8.5% in terms of MSE, MAE, and RSE, respectively. In our evaluation on seven domain-specific datasets, ADTime achieved state-of-the-art (SOTA) performance in 19 out of 20 instances, with improvements of 17.1%, 13.5%, and 10.6% over GPT4TS in terms of MSE, MAE, and RSE, respectively.
It is worth noting that PatchTST and DLinear outperform the other baselines on the public datasets, particularly the larger-scale Traffic and Electricity datasets. On the domain-specific datasets, the LLM-based approaches (Time-LLM and GPT4TS) show significantly better performance than the traditional models. This supports our motivation for using LLMs in time series forecasting. More specifically, these results suggest that prompt-based learning can enhance LLMs’ understanding of domain knowledge, leading to improved forecasting abilities.

4.3. Zero-Shot Forecasting

This task evaluates the cross-dataset adaptability of our proposed scheme, i.e., its temporal prediction performance on dataset A when the model is trained on dataset B without encountering any samples from dataset A. In this section, we evaluate various cross-dataset scenarios using the ETT, PI, and LIMS datasets.
The zero-shot experimental results are shown in Table 3. Our proposed ADTime consistently outperforms all existing baselines across various transfer learning scenarios, demonstrating strong transferability and generalization ability. In particular, the results on real-world refinery datasets PI and LIMS further verify the robustness and scalability of ADTime in practical scenarios. The suboptimal performance of existing models such as Time-LLM and GPT4TS underscores the inherent limitations of generic LLMs in zero-shot settings [37]. These findings suggest that integrating LLMs with well-designed prompts and time series decomposition can significantly improve the accuracy and stability of zero-shot forecasting in real-world domains.

4.4. Ablation Study

Embedding Variants. In order to better understand the effectiveness of the embedding construction module design in ADTime, we constructed three variants of the model and conducted ablation experiments on the ETTh and PI datasets. The experimental results illustrated in Figure 5 demonstrate that removing either the modality alignment or the prompt module leads to diminished time series forecasting capability on the part of LLMs. Text-aligned ablation indicates that eliminating gaps between the text space and temporal data space leads to better prediction results than simple splicing. The degradation of prediction results in the absence of prompts also validates the significance of prompts in guiding the prediction task.
Adaptive Decomposition. We performed an ablation study to verify the significance of adaptive temporal decomposition. Two variants of the module (w/o decomposition and direct decomposition) were constructed and evaluated on the PI and Weather datasets. The experimental results in Figure 6a demonstrate that temporal decomposition is effective for feature extraction. In particular, our adaptive decomposition method selectively identifies significant features, significantly improving the model’s performance.
Furthermore, to validate the generalizability of the adaptive temporal decomposition approach, we modified the PatchTST and TEMPO by incorporating the adaptive module into both. The effectiveness of this modification was evaluated, and results are presented in Table 4. The experimental results confirm that integrating the adaptive temporal decomposition module into existing models consistently enhances their predictive performance. This improvement is evident across different datasets and baseline architectures. It highlights that by selectively focusing on significant components, the module improves the robustness and accuracy of diverse time series models, making it a valuable addition for tackling both general and domain-specific forecasting tasks.
Prompt Variants. We designed two prompt variants in this study: Prompts(brief) provides only domain information and task instructions, while Prompts(precise) additionally summarizes time series characteristics such as the mean value of the variable and the trend value. The prediction performance of the different prompt designs is illustrated in Figure 6b.

4.5. Model Analysis

LLM Variants. We compared the performance of different large language model variants with varying sizes, as shown in Table 5. The experiment reveals several key findings. First, increasing the model size with the same architecture consistently improves forecasting performance, confirming that larger models are more effective at capturing complex temporal dependencies. Second, the performance of different model architectures varies across datasets, indicating that the knowledge learned by each model may be biased toward specific patterns. These results underscore the importance of selecting a model that aligns with the specific characteristics of the dataset in order to achieve optimal accuracy.
Computational Complexity. For the ETTh1-96 dataset with a batch size of 64, ADTime requires approximately 14.25 GB of memory. This is comparable to other LLM-based methods, with a difference of only a few hundred MB. Compared to PatchTST, ADTime’s memory usage is higher by approximately 12.5 GB. However, given a 40 GB GPU, this additional memory cost remains manageable. In terms of processing speed, ADTime achieves 2.68 it/s. While lower than lightweight models such as PatchTST, this is sufficient for real-time processing. Because most time series forecasting tasks operate with hourly or daily sampling intervals, this level of computational efficiency is well within practical requirements, ensuring feasibility in real-world applications. Furthermore, the slight tradeoff in processing speed is justified by the significant improvements in forecasting accuracy and interpretability, making ADTime highly suitable for real-world deployment in scenarios requiring high-stakes decision-making.
T-SNE Visualization. Figure 7 presents the T-SNE visualization of the embedding process for the ETTh dataset at different stages of the multimodal learning pipeline consisting of series and prompt embeddings. Initially, the series embedding shows a sparse distribution of data points, indicating a lack of structured correlation in the time series data. In contrast, prompt embedding exhibits a more compact distribution with the data points clustered together, reflecting the strong correlations established through the prompt-based approach. After the alignment process, the embeddings from both the time series and the prompt overlap closely, demonstrating that text alignment successfully bridges the gap between the temporal and textual spaces. This result confirms the effective integration of time series and prompt embeddings, enhancing the model’s ability to leverage both modalities in a unified manner.

5. Conclusions

In this paper, we propose ADTime, an adaptive LLM-based framework for multivariate time series forecasting. ADTime leverages an adaptive time series process model to dynamically captures temporal features, followed by a prompt and textual alignment module that transforms structured data into textual representation. By integrating these mechanisms, ADTime achieves state-of-the-art predictive performance across multiple domain-specific datasets, particularly in complex real-world scenarios. Our results demonstrate the effectiveness of adaptive decomposition and clustering in enhancing model robustness and generalization. However, challenges remain in domain adaptation. Future research could explore domain-adaptive fine-tuning to improve robustness. Additionally, leveraging LLMs for end-to-end time series forecasting and reasoning represents a promising direction that could offer deeper insights into temporal dynamics.

Author Contributions

Conceptualization, J.P., Y.Z. and T.L.; methodology, J.P. and T.L.; software, validation, writing—original draft preparation, visualization, J.P.; formal analysis, investigation, resources, K.Q. and Y.Z.; data curation, J.P., T.L. and J.Y.; writing—review and editing, T.L. and J.Y.; supervision, project administration, Q.W. and K.Q.; funding acquisition, K.Q. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Energy R&D Center of Petroleum Refining Technology (RIPP, SINOPEC) under Grant 36800000-23-ZC0607-0107.

Data Availability Statement

One part of the data used in this study is confidential and cannot be shared due to restrictions. The public part of the dataset is available in Google Drive (https://drive.google.com/drive/folders/1ZOYpTUa82_jCcxIdTmyr0LXQfvaM9vIy), accessed on 10 April 2025.

Acknowledgments

This work was partially supported by the National Energy R&D Center of Petroleum Refining Technology (RIPP, SINOPEC). The authors would like to express their gratitude for the valuable support and resources provided during this research.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Angryk, R.A.; Martens, P.C.; Aydin, B.; Kempton, D.; Mahajan, S.S.; Basodi, S.; Ahmadzadeh, A.; Cai, X.; Filali Boubrahimi, S.; Hamdi, S.M.; et al. Multivariate time series dataset for space weather data analytics. Sci. Data 2020, 7, 227. [Google Scholar] [CrossRef] [PubMed]
  2. Wang, Y.; Huang, N.; Li, T.; Yan, Y.; Zhang, X. Medformer: A Multi-Granularity Patching Transformer for Medical Time-Series Classification. Adv. Neural Inf. Process. Syst. 2024, 37, 36314–36341. [Google Scholar]
  3. Li, L.; Su, X.; Zhang, Y.; Lin, Y.; Li, Z. Trend modeling for traffic time series analysis: An integrated study. IEEE Trans. Intell. Transp. Syst. 2015, 16, 3430–3439. [Google Scholar] [CrossRef]
  4. Zhu, Z.; Chen, W.; Xia, R.; Zhou, T.; Niu, P.; Peng, B.; Wang, W.; Liu, H.; Ma, Z.; Gu, X.; et al. Energy forecasting with robust, flexible, and explainable machine learning algorithms. AI Mag. 2023, 44, 377–393. [Google Scholar] [CrossRef]
  5. Yu, X.; Chen, Z.; Ling, Y.; Dong, S.; Liu, Z.; Lu, Y. Temporal Data Meets LLM–Explainable Financial Time Series Forecasting. arXiv 2023, arXiv:2306.11025. [Google Scholar]
  6. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  7. Liu, T.; Low, B.K.H. Goat: Fine-tuned llama outperforms gpt-4 on arithmetic tasks. arXiv 2023, arXiv:2305.14201. [Google Scholar]
  8. Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.A.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F.; et al. Llama: Open and efficient foundation language models. arXiv 2023, arXiv:2302.13971. [Google Scholar]
  9. Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language models are unsupervised multitask learners. OpenAI Blog 2019, 1, 9. [Google Scholar]
  10. Nie, Y.; Nguyen, N.H.; Sinthong, P.; Kalagnanam, J. A Time Series is Worth 64 Words: Long-term Forecasting with Transformers. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  11. Zeng, A.; Chen, M.; Zhang, L.; Xu, Q. Are transformers effective for time series forecasting? AAAI Conf. Artif. Intell. 2023, 37, 11121–11128. [Google Scholar] [CrossRef]
  12. Wen, Q.; Zhou, T.; Zhang, C.; Chen, W.; Ma, Z.; Yan, J.; Sun, L. Transformers in time series: A survey. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, Macao, China, 19–25 August 2023; pp. 6778–6786. [Google Scholar]
  13. Liu, Y.; Wu, H.; Wang, J.; Long, M. Non-stationary transformers: Exploring the stationarity in time series forecasting. Adv. Neural Inf. Process. Syst. 2022, 35, 9881–9893. [Google Scholar]
  14. Kitaev, N.; Kaiser, Ł.; Levskaya, A. Reformer: The efficient transformer. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 30 April 2020. [Google Scholar]
  15. Liu, Y.; Hu, T.; Zhang, H.; Wu, H.; Wang, S.; Ma, L.; Long, M. itransformer: Inverted transformers are effective for time series forecasting. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2023. [Google Scholar]
  16. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. AAAI Conf. Artif. Intell. 2021, 35, 11106–11115. [Google Scholar] [CrossRef]
  17. Wu, H.; Xu, J.; Wang, J.; Long, M. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Adv. Neural Inf. Process. Syst. 2021, 34, 22419–22430. [Google Scholar]
  18. Zhou, T.; Ma, Z.; Wen, Q.; Wang, X.; Sun, L.; Jin, R. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In Proceedings of the International conference on machine learning. PMLR, Baltimore, MD, USA, 17–23 July 2022; pp. 27268–27286. [Google Scholar]
  19. Chen, P.; Zhang, Y.; Cheng, Y.; Shu, Y.; Wang, Y.; Wen, Q.; Yang, B.; Guo, C. Pathformer: Multi-scale transformers with adaptive pathways for time series forecasting. arXiv 2024, arXiv:2402.05956. [Google Scholar]
  20. Jin, M.; Zhang, Y.; Chen, W.; Zhang, K.; Liang, Y.; Yang, B.; Wang, J.; Pan, S.; Wen, Q. Position: What Can Large Language Models Tell Us about Time Series Analysis. In Proceedings of the Forty-first International Conference on Machine Learning, Vienna, Austria, 21–27 July 2024. [Google Scholar]
  21. Liu, C.; Yang, S.; Xu, Q.; Li, Z.; Long, C.; Li, Z.; Zhao, R. Spatial-temporal large language model for traffic prediction. In Proceedings of the 2024 25th IEEE International Conference on Mobile Data Management (MDM), Brussels, Belgium, 24–27 June 2024; pp. 31–40. [Google Scholar]
  22. Zhou, T.; Niu, P.; Sun, L.; Jin, R. One fits all: Power general time series analysis by pretrained lm. Adv. Neural Inf. Process. Syst. 2023, 36, 43322–43355. [Google Scholar]
  23. Liu, C.; Xu, Q.; Miao, H.; Yang, S.; Zhang, L.; Long, C.; Li, Z.; Zhao, R. TimeCMA: Towards LLM-Empowered Time Series Forecasting via Cross-Modality Alignment. In Proceedings of the AAAI, Dubai, United Arab Emirates, 20–22 May 2025. [Google Scholar]
  24. Sun, C.; Li, H.; Li, Y.; Hong, S. TEST: Text prototype aligned embedding to activate LLM’s ability for time series. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  25. Jin, M.; Wang, S.; Ma, L.; Chu, Z.; Zhang, J.Y.; Shi, X.; Chen, P.Y.; Liang, Y.; Li, Y.F.; Pan, S.; et al. Time-llm: Time series forecasting by reprogramming large language models. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  26. Cao, D.; Jia, F.; Arik, S.O.; Pfister, T.; Zheng, Y.; Ye, W.; Liu, Y. Tempo: Prompt-based generative pre-trained transformer for time series forecasting. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  27. Wang, S.; Wu, H.; Shi, X.; Hu, T.; Luo, H.; Ma, L.; Zhang, J.Y.; Zhou, J. Timemixer: Decomposable multiscale mixing for time series forecasting. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  28. Cleveland, R.B.; Cleveland, W.S.; McRae, J.E.; Terpenning, I. STL: A seasonal-trend decomposition. J. Off. Stat. 1990, 6, 3–73. [Google Scholar]
  29. Oreshkin, B.N.; Carpov, D.; Chapados, N.; Bengio, Y. N-BEATS: Neural basis expansion analysis for interpretable time series forecasting. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  30. Wu, H.; Hu, T.; Liu, Y.; Zhou, H.; Wang, J.; Long, M. Timesnet: Temporal 2d-variation modeling for general time series analysis. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  31. Liu, S.; Yu, H.; Liao, C.; Li, J.; Lin, W.; Liu, A.X.; Dustdar, S. Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting. In Proceedings of the International Conference on Learning Representations, Virtual, 25–29 April 2022. [Google Scholar]
  32. Zhang, Y.; Yan, J. Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  33. Han, L.; Ye, H.J.; Zhan, D.C. The capacity and robustness trade-off: Revisiting the channel independent strategy for multivariate time series forecasting. IEEE Trans. Knowl. Data Eng. 2024, 36, 7129–7142. [Google Scholar] [CrossRef]
  34. Wang, X.; Zhou, T.; Wen, Q.; Gao, J.; Ding, B.; Jin, R. CARD: Channel aligned robust blend transformer for time series forecasting. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  35. Chen, J.; Lenssen, J.E.; Feng, A.; Hu, W.; Fey, M.; Tassiulas, L.; Leskovec, J.; Ying, R. From Similarity to Superiority: Channel Clustering for Time Series Forecasting. arXiv 2024, arXiv:2404.01340. [Google Scholar]
  36. Li, X.L.; Liang, P. Prefix-tuning: Optimizing continuous prompts for generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing; Volume 1: Long Papers; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 4582–4597. [Google Scholar]
  37. Gruver, N.; Finzi, M.; Qiu, S.; Wilson, A.G. Large language models are zero-shot time series forecasters. Adv. Neural Inf. Process. Syst. 2024, 36, 861. [Google Scholar]
Figure 1. The model framework of ADTime.
Figure 1. The model framework of ADTime.
Make 07 00035 g001
Figure 2. Channels are grouped into C clusters; intra-cluster channels use the CD strategy, while inter-cluster channels are independently passed through the projection layer.
Figure 2. Channels are grouped into C clusters; intra-cluster channels use the CD strategy, while inter-cluster channels are independently passed through the projection layer.
Make 07 00035 g002
Figure 3. Reprogramming architecture.
Figure 3. Reprogramming architecture.
Make 07 00035 g003
Figure 4. Prompt example.
Figure 4. Prompt example.
Make 07 00035 g004
Figure 5. Ablation study of the embedding module: (a) experimental results on the ETTh dataset and (b) experimental results on the PI dataset.
Figure 5. Ablation study of the embedding module: (a) experimental results on the ETTh dataset and (b) experimental results on the PI dataset.
Make 07 00035 g005
Figure 6. Ablation study of model design: (a) decomposition module ablation experiment on the ETTh and PI datasets and (b) prompt variants on the ETTh and PI datasets.
Figure 6. Ablation study of model design: (a) decomposition module ablation experiment on the ETTh and PI datasets and (b) prompt variants on the ETTh and PI datasets.
Make 07 00035 g006
Figure 7. T-SNE visualization of embeddings at different stages for the ETTh dataset.
Figure 7. T-SNE visualization of embeddings at different stages for the ETTh dataset.
Make 07 00035 g007
Table 1. Few-shot learning on 5% training data from the public datasets. All results are averaged from four different forecasting horizons: H { 96 , 192 , 336 , 720 } . Lower values indicate better performance. Red: best; Blue: second-best.
Table 1. Few-shot learning on 5% training data from the public datasets. All results are averaged from four different forecasting horizons: H { 96 , 192 , 336 , 720 } . Lower values indicate better performance. Red: best; Blue: second-best.
MethodsOursTEMPOTime-LLMGPT4TSPatchTSTDlinearAutoformerInformer
MetricMSEMAERSEMSEMAERSEMSEMAERSEMSEMAERSEMSEMAERSEMSEMAERSEMSEMAERSEMSEMAERSE
ETTh1960.6020.5140.7370.6420.5501.0450.5350.6720.7790.5370.5250.8330.6890.5740.8180.6170.5840.9870.7900.6500.8501.0100.7400.960
1920.6860.5800.7420.8050.5981.2240.9710.6780.9380.7320.6150.7810.8590.6670.8460.7140.6210.7660.9750.7360.9021.2110.8401.058
3360.9260.6600.9201.3230.8031.1171.3550.8171.1101.1080.7600.9641.1750.7930.9920.9470.7620.7910.9650.7290.8991.6210.8951.182
avg0.7380.5840.8000.9230.6511.1290.9540.7220.9420.7920.6340.8590.9070.6780.8850.7590.6550.8480.9100.7050.8841.2810.8251.067
ETTh2960.3880.4120.5020.4070.4290.5430.3950.4150.5070.5310.4940.7230.4010.4220.6370.4360.4870.6730.5900.5900.6103.6201.5201.520
1920.3980.4090.5430.4400.4430.5690.5160.4760.5760.3910.4140.7940.4120.4460.8140.6030.4980.6980.5190.4200.7163.6931.3141.649
3360.3940.4110.5180.4470.4540.6010.5240.4850.5780.3930.3930.7360.4220.4510.8400.8940.5140.9090.5350.4310.7473.4531.6602.021
avg0.3930.4110.5210.4310.4420.5710.4780.4590.5540.4390.4340.7510.4120.4400.7640.6450.4990.7600.5480.4800.6913.5881.4981.730
ETTm1960.4110.4100.6140.5350.4780.8350.4780.4380.6580.3990.4150.6310.4290.4290.6800.4220.4740.8150.7800.6100.8401.0400.7600.970
1920.4550.4500.6620.5490.4950.8550.4790.4510.6590.4900.4660.6390.5990.5330.7060.4270.4340.5960.9070.6770.8691.4760.7250.936
3360.4350.4310.6460.6350.5310.8880.4430.4240.6340.5870.5190.6990.6830.5920.7540.4930.4720.6410.9230.6950.8781.8780.8390.949
7200.4870.4670.6940.6670.5531.0510.5930.5120.7330.7280.5910.7811.0100.7110.9200.5630.5250.6871.0360.7620.9311.0960.8441.110
avg0.4470.4390.6540.5970.5140.9070.4980.4560.6710.5510.4980.6870.6800.5660.7650.4760.4760.6850.9120.6860.8791.3720.7920.991
ETTm2960.1730.2530.3740.2250.3070.4890.2400.3110.3970.1940.2770.4410.2190.2990.4680.3200.3800.4600.3700.3400.4903.5901.5101.530
1920.2270.2480.3960.3020.3500.4530.2770.3330.4260.2500.2880.4860.2630.3220.5060.3270.3810.6270.3110.3220.5753.5981.4751.317
3360.2500.3250.4540.3480.3820.5100.3420.3750.4720.2960.3460.5420.2980.3410.5600.3790.4240.6940.3310.3330.6033.7921.2131.380
7200.4070.4200.5150.4270.4210.5620.4710.4450.5520.4020.4160.6960.3710.3930.6590.4640.4880.7940.3850.3760.6753.6231.5031.548
avg0.2640.3110.4350.3250.3650.5040.3320.3660.4620.2860.3320.5410.2880.3390.5490.3730.4180.6440.3490.3430.5863.6511.4251.444
Weather960.1830.2360.4380.1930.2390.6220.1930.2450.5790.1850.2440.4500.1920.2420.4430.2800.3300.7000.3000.3600.7200.5100.5100.940
1920.1930.2570.6040.2460.2890.7060.2440.2820.6500.2430.2830.6480.2640.3060.6760.2270.2810.6270.3810.4010.8130.7080.5991.108
3360.2840.3190.7180.3260.3430.8510.2900.3140.7080.3010.3300.7210.3140.3390.7360.2790.3240.6940.4040.4110.8350.6800.5831.083
7200.3470.3710.7910.3720.3810.9350.3780.3730.8090.3710.3810.8020.3950.3960.8270.3640.3880.7940.4310.4250.8630.6330.5521.047
avg0.2520.2960.6380.2840.3130.7780.2770.3030.6860.2750.3090.6550.2910.3210.6710.2880.3310.7040.3790.3990.8080.6330.5611.044
ECL960.1820.3040.4080.1800.2920.4050.2150.3280.4610.1460.2420.3820.1440.2390.3790.2300.2610.3920.5100.5500.7101.3300.9301.150
1920.1580.2410.3820.1820.2850.4470.2250.3340.4710.1720.2730.4130.2250.3290.4720.1570.2520.3940.4360.4950.6561.3260.9471.145
3360.1780.2770.4140.2090.3090.5240.2330.3410.4810.1870.2860.4300.3200.4150.5630.2680.2680.4140.4400.5030.6601.3250.9501.146
7200.2240.3120.4490.2790.3550.5740.2830.3770.5310.2410.3310.4900.2920.3820.5390.2180.3090.4650.5190.5510.7191.2790.9261.128
avg0.1860.2840.4130.2130.3100.4880.2390.3450.4860.1860.2830.4290.2450.3410.4880.2180.2720.4160.4760.5250.6861.3150.9381.142
Traffic960.3920.3200.6030.4380.3110.6250.6530.4430.6690.4340.3200.6590.4130.2860.6420.4600.3460.6340.8750.5131.0031.7300.9001.090
1920.4190.3110.5740.4960.3550.6870.6900.4680.6860.45970.3290.5600.4410.2980.5480.4470.3140.5720.8760.5380.7721.5470.8181.027
3360.4370.3140.5990.5030.3560.6810.6900.4650.6830.4790.3340.5690.4600.3090.5570.4840.3400.5720.9620.5850.8061.5800.8291.033
avg0.4160.3150.5920.4790.3410.6640.6770.4590.6790.4570.3280.5960.4380.2980.5830.4640.3330.5920.9040.5450.8611.6190.8491.050
For the ETTh and Traffic datasets, the 720 horizon is not included because 5% of the time series was insufficient to constitute a training set.
Table 2. Few-shot learning on 5% training data from the refinery datasets. All results are averaged from four different forecasting horizons: H { 96 , 192 , 336 } . Lower values indicate better performance. Red: best; Blue: second-best.
Table 2. Few-shot learning on 5% training data from the refinery datasets. All results are averaged from four different forecasting horizons: H { 96 , 192 , 336 } . Lower values indicate better performance. Red: best; Blue: second-best.
MethodsOursTEMPOTime-LLMGPT4TSPatchTSTDlinearAutoformerInformer
MetricMSEMAERSEMSEMAERSEMSEMAERSEMSEMAERSEMSEMAERSEMSEMAERSEMSEMAERSEMSEMAERSE
PI1961.2260.986 0.1181.2341.1300.1191.2481.0600.1181.1751.0350.1151.9711.4350.1491.6691.3530.1371.8031.7120.1365.9683.7772.014
1920.9710.9920.1201.6281.2770.1431.5341.2150.1321.6691.3110.1371.5931.3310.1342.1962.8170.2851.9651.7330.1495.9263.7491.725
3361.9291.3340.1422.0221.4230.1571.9381.4040.1482.0511.4470.1523.6492.4660.2031.9632.5240.3472.1121.7990.1555.9083.7431.628
avg1.3751.1040.1261.6281.2770.1401.5731.2260.1331.6321.2640.1352.4051.7440.1621.9432.2310.2561.9601.7480.1465.9343.7561.789
PI2960.4100.6510.3660.4420.7240.4120.4170.6700.3930.4580.6830.4120.4300.7730.3990.3820.7100.3760.4891.0090.4092.7301.8211.540
1920.5520.7110.3710.5550.8420.4380.5090.7640.4290.6521.0160.4860.5370.8000.4410.5420.8690.4430.5181.0910.4332.8062.2101.829
3360.4980.6770.3630.5870.8790.4450.5180.8090.4260.6861.0050.4900.6660.9300.4830.6081.0840.4610.5021.1060.4192.8992.6611.726
avg0.4870.6790.3660.5280.8150.4320.4810.7480.4160.5990.9010.4630.5440.8340.4410.5110.8880.4270.5031.0690.4202.8122.2311.698
PI3960.9890.8300.1072.0161.2380.1131.3601.0880.0941.0960.9960.0841.5721.2320.1011.6831.2290.1042.2361.7910.1342.5681.6630.240
1920.9660.8830.1141.9531.2250.1172.2261.3540.1191.2661.1480.0901.8981.3550.1101.6271.4970.3222.9601.8930.1372.6801.6580.231
3361.3291.2910.1062.1391.4280.1733.7161.7070.1531.9861.3680.1122.1871.4920.1171.5471.4570.3124.1812.1020.1622.6951.7540.223
avg1.0941.0010.1092.0361.2970.1342.4341.3830.1221.4491.1700.0951.8861.3600.1091.6191.3940.2463.1261.9290.1452.6481.6910.231
PI4960.1790.1840.3640.1970.2190.3950.1860.2100.3630.2900.2810.4540.2370.2550.4100.2390.2220.3970.4020.3660.5341.4160.6901.002
1920.2040.2110.3560.2300.2770.4200.2080.2260.3840.2190.2840.4480.2290.2450.4040.2290.2450.4040.3890.4090.5261.5450.7411.048
3360.2630.2720.4540.2790.3220.4650.2690.2660.4370.2790.2790.4450.2500.2540.4210.2940.2900.4570.6230.5810.6651.7140.8711.103
avg0.2150.2220.3910.2350.2730.4270.2210.2340.3950.2630.2810.4490.2390.2510.4120.2540.2520.4190.4710.4520.5751.5580.7671.051
PI5960.6510.2560.3260.6540.2650.3620.6950.2620.3580.6700.2680.3460.8080.3090.3860.7530.2870.3720.9120.3910.4105.8560.9251.038
1920.5250.2740.2870.7750.2980.4140.7880.2840.3810.7510.3890.5170.8250.3110.3900.9000.3020.4071.0310.4390.4365.9571.0001.048
3360.6630.3720.3320.8780.3270.3910.9630.3470.4210.8650.3180.3990.8250.3110.3901.2680.4180.4831.3680.6750.5025.7441.0331.028
avg0.6130.3010.3150.7690.2970.3890.8160.3020.3870.7620.3250.4210.8190.3110.3880.9740.3360.4211.1040.5020.4495.8530.9861.038
LIMS1960.9550.5730.7811.2770.7021.1281.7260.8641.0451.0200.6060.8071.0930.6380.8302.0781.1131.1521.1180.6900.8382.0141.1021.124
1921.1900.6470.8430.9870.6001.1621.6170.8080.9821.1920.6550.8431.7110.8471.0112.1591.1471.1351.1920.7160.8362.2921.2261.158
avg1.0730.6100.8121.1320.6511.1451.6710.8361.0131.1060.6300.8251.4020.7420.9202.1191.1301.1441.1550.7030.8372.1531.1641.141
LIMS2960.4300.4580.4780.4980.4770.6630.6240.5560.5970.5460.4890.5040.5230.4900.5462.1571.1311.1090.6810.6070.6241.9411.1041.054
1920.4380.4490.4820.5060.5270.8970.6860.5790.6310.5450.4800.5080.5230.4830.5511.7371.0251.0050.5740.5330.5782.2251.1881.139
3360.4330.4780.5000.5600.5720.6640.6540.5460.6200.5450.4750.5120.5700.4990.5792.2001.1461.1370.5770.5180.5872.1701.1701.139
avg0.4340.4610.4860.5210.5260.7410.6550.5600.6160.5450.4810.5080.5390.4910.5592.0311.1001.0830.6100.5530.5972.1121.1541.111
For the LIMS1 dataset, the 360 horizon is not included because 5% of the time series was insufficient to constitute a training set.
Table 3. Zero-shot learning results. Red: best; Blue: second-best.
Table 3. Zero-shot learning results. Red: best; Blue: second-best.
MethodsOursTime-LLMTEMPOGPT4TSPatchTST
MetricMSEMAERSEMSEMAERSEMSEMAERSEMSEMAERSEMSEMAERSE
ETTh1→ETTh20.3670.4220.4310.3750.4240.4730.3920.4360.4600.4060.4220.4600.3800.4050.443
ETTh2→ETTh10.6110.5130.6990.6400.5240.7610.7420.6060.8390.7030.5780.7350.6650.5330.747
ETTm1→ETTm20.2030.2840.3640.2170.2920.3680.2380.2890.3790.2640.2950.3930.2960.3340.408
ETTm2→ETTm10.4420.4250.6330.4780.4380.6580.5150.4780.6620.7690.5670.6790.5680.4920.639
PI5→PI60.7030.2480.3290.7310.2640.3561.7040.2600.6910.9800.3400.4250.8290.3260.391
PI6→PI50.2040.2040.3640.1940.2150.3710.2120.2110.3520.2100.2270.3860.2250.2420.400
LIMS1→LIMS20.5030.4700.5040.8500.6720.6960.6280.5550.6700.5750.4950.5220.6010.5250.592
LIMS2→LIMS11.0750.6750.8101.2200.6850.8771.2700.6971.1161.1760.6840.7941.2980.6920.872
Table 4. Performance comparison of different methods. Black: best performance.
Table 4. Performance comparison of different methods. Black: best performance.
VariantsWeather-96Weather-192PI5-96PI5-192
MSEMAEMSEMAEMSEMAEMSEMAE
ADTime0.1830.2360.1930.2470.6510.2560.5250.274
TEMPO0.1930.2390.2460.2890.6540.2650.7750.298
TEMPO +0.1870.2370.1990.2570.6520.2590.7250.286
PatchTST0.1920.2420.2640.3060.8080.3090.8250.311
PatchTST +0.1860.2380.2130.2690.7790.2770.7810.293
Table 5. Forecasting performance with language model variants; GPT(K) refers to the GPT backbone model with K layers; similarly, LLaMA(K) denotes the LLaMA backbone with K layers. Black: best performance.
Table 5. Forecasting performance with language model variants; GPT(K) refers to the GPT backbone model with K layers; similarly, LLaMA(K) denotes the LLaMA backbone with K layers. Black: best performance.
ModelETTh1-96ETTh1-192ETTh2-96ETTh2-192PI4-96PI4-192PI5-96PI5-192
MSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAE
Llama3.2 (32)0.5610.5010.6470.5380.3750.3930.3720.4260.1520.1660.1900.1930.6790.2670.7790.284
Llama3.2 (8)0.5920.5070.6980.5510.3610.4160.3770.4270.1590.1780.1920.2030.6940.2600.6830.280
GPT-2 (12)0.6860.5540.6320.5290.3230.3680.3470.4090.1790.1920.2000.2100.6750.2570.7360.277
GPT-2 (6)0.6020.5140.6860.5800.3880.4120.3980.4090.1790.1840.2040.2110.6510.2560.5250.274
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pei, J.; Zhang, Y.; Liu, T.; Yang, J.; Wu, Q.; Qin, K. ADTime: Adaptive Multivariate Time Series Forecasting Using LLMs. Mach. Learn. Knowl. Extr. 2025, 7, 35. https://doi.org/10.3390/make7020035

AMA Style

Pei J, Zhang Y, Liu T, Yang J, Wu Q, Qin K. ADTime: Adaptive Multivariate Time Series Forecasting Using LLMs. Machine Learning and Knowledge Extraction. 2025; 7(2):35. https://doi.org/10.3390/make7020035

Chicago/Turabian Style

Pei, Jinglei, Yang Zhang, Ting Liu, Jingbin Yang, Qinghua Wu, and Kang Qin. 2025. "ADTime: Adaptive Multivariate Time Series Forecasting Using LLMs" Machine Learning and Knowledge Extraction 7, no. 2: 35. https://doi.org/10.3390/make7020035

APA Style

Pei, J., Zhang, Y., Liu, T., Yang, J., Wu, Q., & Qin, K. (2025). ADTime: Adaptive Multivariate Time Series Forecasting Using LLMs. Machine Learning and Knowledge Extraction, 7(2), 35. https://doi.org/10.3390/make7020035

Article Metrics

Back to TopTop