Next Article in Journal
Supply–Demand Matching of Engineering Construction Materials in Complex Mountainous Areas Based on Complex Environment Two-Stage Stochastic Programing
Previous Article in Journal
Dynamics of Activation and Regulation of the Immune Response to Attack by Viral Pathogens Using Mathematical Modeling
Previous Article in Special Issue
Bayesian Joint Modeling Analysis of Longitudinal Proportional and Survival Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bayesian Model Selection for Addressing Cold-Start Problems in Partitioned Time Series Prediction

by
Jaeseong Yoo
1 and
Jihoon Moon
2,*
1
Statistical Ground, Seoul 06979, Republic of Korea
2
Department of AI and Big Data, Soonchunhyang University, Asan 31538, Republic of Korea
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(17), 2682; https://doi.org/10.3390/math12172682
Submission received: 16 July 2024 / Revised: 23 August 2024 / Accepted: 26 August 2024 / Published: 28 August 2024
(This article belongs to the Special Issue Bayesian Statistical Analysis of Big Data and Complex Data)

Abstract

:
How to effectively predict outcomes when initial time series data are limited remains unclear. This study investigated the efficiency of Bayesian model selection to address the lack of initial data for time series analysis, particularly in cold-start scenarios—a common challenge in predictive modeling. We utilized a comprehensive approach that juxtaposed observational data against various candidate models through strategic partitioning. This method contrasted traditional reliance on distance measures like the L2 norm. Instead, it applied statistical tests to validate model efficacy. Notably, the introduction of an interactive visualization tool featuring a slide bar for setting significance levels marked a significant advancement over conventional p-value displays. Our results affirm that when observational data align with a candidate model, effective predictions are possible, albeit with necessary considerations of stationarity and potential structural breaks. These findings underscore the potential of Bayesian methods in predictive analytics, especially when initial data are scarce or incomplete. This research not only enhances our understanding of model selection dynamics but also sets the stage for future investigations into more refined predictive frameworks.

1. Introduction

How to effectively predict outcomes in numeric time series data when faced with a cold-start problem where little or no initial data are available remains unclear. The cold-start problem poses a significant challenge to automated data modeling for computer-based information systems, especially when user and item data are insufficient to make accurate predictions [1,2]. This issue is particularly notorious in recommendation systems that utilize information filtering techniques to tailor item displays to user preferences, typically referencing user profiles based on specific features [3]. However, the term “cold-start” becomes less straightforward when it is applied to numeric time series data [4]. In these cases, if the data-generating process (DGP) is accurately known, predictions can be reliably made, sometimes even without any observed data [5]. This capability underscores the importance of understanding and identifying the general DGP, as it allows for effective prediction even at the onset of data collection. Consequently, a substantial body of research has been dedicated to understanding the general DGP rather than directly addressing the cold-start issue in time series. This research is of paramount importance because it addresses the foundational challenge of making informed predictions in the absence of substantial initial data, a common scenario in many practical applications, including economics and environmental science.
The objective of this study was to address the issue of cold start in time series analysis, particularly in the context of specific conditions that might challenge conventional data modeling approaches. These specific conditions include the following:
  • Insufficient length of observed time series: It is typically recommended that a sample size of at least 50 should be used for data analysis, though this is not a strict rule [6]. The necessary sample size can vary depending on data characteristics, domain, and analysis methods [7,8]. A cold-start problem exists if available data are insufficient and below the suggested minimum sample size. This scenario can greatly affect the reliability and accuracy of analyses performed.
  • Incomplete cyclical or seasonal variation: To achieve accurate predictions, it is necessary to observe a complete cycle for time series exhibiting seasonal or cyclical variations. If available data only cover a partial cycle, as illustrated in Figure 1, it becomes challenging to identify repeating patterns and make accurate forecasts. This limitation can significantly hinder the effectiveness of predictive analytics in such cases.
  • Post-structural break and insufficient data: The occurrence of structural breaks in a time series with significant shifts in data distribution further complicates the process of forecasting [9]. As illustrated in Figure 2, inadequate data collection following a structural break can hinder the ability to forecast future trends accurately. This limitation is critical as it affects the prediction of whether the observed anomaly will lead to further structural changes.
  • Lack of data on specific features or items: The introduction of new features or items into a dataset can present forecasting challenges when these additions lack sufficient data to support them. To understand and predict the impact of these new elements, it is necessary to have adequate historical data. To make reliable predictions, it is necessary to have a robust dataset that captures the characteristics of new elements fully.
Figure 1. Cases with less than one full seasonal cycle. The letter “A” indicates an incomplete seasonal cycle within the known area, making it difficult to identify repeating patterns and accurately predict future values.
Figure 1. Cases with less than one full seasonal cycle. The letter “A” indicates an incomplete seasonal cycle within the known area, making it difficult to identify repeating patterns and accurately predict future values.
Mathematics 12 02682 g001
Figure 2. Insufficient data observation following a structural break. The arrows labeled “A” and “B” mark significant structural breaks in the time series, indicating sharp shifts in the distribution of the data. The area following the second structural break “B” shows insufficient data for accurate forecasting, making it difficult to predict future trends or detect further structural changes.
Figure 2. Insufficient data observation following a structural break. The arrows labeled “A” and “B” mark significant structural breaks in the time series, indicating sharp shifts in the distribution of the data. The area following the second structural break “B” shows insufficient data for accurate forecasting, making it difficult to predict future trends or detect further structural changes.
Mathematics 12 02682 g002
The partitioning of time series data has historically employed methods based on the autocorrelation function (ACF) of related time series as originally proposed by Chow and Lin [10] and Denton [11]. When reference time series are unavailable, interpolated methods have been utilized to segment time series data. This approach is particularly advantageous when the scope or temporal constraint of the research necessitates disaggregation or aggregation of time series data. Moreover, methods that employ the ACF have evolved into Box and Jenkins’ techniques and their derivatives [12]. As these techniques have been refined and adapted to the complexities of modern data analysis, they have proven to be crucial in addressing the nuanced effects of seasonal and trend variations in predictive accuracy. The impact of seasonal and trend variations in time series data on various predictions is significant. In light of this, the Box and Jenkins method has been extensively utilized. With the advancement of artificial neural networks (ANNs) in the 2000s, a combined approach with Box and Jenkins’ techniques has emerged [13]. However, these methods often fail to capture seasonal effects in nonstationary time series data effectively. To minimize variations, Puma-Villanueva et al. [14,15] have segmented time series data, while Sarkar et al. [16] and Leverger et al. [17] have proposed effective partitioning methods to enhance classification accuracy.
The cold-start problem has been a persistent issue in recommender systems, particularly within collaborative filtering technologies. Despite the development of numerous solutions, no perfect solution has been identified. However, the creation of an item–item matrix to determine correlations among items has emerged as a prevalent alternative. This matrix is then used to infer user preferences based on the most recent data. Xie et al. [18] have found that the cold-start problem in time series arises from issues such as missing data and high dimensions. They have attempted long-term forecasting by considering metadata, high-dimensional structures, and seasonality. Building on these insights, we will examine how these approaches are applied in practical scenarios, with a particular focus on challenges of data scarcity at the outset of analysis. Consider the scenario where we are provided with time series data as illustrated in Figure 3, which depicts an example of observed time series data. If the DGP for this series is known, one can directly utilize the DGP. Alternatively, if the DGP is unknown, it is possible to estimate the DGP from available data, enabling the kind of predictions shown in Figure 4, where the application of DGP on observed data is displayed. However, difficulties may arise in instances where a cold-start problem is present and where it is impossible to make predictions with limited data initially available.
Bayesian model selection is particularly well suited to address the cold-start problem in time series analysis because it allows for the integration of prior knowledge and statistical evidence, even when data are sparse or incomplete. This method facilitates the generation of informed decisions on the most likely models that could have generated the observed data. Before we explore the specific scenario involving Bayesian model selection, it is essential to address the following research questions that guide our investigation:
  • How can we accurately predict time series outcomes when only limited data are initially available?
  • Among various candidate models proposed, which is most likely to have generated the observed time series data?
  • How effectively can Bayesian model selection overcome challenges presented by the cold-start problem in time series prediction?
These questions serve to frame our examination of model selection and the application of Bayesian principles to resolve uncertainties in the cold-start scenario. We propose that the time series could originate from one of several candidate models. This setup requires knowledge of these models and introduces a Bayesian model selection scenario, where we must decide which of the competing models is the most probable given the data. This decision process is demonstrated in Figure 3, which displays probabilities of the observed time series data belonging to different candidate models.
This study employed Bayesian model selection in cases where insufficient data prevented a comprehensive analysis of the observed time series. The aim of this study was to derive inferences under these constraints and present results in a visually comprehensible format, facilitating the interpretation of complex statistical decisions. The following contributions of this paper address critical gaps in time series analysis and provide practical tools for managing data scarcity and uncertainty in predictive modeling:
  • This paper demonstrates how Bayesian model selection can enhance forecasting accuracy when traditional methods are inadequate due to a lack of data. A robust framework is presented to generate reliable forecasts from limited observations, ensuring that predictions remain viable even when datasets are sparse.
  • The proposed innovative approach incorporates prior knowledge into model selection, which is of value when historical data are unavailable or do not reflect current circumstances. Integrating prior knowledge makes model predictions more reliable in challenging situations.
  • The proposed visualization techniques can simplify complex Bayesian results, increasing stakeholder understanding and confidence in model predictions. This approach enables a deeper comprehension of statistical outcomes, improving the decision-making quality.
The remainder of this paper is structured as follows: Section 2 reviews related studies in the field. Section 3 provides requisite background information for this study. Section 4 offers a comprehensive account of conducted experiments. Section 5 discusses visualization techniques. Finally, Section 6 presents the conclusion.

2. Related Work

Related work on cold-start problems in recommendation systems and anomaly detection in time series data is categorized into classification and regression problems. This section reviews previous studies, highlighting differences between the present research and previous studies. Moreover, several studies have addressed classification problems in the context of cold-start and anomaly detection.
For example, Xu et al. [19] presented a variational embedding learning framework (VELF) to address the cold-start problem in predicting click-through rate. The VELF addressed data sparsity by learning probabilistic embeddings and applying trainable, regularized priors using information regarding users and advertisements. Experiments revealed that the VELF outperformed traditional methods and enhanced generalization and robustness in cold-start scenarios. In addition, Al Rossais et al. [20] proposed an approach to improve cold-start recommendations by generating item-based stereotypes from metadata without considering user-item ratings. Their experiments on MovieLens/IMDb (MovieLens, Minneapolis, MN, USA) and Amazon datasets (Amazon Inc., Seattle, WA, USA) found that these stereotypes enhanced both recommendation quality and computational performance and outperformed traditional singular value decomposition-based approaches. Pirasteh et al. [21] designed an enhanced hybrid collaborative filtering method to improve personalized recommendations by combining similarity measures. This method combines user and item similarities based on ratings and genres, addressing cold-start problems and outperforming conventional collaborative filtering techniques.
Further, Rohani et al. [22] introduced an enhanced content-based algorithm using social networking to address the cold-start problem in recommender systems by incorporating user preferences with those of friends and faculty. The efficacy of the enhanced content-based algorithm using social networking was evaluated on the MyExpert academic social network (University of Malaya, Kuala Lumpur, Malaysia), demonstrating significantly enhanced recommendation accuracy compared with random, collaborative, and content-based algorithms. Ni et al. [23] predicted student performance on learner-sourced questions by integrating signed graph neural networks with large language model embeddings. Their method modeled student responses using a signed bipartite graph and employed a contrastive learning framework to enhance noise resilience, significantly outperforming existing baselines in predictive accuracy and robustness. In addition, Tey et al. [24] designed a social network-based recommender system to address the cold-start problem by leveraging indirect relationships between users and their friends’ friends. The system integrated user preferences and social media interactions. It significantly improved recommendation accuracy using data from Yelp (Yelp Inc., San Francisco, CA, USA).
Kuznetsov and Kordík [25] addressed the cold-start problem in recommendation systems with ontologies and knowledge graphs to enhance text-based methods. Their approach used ontologies to generate a knowledge graph capturing implicit and explicit characteristics of item–text attributes, enriching item profiles with semantically similar keywords. Their experimental evaluations demonstrated the efficacy of this method compared with state-of-the-art text feature extraction techniques. Recently, Li et al. [26] proposed a novel reinforcement learning approach for time series anomaly detection incorporating human feedback. This approach applied an ensemble of unsupervised anomaly scoring and devised reward strategies to guide the learning process, significantly outperforming five state-of-the-art models on the F1 score and the precision–recall metric of the area under the curve.
Moreover, numerous studies have examined regression problems regarding cold-start forecasting and anomaly detection. For example, Fatemi et al. [27] presented the cold causal demand forecasting framework, combining causal inference with deep learning models to enhance multivariate time series forecasting (TSF) in the context of the cold-start problem. Their study applied several critical techniques, including graph neural networks for representation learning based on causal influence, long short-term memory networks for capturing historical data, and similarity-based approaches using Gaussian mixture models and the extended Frobenius norm to leverage data from similar data centers. Xie et al. [28] designed a unified framework for long-range and cold-start forecasting of seasonal profiles in a time series. The framework combined high-dimensional regression and matrix factorization to address forecasting challenges posed by limited historical data, showing robust performance across multiple datasets.
Additionally, Ryu et al. [29] addressed the problem of cold start in web-service quality of service predictions using location-based matrix factorization with preference propagation, combining invocation similarity and neighborhood similarity, and applying location data for users and services. The method outperformed existing methods for cold-start and warm-start scenarios in their experiments. Xie et al. [18] introduced a unified framework to address long-range forecasts, missing data, and cold-start problems in time series data. The framework applied repeated patterns over fixed periods and employed metadata using low-rank decompositions, yielding accurate predictions and imputing missing values. Chen et al. [30] presented FrAug, a novel frequency domain data augmentation technique for improving TSF. The FrAug technique includes frequency masking and frequency mixing methods. It significantly enhanced forecasting accuracy and mitigated performance degradation under distribution shifts, making it effective for cold-start forecasting.
Key aspects of the proposed approach compared with referenced studies in related work are outlined below to clarify differences between them (see Table 1):
  • This study is focused on TSF, whereas most of the related studies have concentrated on recommendation systems or anomaly detection. The proposed approach addresses challenges of insufficient data by partitioning time series data and applying Bayesian inference to select the most probable model. R version 4.3.3 (R Foundation for Statistical Computing, Vienna, Austria) and RStudio version 2023.12.1.402 (Posit Software, PBC, Boston, MA, USA) were used as computational tools to ensure accurate and reproducible results. This approach contrasts with those of other studies, which often rely on augmenting the data with additional information or using hybrid models.
  • The proposed approach is tailored to numeric time series data, whereas related studies often involve categorical data in recommendation systems or mixed data types in anomaly detection. In addition, the proposed approach differs from previous studies in that Bayesian model selection is employed, which is distinct from typical methods used in related work, such as collaborative filtering, matrix factorization, and deep learning techniques.

3. Bayesian Time Series Analysis

Bayesian time series analysis provides a robust framework that integrates prior knowledge with observed data, thereby facilitating more reliable model selection and inference, particularly in situations where the precise DGP is uncertain. This method offers flexibility by incorporating uncertainty and past information, making it a valuable tool in a diverse range of time series applications.

3.1. Bayesian Model Selection

A general DGP is assumed as follows:
z t = f z t 1   + ε t ,   where   ε t N ( μ ^ ε , σ ^ ε 2 ) .
In most situations, when data are provided, the exact DGP is unknown. The DGP (t) can only be estimated using sample data derived from some DGP. The probability p[X|t] that X occurs can be calculated from the estimated DGP (t). However, in a cold-start problem, estimating this DGP or estimating p[X|zt] is infeasible.
Instead, if M DGPs could be tentatively considered ground truths, these might be represented by already-known equations or datasets. These M DGPs are called candidate models, and the mth candidate model is labeled Modelm. Under this setup, the following equation for calculating p[Modelm|X] (known as the posterior probability) indicates the probability that given data X are generated from Modelm:
p[Modelm|X] = p[Modelm, X]/p[X] =
(p[X|Modelm] ⋅ p[Modelm])/Σ (from i = 1 to M) (p[X|Modeli] ⋅ p[Modeli]),
where p[Modelm] (with m = 1, 2, …, M) represents the prior probability of each model, assigned uniformly assuming that no prior knowledge favors one model over the others.
In this study, Bayesian model selection is of paramount importance, as it integrates prior information (e.g., the uniform prior p[Modelm]) with observed data to calculate the posterior probability p[Modelm|X]. This approach allows for the selection of the most appropriate model among multiple candidates based on both prior beliefs and the likelihood of the observed data fitting each model. The posterior probability is calculated using Bayes’ theorem, whereby the initial model probabilities are updated based on how well each model explains the observed data.
The prior probability, p[Modelm], represents the initial assumption about the likelihood of each model before any data are considered. In the absence of prior knowledge that favors one model over another, a uniform prior is employed, giving each model equal weight initially. This is particularly advantageous in cold-start problems, where there is limited prior information available, as it allows for a more objective evaluation of the models.
The probability p[Modelm|X] indicates the likelihood of the observed data X occurring under each candidate model. The likelihood is calculated using statistical tests and distance metrics, such as the L2 norm, Pearson distance, or Wasserstein distance, depending on the characteristics of the dataset. These distance measures assist in quantifying the fit between the observed data and the predictions made by each model.
To further elucidate the selection of the most appropriate model, our criterion is primarily based on the analysis of results from various distance measurements under varying levels of statistical significance. This methodology entails a comparison of the capacity of each candidate model to reproduce the observed data, employing specific statistical tests and distance metrics, such as the L2 norm and others tailored to the characteristics of the dataset.
The contribution of each distance measurement is distinct, and significance thresholds are established to ascertain the probability that a given model represents the generative process responsible for the observed data. This approach ensures that the selected model not only exhibits a high degree of fit with the data but also adheres to predetermined significance levels, reflecting a robust integration of prior knowledge and statistical evidence.
In practice, Bayesian model selection is performed by calculating the posterior probability for each candidate model. The posterior probability reflects the probability that the observed data X were generated by a particular model, taking into account both the prior probability of the model and the likelihood of the data given the model. The model with the highest posterior probability is then selected as the optimal representation of the DGP. This process ensures thorough evaluation of models based on their ability to explain the observed data while integrating prior knowledge through the Bayesian framework.

3.2. Partitioned Time Series

As noted in Section 1, making predictions by predicting them is only feasible when the DGP can be easily estimated. However, candidate models often have missing data or structural changes in characteristics of influencing features, among other complexities, making it challenging to estimate the DGP. Although it would be ideal if the DGP could be assumed to remain constant, structural changes in a time series can occur at any time, requiring preparation for these changes.
When these changes occur, the “belief” regarding the model to which Y belongs must be updated. Re-estimating the DGP each time is highly inefficient. Therefore, the following approach is considered to estimate the marginal likelihood:
  • Observed data are denoted by X = {x1, x2, …, xu};
  • Given M candidate models, each is denoted by Modelm = Ym = {ym,1, ym,2, …, ym,nm}, where m = 1, 2, …, M, and nm is the length of Modelm.
Figure 5 illustrates the partitioning of data from each model. If the length of each partitioned data segment is u, equal to the length of X, then the maximum number of partitioned data vectors pm obtained from Modelm is nmu (qmnmu). Partitioned data vectors are pm = {pm,1, pm,2, …, pm,q_m}. If Modelm is a model for missing data, vectors shorter than u that occur immediately before or after the missing data are disregarded. If ym,1, …, ym,u are not missing data, then pm,1 = {ym,1, …, ym,u}. The relationship between each partitioned data vector pm,j (j = 1, …, qm) and X, denoted as r(pm,j, X), is then calculated.
Advantages of this approach include the following:
  • When new data arrive, all existing data are not needed, reducing the burden even when data continue to stream in real time;
  • This approach is advantageous for addressing structural breaks in model data. It decreases the need for complex analyses considering nonstationarity, reducing the burden of addressing long-term time series;
  • This approach simplifies handling missing data. Although this study employs a method that entirely excludes missing data, imputation is possible when necessary. Splitting data into shorter segments can offset irregularities, facilitating imputation;
  • Distributed computing is feasible. The process of determining the relationship between each partitioned data point pm and the observed data point X, denoted as r(pm,j, X), can be managed in parallel.

3.3. Analysis of Similarity between Observed and Partitioned Data

Section 3.2 discusses how to partition the data for each model. Partitioned data are compared with observed data. Sections below explore methods to make this comparison.

3.3.1. Measuring the Distance between Observed and Partitioned Data

Figure 6 presents how to determine differences between observed and partitioned data. However, using a random distance measure to consider the relationship between the two time series must be avoided. For example, if observed data X = {0.1, 0.4, 0.7, 1.0} and partitioned data from Modelm and Modelo are pm,q_i = {0.0, 0.4, 0.8, 1.2} and po,q_j = {1.3, 1.6, 1.9, 2.1}, respectively, increments of X and po,q_j are consistent at 0.3, whereas those of pm,q_i are at 0.4. This result could imply that X and po,q_j are generated from the same DGP.
However, when relationships were calculated using the L2 norm, values obtained with pm,q_i were lower, suggesting that they were more similar in distance to the observed data X, although they originated from different generative processes. This outcome illustrates a limitation of using straightforward distance measures such as the L2 norm to determine data origins. Thus, “similar values” and “values generated from the same process” must be identified. Using the L2 norm might imply the former, potentially missing a time series generated using different processes.
In contrast, if a particular data-generating function is known and used to generate data, various forms of a time series can be obtained from a single DGP, as depicted in Figure 7. In a statistical approach, current data can be considered a sample that accurately reflects characteristics of a specific population. Therefore, a direct comparison of distances might be inappropriate for verifying a time series generated under the same DGP. The L2 norm values can vary significantly depending on how the sampling was conducted.

3.3.2. Statistical Testing between Observed and Partitioned Data

Upon obtaining several partitioned data vectors qm from a model, their distribution can be assessed, and a joint distribution of these partitions can be considered (Figure 8). Each partitioned data vector only reflects characteristics of that segment, not necessarily the entire model. When partition sizes are small, this limitation can become significant. Similarly, initially observed data X may not fully reflect overall characteristics of the model, which is a fundamental reason for the failure to estimate the DGP directly from X alone.
When comparing distributions of observed data f(X) with those of qm partitioned data vectors g(pm,j), some might be similar while others might be different. If many g(pm,j) followed a specific distribution, with f(X) also following such a distribution, observed data X probably originated from this model. This scenario can be represented using the marginal likelihood p[Y|Modeli] for determining the probability that these data came from the ith candidate model. This approach does not assume structural changes. The high marginal likelihood indicates that many sections of the distribution are elevated.
Furthermore, using distribution testing (Figure 9) does not allow the discernment of trend differences. If distance measures are employed for comparison, observing how distances evolve over time can reveal trend differences. However, when employing distribution comparison tests, trend tests must be separately conducted. Nonetheless, if time series data are generated from a specific DGP, statistical independence tests can be used for analyzing data distribution before estimating the DGP. This method also accounts for distances between distributions. However, data samples are from a joint distribution. Hence, statistical testing is appropriate.

3.4. Analytical Procedures

Observed data X and M candidate models are assumed to be known beforehand with a cold-start problem. The following steps outline the proposed approach:
  • Data for each candidate model are partitioned. This study followed the method described in Section 3.2 to partition data from each candidate model (designated as pm). If M candidate models exist, M partitioned datasets pm are also generated. If the length of observed data is u, each partitioned dataset is also set to length u. If data are standardized, standardization is applied separately to each pm;
  • Users must determine whether to use a distance measure or a statistical test to calculate r(pm,j, X). If distance is chosen, r(pm,j, X) is determined by d(pm,j, X), measuring direct discrepancies between datasets. If a statistical test is selected, r(pm,j, X) corresponds to a p-value that evaluates the independence between pm,j and X, indicating the likelihood of a statistical relationship;
  • When a statistical test is employed to determine r(pm,j, X), trend differences between pm,j and X are not assessable. Therefore, test results for the trend must also be considered. Trend tests for pm,j and X are conducted to obtain p-values. Closer trends of pm,j and X indicate a smaller difference between p-values from trend tests of pm,j and X, which results in a value that approaches zero;
  • When calculating marginal likelihood, because qm instances of pm,j exist, qm instances of r(pm,j, X) are also attained. By setting a significance level for r(pm,j, X), the probability p[X|Modelm] can be calculated, indicating the likelihood of X belonging to Modelm. Probability calculations regarding significance levels are detailed in Equations (3) and (4):
    p[XModelm] = p[r(pm,j, X) ≤ SignIf.indep.],
    p[XModelm] = p[(p-valueindep.(pm,j, X) ≤ SignIf.indep.)/(p-valuetrend.(pm,j) ≤ SignIf.trend)];
  • Posterior probability is calculated using the method described in Equation (2). The prior probability is uniformly assigned across all M candidate models. Comparing posterior probabilities across all M candidate models could identify the model that most likely generated observed data X, enhancing our understanding of the reliability of the model.
The process of evaluating which candidate model best explains observed data is streamlined by structuring the analysis in this way, considering statistical relationships and trends between datasets. This approach guarantees robust decision-making by applying a comprehensive statistical evidence base.

4. Results

4.1. Statistical Testing or Distance

In this experiment, methods employed to calculate r(pm,j, X) include statistical tests, specifically the Kolmogorov–Smirnov test and runs test for two samples. The Cox–Stuart trend test was applied for trend testing. Although these tests employ historical data regarding distribution and trends, the Bayesian model utilized in this study also incorporates prior knowledge, thereby enhancing the analytical process and providing a more robust foundation for the findings. Table 2 lists the distances considered for calculating r(pm,j, X). The prior information employed in the Bayesian framework is derived from assumptions based on past observations and pertinent domain knowledge, which facilitates the model’s inference process. This prior knowledge was selected based on the distinctive characteristics of the synthetic data presented in Table 3, including stability, variance, and trends. Its suitability in the context of the time series behaviors examined in the study was then empirically validated.
The reason for extensively examining distances, more so than statistical tests, was that the methods for measuring distances between two time series varied significantly more than those for the statistical testing of their distributions. However, the primary motivation was to verify the caution advised in Section 3.3.1 empirically against randomly using distance measures when comparing two time series. The incorporation of prior knowledge into the Bayesian model facilitated a more structured approach to distance measure selection, thereby reducing the probability of random or inappropriate choices. This contributed to the development of a more robust model, as evidenced by the results.
The synthetic models presented in Table 3 provide a basis for evaluating the distance measures in question. Table 3 outlines four categories of synthetic data generated to represent diverse time series behaviors, including stationary time series, unstable variance, trend changes, and the presence of a unit root. The aforementioned synthetic models permit an evaluation of the efficacy of distance functions and statistical tests in identifying the true model under a variety of conditions. Prior knowledge was incorporated into the Bayesian model in a category-specific manner, which helped to improve model accuracy by providing additional context about the expected behavior of the data.
By establishing a connection between the synthetic models presented in Table 3 and the distance functions outlined in Table 2, we ensure a comprehensive examination of the performance of each method across diverse types of time series data. This methodology allows us to elucidate the strengths and limitations of distance measures and statistical tests in various contexts, which are further elaborated in Section 4.2. Furthermore, the incorporation of prior information through the Bayesian framework not only enhanced the interpretability of the results but also demonstrated the potential advantages of integrating prior knowledge into time series analysis.

4.2. Synthetic Model

For objective experimentation, the following synthetic autoregressive model of length 1000 was considered:
ym,t = α × yt−1 + t,
where ym,1 = 0, tN(μ, σ2), and t = 1, …, 1000 (where m was the index of the candidate model, and α, μ, and σ denoted parameters that could be adjusted to control the stationarity of the synthetic model).
  • Stationary time series: The model is considered stable when σ2 is less than 1. Although diversifying a stable model is challenging, variations can be introduced in the range, where 0 < σ2 < 1, μ = 0, and 0 < α <1;
  • Unstable variance: The σ value should be set to ≥1 because a larger σ2 yields more dynamic movement in the data;
  • Trend changes: The μ value can be applied to determine a stochastic trend. No trend exists when μ = 0. If μ > 0, an upward trend exists. If μ < 0, a downward trend exists. A larger absolute value of μ indicates a steeper trend slope;
  • Presence of a unit root: The α term is the coefficient. For a nonstationary model, α should be set to ≥1. If α = 1, a unit root is present.
As presented in Table 3, four types of synthetic data with set parameters were created to analyze the effects of varying these parameters. Observed data X were matched with the type for each model.
The significance level for r(pm,j, X) was determined using quantiles. However, users can make interactive selections. Settings were adjusted to 0.1, 0.3, 0.5, 0.7, and 0.9 to observe changes. If the method for calculating r(pm,j, X) operates correctly, whether using distance or statistical testing, a lower significance level is expected to increase the likelihood of selecting the correct model.
Table 4, Table 5, Table 6 and Table 7 present a comprehensive overview of the performance of various distance measures and statistical tests across a range of experimental conditions, including stationary time series, dynamic variance, trend changes, and the presence of a unit root. The tables provide detailed insights into the relative merits and limitations of each method in identifying the true models under these diverse scenarios. Table 4 is devoted to an examination of the effectiveness of the methods in the context of a stationary time series. Table 5, Table 6 and Table 7 are dedicated to an evaluation of the methods under more intricate circumstances, such as unstable variance, trend alterations, and unit roots.
To evaluate the effectiveness of distance metrics and statistical tests employed to calculate r(pm,j, X), it is essential that observed data and parameters exhibit a high posterior probability for the same model. A high posterior probability for the same model ensures that the model aligns well with the data, thereby providing a robust framework for identifying the true model even under challenging conditions. In an ideal scenario, even when a high significance level is set, the true model should be consistently identified. In experiments concerning candidate models for stationary time series, as presented in Table 4, setting the significance level at the 10% quantile revealed that 13 out of 29 metrics, including the L2 norm, Manhattan, and Chebyshev distances, along with Kolmogorov–Smirnov and runs tests, successfully identified the true model. The results presented herein demonstrate the efficacy of these metrics in stable time series data, particularly under lower significance levels, where model identification is more precise.
However, as the significance level increased, the majority of methods were unable to identify the true model. The challenge of identifying the optimal model at progressively higher levels of statistical significance underscores the inherent trade-offs involved in selecting an appropriate level of significance, particularly in the context of more complex datasets. From the perspective of a practitioner, the absence of a definitive criterion for setting the appropriate significance level presents a significant challenge. If the significance level is set too low, issues of reproducibility may arise. Conversely, setting it too high could lead to results that are difficult to interpret as meaningful, potentially undermining the novelty of findings. Notwithstanding the aforementioned challenges, the outcomes illustrated in Table 4 indicate that distance metrics such as the Chebyshev distance, the Wasserstein distance, and the Kolmogorov–Smirnov test exhibited resilience across varying levels of statistical significance, consistently identifying the authentic model in stationary time series scenarios. This resilience suggests that these metrics are robust and reliable in stable conditions.
In the examination of candidate models exhibiting unstable variance, specifically where the variance parameter σ2 > 1, the majority of methods failed to identify the true model even when the significance level was reduced to 0.1. As illustrated in Table 5, the findings reveal that distance-based measures, such as the Chebyshev and Wasserstein distances, demonstrated superior performance compared to traditional statistical tests in scenarios characterized by high variance. These findings highlight the significant advantage of distance-based techniques in addressing fluctuations in variance, rendering them highly effective for data exhibiting dynamic variance. Despite these challenges, certain methodologies, notably the Chebyshev distance and the Kolmogorov–Smirnov test, demonstrated robust performance under these conditions.
In scenarios where both observed data and candidate models exhibit trends, it is of paramount importance to identify models that not only match the trend direction but also its slope. The complexity of trend detection represents a significant challenge in identifying the most appropriate model, as it is not a straightforward process. For example, candidate models A, B, and C were generated with increasing trends, characterized by parameters µ of 0.5, 1.0, and 2.0, respectively. In contrast, models D, E, and F were generated with decreasing trends, showing parameters µ of −0.5, −1.0, and −2.0, respectively. As illustrated in Table 6, while numerous distance measures, including the L2 norm, Manhattan, and Chebyshev distances, were effective in identifying the trend direction, Pearson distance was the sole method that correctly identified both the trend direction and slope. This underscores Pearson distance’s distinctive capability in detecting subtle variations in trend patterns. These findings are presented in Table 6, emphasizing the necessity of additional processes, such as detrending, when working with data that exhibit trends.
In experiments designed to assess the presence of a unit root, Pearson distance, Wasserstein distance, the Kolmogorov–Smirnov test, and the runs test demonstrated relatively effective performances in correctly selecting the true model. As demonstrated by the findings in Table 7, the Wasserstein and Pearson distances exhibit superior performance in identifying unit roots, a task that frequently presents a challenge for other methods. A comprehensive analysis revealed that, regardless of parameter configurations, the Kolmogorov–Smirnov test consistently identified the true model when calculating r(pm,j, X).
Moreover, this study demonstrates that the Wasserstein distance is particularly useful when considering the presence of a unit root. Moreover, the findings indicate that the Wasserstein distance, with its capacity to process nonstationary data, is a valuable tool for practitioners. Additionally, in the event that the data exhibit dynamic variance, it is recommended to consider both Chebyshev and Wasserstein distances, along with Pearson distance, when accounting for trends. The comprehensive results demonstrate that the application of multiple tailored approaches, based on the specific characteristics of the data, can markedly enhance model accuracy and provide reliable insights. Given that each distance measure possesses distinctive strengths under different conditions, the utilization of a combination of these methods allows for a more nuanced and comprehensive analysis of time series data.

4.3. Practical Applications in the Energy Sector

The European Network of Transmission System Operators for Electricity (ENTSO-E) plays a pivotal role in the implementation of energy policies across the European Union [32]. It contributes to the achievement of Europe’s energy and climate goals by aggregating and disseminating critical energy data, including monthly power consumption, production, and annual net generating capacities, through its online platform. In the context of this research, we conducted an analysis of hourly net generating capacity data from 1 January 2006 to 31 December 2015. The subset of data pertaining to Austria (AT), specifically from t = 3000 to t = 3999, was treated as observed data in this study. The dataset is systematically cataloged in Table 8, which provides a detailed view of power generation dynamics across 35 European nations.
As demonstrated in Figure 9, the power data exhibited strong seasonal variations along with pronounced trend changes. In the short term, these points of trend shifts might also indicate structural changes. However, the presence of a unit root cannot be ascertained through visual means alone. Unlike synthetic data, the actual dataset often contains numerous instances where data collection has not occurred, resulting in significant gaps or missing values. As previously discussed in Section 3.2, our approach to handling missing data involved the exclusion of segments containing gaps during the dataset partitioning process. This method simplifies the handling of missing values while simultaneously avoiding an increase in uncertainty that typically accompanies imputation techniques. However, one disadvantage of this approach is the difficulty in obtaining reliable estimates of r(pm,j, X) for segments where missing values are concentrated.
The analysis, as documented in Table 9 and Table 10, employed the Wasserstein distance, the Kolmogorov–Smirnov test, and the runs test to assess r(pm,j, X). These methodologies successfully identified the true model, demonstrating robustness in both synthetic and real data scenarios. This is further elaborated in Section 4.2. These methods were demonstrated to be particularly effective for stationary time series within synthetic models. The consistent performance observed in the ENTSO-E dataset was likely attributable to the inherent periodic nature of power data, which remained stable over extended periods. Furthermore, distinctive features of each candidate model in real data scenarios enhanced the clarity of comparative results, as evidenced by detailed evaluations presented in Table 9 and Table 10.

4.4. Discussion

The objective of this study was to assess the efficacy of the L2 norm, a commonly utilized metric in time series analysis, in differentiating between models exhibiting variation solely in their parameter values. Table 9 and Table 10 present a comprehensive comparison of alternative distance metrics in both synthetic and real-world energy data scenarios, emphasizing their reliability. However, the results indicate that while certain methods, such as the Wasserstein distance and the Kolmogorov–Smirnov test, demonstrated effectiveness in specific contexts, the reliability of the L2 norm as a metric for model comparison was called into question.
The experiment yielded stationary time series data across a range of autoregressive (AR), moving average (MA), and ARMA models with varying parameters but comparable structures. The utilization of stationary time series data facilitated a direct evaluation of the efficacy of the L2 norm in differentiating between models exhibiting variation solely in their parameter values. In each instance, models were generated with specific parameter sets, and the L2 norm of the discrepancies between the observed and split datasets was calculated. It was postulated that the L2 norm would consistently identify the model that generated the observed data, particularly when m = 3.
In this experiment, the following types of data were generated: A total of 30 models were generated by setting five parameters for each of the six types. The only varying factor was variance.
  • Type A: AR(1), ym,t = φ0 + φm,1yt−1 + εt;
  • Type B: AR(2), ym,t = φ0 + φm,1yt−1 + φm,2yt−2 + εt;
  • Type C: MA(1), ym,t = θ0 + θm,1yt−1 + εt;
  • Type D: MA(2), ym,t = θ0 + θm,1yt−1 + θ0 + θm,2yt−2 + εt;
  • Type E: ARMA(1), ym,t = φ0 + φm,1yt−1 + θm,1yt−1 + εt;
  • Type F: ARMA(2), ym,t = φ0 + φm,1yt−1 + φ0 + φm,2yt−2 + θm,1yt−1 + θm,2yt−2 + εt
    where m = 1, …, 5, εt ~ N(0, 12), t = 1, …, 3600;
  • φ1,1 = 0.1, φ2,1 = 0.3, φ3,1 = 0.5, φ4,1 = 0.7, φ5,1 = 0.9;
  • φ1,2 = −0.1, φ2,2 = −0.3, φ3,2 = −0.5, φ4,2 = −0.7, φ5,2 = −0.9;
  • θ1,1 = 0.1, θ2,1 = 0.3, θ3,1 = 0.5, θ4,1 = 0.7, θ5,1 = 0.9;
  • θ1,2 = −0.1, θ2,2 = −0.3, θ3,2 = −0.5, θ4,2 = −0.7, θ5,2 = −0.9.
For each type, data of length 100 were generated from the model with m = 3, and t = 1, …, 100 was considered as the observed data.
yo,t_o = {ym,t_o}, to = 1, …, 100.
Then, split data of length 100 were generated from the data corresponding to t = 101, …, 3600. In this way, 3400 split datasets were generated for each model. The L2 norm of the difference between yo,t_o and each split dataset was then computed. Figure 10 exhibits the process of generating split datasets. The red vertical line separates the observed data yo,t_o on the left from the portion of the data used to create the split datasets on the right.
After discussing the challenges of using the L2 norm as a distance measure, we provide further empirical results in Table 11. This table compares the L2 norms obtained for different models under different parameter settings and clearly shows how increasing the value of m affects the magnitude of the L2 norm. The key observation here is that even when m = 3, for which the original data were generated, the L2 norm did not consistently show the lowest values as initially expected. This finding suggests that the L2 norm may not be an appropriate metric for model comparison in this context, even when the experiment is repeated 100 times.
In addition, Figure 11 provides a graphical representation of the L2 norm distributions for each model type. The histograms show the frequency of L2 norms across 3400 generated datasets, illustrating how the norms vary across different parameter settings. The sharp peaks indicate that for certain models, there is a higher concentration of L2 norms near the lower range, while for others, the distribution is more spread out. This variability in the distribution highlights the inconsistency of using the L2 norm for model selection and reinforces the conclusion that alternative methods may be necessary, especially when trends or other nonstationary factors are present.
While these results highlight the limitations of the L2 norm, alternative distance metrics such as those presented in Table 9 and Table 10, such as the Wasserstein distance, the Kol-Mogorov-Smirnov test, and various statistical measures, showed more consistent performance in identifying the true model. These alternative methods provided better robustness, especially in real-world datasets with inherent complexities such as missing data and nonstationarity. This underscores the importance of using multiple metrics for model comparison rather than relying solely on the L2 norm.
The histograms in Figure 11 show how the L2 norms are distributed for each model, providing further insight into the suitability of this metric for different types of time series. The visual representation complements the numerical data in Table 11 and further supports the argument that while the L2 norm can provide some information, its overall effectiveness as a model selection criterion remains questionable.
To address the limitations of the preceding experiment, which may have involved an insufficient number of parameter combinations, an expanded experiment was conducted. The objective was to further investigate the inconsistencies observed in the L2 norm’s performance by considering a wider range of parameter combinations. The goal was to more thoroughly evaluate the L2 norm’s reliability as a model comparison tool. The following model configuration was considered:
yt = φ0 + φ1yt−1 + φ2yt−2 + θ1εt−1 + θ2εt−2 + εt,
where t = 1, …, 3600 and θ0 = 100. The remaining parameters, φ1, φ2, θ1, θ2, and the standard deviation σ of εt ~ N(0, σ2), are subject to the following variations:
  • p1 = φ1 = {0, 0.1, 0.5, 0.9};
  • p2 = φ2 = {0, −0.1, −0.5, −0.9};
  • q1 = θ1 = {0, 0.1, 0.5, 0.9};
  • q2 = θ2 = {0, −0.1, −0.5, −0.9};
  • σ = {0.1, 0.5, 1, 2, 3}.
The aforementioned combinations result in 1280 distinct models, which are designated as Modelp_1,p_2,q_1,q_2,σ. A total of 3500 data subsets of length 100 were generated for each model. Moreover, a dataset of length 100 was generated from Model0.5,0,0.5,0,1 and employed as the observed data. The model equation is as follows:
yo,t_o = φ0 + φ1yt−1 + θ1εt−1 + εt = 100 + 0.5yt−1 + 0.5εt−1 + εt, to = 1, …, 100, εt ~ N(0, 1).
Following this, the L2 norm of the differences between yo,t_o and each generated subset of data was computed. Since yo,t_o was derived from Model0.5,0,0.5,0,1, it was expected that the L2 norm would be lowest when comparing against this same model. However, the experimental results shown in Table 12 reveal that the L2 norm did not consistently identify the correct model.
Despite comparing stationary time series that only differ in parameter values, the L2 norm-based testing showed that the rank of the model generating the observed data, Model0.5,0,0.5,0,1, appeared in the late 500s out of the 1280 models. This suggests that using the L2 norm alone to compute the Bayes factor could result in a stronger belief in 500 other models rather than the original model that generated the observed data.
As illustrated in Table 13, the issue of ranking is further emphasized when different summary statistics are considered. While the model responsible for generating the observed data exhibits a high ranking in terms of minimum and maximum values, it performs poorly in other key statistics, such as the first quartile, median, and mean. This variability across summary statistics indicates that the L2 norm may not be a sufficient criterion for robust model comparison, particularly when dealing with more complex or real-world data scenarios.
In scenarios that are more complex, such as those in which nonstationary elements like trends are introduced, the limitations of traditional distance metrics like the L2 norm become even more pronounced. This underscores the necessity for sophisticated methodologies, such as graph neural networks (GNNs) and transfer learning. GNNs are particularly well suited to the analysis of graph-structured data, rendering them an excellent choice for complex time series such as energy networks, where interdependencies exist across multiple time points or nodes. However, the application of GNNs or transfer learning in this study presented a number of challenges, including the necessity for extensive labeled datasets and the complexity of model tuning.
In conclusion, this study demonstrates the shortcomings of relying exclusively on conventional distance metrics, such as the L2 norm, for model comparison in time series analysis. By integrating advanced techniques, such as GNNs and transfer learning, with more reliable statistical measures, future research can overcome these limitations. This hybrid approach has the potential to enhance the precision and computational efficiency of time series models, particularly in domains where accurate predictions are crucial for decision-making, including energy management and beyond.

5. Visualization

While the process of selecting a candidate model by computing the posterior probability from observed data is crucial, as depicted in Figure 4, it is equally important to determine “at which point” the observed data resemble the candidate model. However, this comparison can be challenging to express numerically. It may not be easily understood by the user if presented in such a format. Figure 12 and Figure 13 provide a detailed visualization framework to help illustrate these concepts. Figure 12 shows the components necessary for dataset selection and model analysis, while Figure 13 focuses on the visualization of marginal likelihood. The simplest and most effective way to convey this information to users is through these visualizations, which enhance the user’s ability to understand complex statistical relationships clearly and intuitively.
In particular, when employing distance to compute p[X|Modelm], as illustrated in Equation (3), it is essential to initially ascertain whether r(pm,j, X) ≤ SignIf.indep. with the candidate model. This can be accomplished by graphically representing the candidate model as a line graph and then highlighting sections that meet this criterion in a specific color, as depicted in Figure 13 “A”. If the method to compute p[X|Modelm] involves statistical tests rather than distance, as indicated in Equation (4), it is also necessary to determine whether p-valuetrend(pm,j) ≤ SignIf.trend. These sections can be independently highlighted, as shown in Figure 13 “B”. Consequently, by visualizing both [indep.(pm,j, X) ≤ SignIf.indep.] and p-valuetrend(pm,j) ≤ SignIf.trend, as shown in Figure 13 “A∩B”), it becomes easier for users to identify sections of the data that are considered similar to the observed data.
The determination of an appropriate threshold for r(pm,j, X) to judge similarity is a highly challenging endeavor due to variability observed across different data states and domains. In particular, when employing statistical tests, the conventional approach has been to reject the null hypothesis when p-values are below traditional thresholds of 0.1, 0.05, or 0.01. Nevertheless, these standards are currently the subject of controversy regarding reproducibility due to an increasing number of cases of misuse involving the manipulation of crucial parameters such as standardization, repetition of experiments, and randomization methods. In the medical field, which relies heavily on statistical analysis, there has been a recent consensus to lower the threshold for rejecting the null hypothesis to 0.001 to address these controversies. However, it is still challenging to assert that this is a comprehensive solution.
Regardless of the absolute value of r(pm,j, X), an analytical method can be considered reproducible if results from analyzing a newly generated independent dataset in the same manner are consistent with those from previous similar analyses. It is therefore desirable for the analysis to yield reproducible results. To facilitate user adaptation to their specific data states or domains within a visualization framework, it is essential to allow users to adjust the significance level directly.
In the course of conducting experiments for this paper, the value of r(pm,j, X) was calculated for a number of partitioned datasets. Results were then examined by adjusting the significance level based on their quantiles. In particular, in Section 4.2, when experimenting with a synthetic model, results for five different significance levels were summarized in a large table. However, rather than displaying such an extensive table, a slide bar was used to interactively adjust the significance level, as shown in Figure 14. This method is an efficient and clear method for demonstrating changes in results.
Once a candidate model with a high likelihood of encompassing the observed data is selected, the subsequent step is to proceed with a prediction. Nevertheless, selecting a candidate model does not necessarily resolve the inherent cold-start problem in predictions using observed data. Nevertheless, it is possible to directly compare the observed data with a partitioned data vector deemed similar. By displaying the observed data and the partitioned data vector on a line graph as illustrated in Figure 15 and simultaneously reviewing the graph, distance, or p-values if a statistical test was performed, users can conduct a personal assessment of the similarity of the data in question and the extent to which this similarity can be confirmed.

6. Conclusions

This study effectively addressed key research questions posed at the outset, providing valuable insights into time series analysis in cold-start scenarios with limited initial data. Through the application of Bayesian model selection, we were able to not only predict responses with higher accuracy than those achieved with traditional distance measures like the L2 norm but also highlight the inadequacies of traditional distance measures, such as the L2 norm, in model comparison.
  • Our findings demonstrate that Bayesian model selection can significantly enhance predictive accuracy when faced with sparse data. By partitioning models and analyzing each vector with statistical tests, we bypassed the traditional reliance on distance measures. This approach was proven to be particularly beneficial in scenarios where the conventional methods would likely fail due to insufficient data.
  • Another significant contribution of this study is the development of a new visualization technique that employs a slide bar for interactively setting significance levels. This method stands in stark contrast to traditional star-marked displays. It offers a dynamic tool for researchers to adjust and interpret significance with greater clarity, thus reducing potential misunderstandings about p-value implications.
  • The operational aspect of employing Bayesian model selection in real-world scenarios was also explored. We found that once the observational data aligned with a candidate model, effective predictions could be made using the model. However, this is contingent upon assumptions of stationarity and the absence of structural breaks, which our research identified as areas requiring further investigation to fully harness the potential of using Bayesian methods for solving cold-start problems.
By systematically addressing these research questions, our study not only advances the field of time series analysis but also equips practitioners with more robust tools for handling the complexities of cold-start scenarios. The insights gained underscore the importance of innovative statistical approaches and the need for the continual refinement of analytical tools to adapt to evolving data challenges.
Despite advances presented in this study, certain limitations must be acknowledged, along with potential directions for future research that could further refine and expand our findings. The effectiveness of Bayesian model selection, as demonstrated in this study, heavily relies on the quality and granularity of the observational data. In scenarios where data are excessively sparse or noisy, the reliability of model alignment and subsequent predictions could be compromised. Our approach assumes stationarity and the absence of structural breaks within the dataset. These assumptions may not hold true in all practical applications, potentially affecting the robustness of predictions. Implementation of the proposed visualization tool and statistical tests requires a certain level of statistical and computational expertise, which may not be readily available in all research or applied settings.
Future studies could explore the application of Bayesian model selection across more diverse datasets, including those with higher levels of noise and nonstationarity. This could help us understand the limits and scalability of our approach. Developing algorithms that can automatically detect and adjust for structural breaks and nonstationarity within the data could significantly enhance the applicability and accuracy of Bayesian model selection in real-world scenarios. There is a need to develop more intuitive and accessible tools that can bring the power of advanced statistical methods to a wider audience. Simplifying the implementation of our visualization technique could facilitate its adoption in nonspecialist contexts, enhancing its practical utility. By addressing these limitations and exploring these avenues for future research, the field can move towards more generalized and robust methods for dealing with cold-start problems and time series analysis under challenging conditions.

Author Contributions

Conceptualization, J.Y.; methodology, J.Y.; software, J.Y. and J.M.; validation, J.Y. and J.M.; formal analysis, J.Y.; investigation, J.Y.; resources, J.M.; data curation, J.M.; writing—original draft preparation, J.Y.; writing—review and editing, J.M.; visualization, J.Y. and J.M.; supervision, J.M.; project administration, J.M.; funding acquisition, J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the MSIT (Ministry of Science, ICT), Korea, under the National Program for Excellence in SW, supervised by the IITP (Institute of Information and communications Technology Planning and Evaluation) in 2021 (2021-0-01399) and the Soonchunhyang University Research Fund.

Data Availability Statement

This study primarily used data from the ENTSO-E Transparency Platform, essential for our analysis. These data can be accessed at https://transparency.entsoe.eu (accessed on 6 June 2024). Additionally, certain datasets were developed specifically for this research. They are detailed in the methods section of the paper. Full details on the data used can be found within the manuscript, ensuring transparency and possibility of replication by others.

Acknowledgments

We would like to sincerely thank the editor for expertly guiding the review process and the two anonymous reviewers for their valuable feedback and thoughtful suggestions.

Conflicts of Interest

Author Jaeseong Yoo was employed by the company Statistical Ground. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The Statistical Ground company had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Stachl, C.; Pargent, F.; Hilbert, S.; Harari, G.M.; Schoedel, R.; Vaid, S.; Bühner, M. Personality research and assessment in the era of machine learning. Eur. J. Personal. 2020, 34, 613–631. [Google Scholar] [CrossRef]
  2. Abdelrahman, G.; Wang, Q.; Nunes, B. Knowledge tracing: A survey. ACM Comput. Surv. 2023, 55, 224. [Google Scholar] [CrossRef]
  3. Herlocker, J.L.; Konstan, J.A.; Terveen, L.G.; Riedl, J.T. Evaluating collaborative filtering recommender systems. ACM Trans. Inf. Syst. 2004, 22, 5–53. [Google Scholar] [CrossRef]
  4. Hasselmann, K.; Sausen, R.; Maier-Reimer, E.; Voss, R. On the cold start problem in transient simulations with coupled atmosphere-ocean models. Clim. Dyn. 1993, 9, 53–61. [Google Scholar] [CrossRef]
  5. Liu, H.; Ong, Y.S.; Shen, X.; Cai, J. When Gaussian process meets big data: A review of scalable GPs. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 4405–4423. [Google Scholar] [CrossRef] [PubMed]
  6. Hanke, J.E.; Wichern, D. Business Forecasting, 9th ed.; Pearson: Upper Saddle River, NJ, USA, 2009. [Google Scholar]
  7. Thomas, L. Retrospective power analysis. Conserv. Biol. 1997, 11, 276–280. [Google Scholar] [CrossRef]
  8. Legg, C.; Nagy, L. Why most conservation monitoring is, but need not be, a waste of time. J. Environ. Manag. 2006, 78, 194–199. [Google Scholar] [CrossRef] [PubMed]
  9. Hansen, B.E. The new econometrics of structural change: Dating breaks in U.S. labor productivity. J. Econ. Perspect. 2001, 15, 117–128. [Google Scholar] [CrossRef]
  10. Chow, G.C.; Lin, A.L. Best linear unbiased interpolation, distribution, and extrapolation of time series by related series. Rev. Econ. Stat. 1971, 53, 372–375. [Google Scholar] [CrossRef]
  11. Denton, F.T. Adjustment of monthly or quarterly series to annual totals: An approach based on quadratic minimization. J. Am. Stat. Assoc. 1971, 66, 99–102. [Google Scholar] [CrossRef]
  12. Box, G.E.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  13. Zhang, G.P.; Qi, M. Neural network forecasting for seasonal and trend time series. Eur. J. Oper. Res. 2005, 160, 501–514. [Google Scholar] [CrossRef]
  14. Puma-Villanueva, W.J.; dos Santos, E.P.; Von Zuben, F.J. Data Partition and Variable Selection for Time Series Prediction Using Wrappers. In Proceedings of the 2006 IEEE International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 16–21 July 2006; pp. 4740–4747. [Google Scholar]
  15. Puma-Villanueva, W.J.; dos Santos, E.P.; Von Zuben, F.J. Long-Term Time Series Prediction Using Wrappers for Variable Selection and Clustering for Data Partition. In Proceedings of the 2007 IEEE International Joint Conference on Neural Networks (IJCNN), Orlando, FL, USA, 12–17 August 2007; pp. 3068–3073. [Google Scholar]
  16. Sarkar, S.; Mukherjee, K.; Jin, X.; Ray, A. Optimization of Time-Series Data Partitioning for Parameter Identification in Nonlinear Dynamical Systems. In Proceedings of the ASME 2010 Dynamic Systems and Control Conference (DSCC2010), Cambridge, MA, USA, 12–15 September 2010; Volume 1, pp. 867–874. [Google Scholar]
  17. Leverger, C.; Lemaire, V.; Malinowski, S.; Guyet, T.; Roze, L. Day-Ahead Time Series Forecasting: Application to Capacity Planning. In Proceedings of the 3rd Workshop on Advanced Analytics and Learning of Temporal Data (AALTD), Dublin, Ireland, 10–14 September 2018. [Google Scholar]
  18. Xie, C.; Tank, A.; Fox, E. A Unified Framework for Missing Data and Cold Start Prediction for Time Series Data. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  19. Xu, X.; Yang, C.; Yu, Q.; Fang, Z.; Wang, J.; Fan, C.; He, Y.; Peng, C.; Lin, Z.; Shao, J. Alleviating Cold-Start Problem in CTR Prediction with a Variational Embedding Learning Framework. In Proceedings of the ACM Web Conference 2022, Virtual, 25–29 April 2022; pp. 27–35. [Google Scholar]
  20. AlRossais, N.; Kudenko, D.; Yuan, T. Improving cold-start recommendations using item-based stereotypes. User Model. User-Adap. Interact. 2021, 31, 867–905. [Google Scholar] [CrossRef]
  21. Pirasteh, P.; Bouguelia, M.R.; Santosh, K.C. Personalized recommendation: An enhanced hybrid collaborative filtering. Adv. Comput. Intell. 2021, 1, 1. [Google Scholar] [CrossRef]
  22. Rohani, V.A.; Kasirun, Z.M.; Kumar, S.; Shamshirband, S. An Effective Recommender Algorithm for Cold-Start Problem in Academic Social Networks. Math. Probl. Eng. 2014, 2014, 123726. [Google Scholar] [CrossRef]
  23. Ni, L.; Wang, S.; Zhang, Z.; Li, X.; Zheng, X.; Denny, P.; Liu, J. Enhancing Student Performance Prediction on Learner-Sourced Questions with Sgnn-Llm Synergy. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–12 March 2024; Volume 38, pp. 23232–23240. [Google Scholar]
  24. Tey, F.J.; Wu, T.Y.; Lin, C.L.; Chen, J.L. Accuracy improvements for cold-start recommendation problem using indirect relations in social networks. J. Big Data 2021, 8, 98. [Google Scholar] [CrossRef]
  25. Kuznetsov, S.; Kordík, P. Overcoming the Cold-Start Problem in Recommendation Systems with Ontologies and Knowledge Graphs. In Proceedings of the European Conference on Advances in Databases and Information Systems, Vienna, Austria, 5–8 August 2023; Springer: Cham, Switzerland, 2023; pp. 591–603. [Google Scholar]
  26. Li, H.; Xu, H.; Peng, W. Deep Reinforced Active Learning for Time Series Anomaly Detection. In Proceedings of the International Conference on Intelligent Computing, Singapore, 3–6 July 2023; Springer: Singapore, 2023; pp. 115–128. [Google Scholar]
  27. Fatemi, Z.; Huynh, M.; Zheleva, E.; Syed, Z.; Di, X. Mitigating Cold-Start Problem Using Cold Causal Demand Forecasting Model. arXiv 2023, arXiv:2306.09261. [Google Scholar]
  28. Xie, C.; Tank, A.; Greaves-Tunnell, A.; Fox, E. A Unified Framework for Long Range and Cold Start Forecasting of Seasonal Profiles in Time Series. arXiv 2017, arXiv:1710.08473. [Google Scholar]
  29. Ryu, D.; Lee, K.; Baik, J. Location-based web service QoS prediction via preference propagation to address cold start problem. IEEE Trans. Serv. Comput. 2018, 14, 736–746. [Google Scholar] [CrossRef]
  30. Chen, M.; Xu, Z.; Zeng, A.; Xu, Q. Fraug: Frequency Domain Augmentation for Time Series Forecasting. arXiv 2023, arXiv:2302.09292. [Google Scholar]
  31. Ebrahimi, A.; Ghobaei-Arani, M.; Saboohi, H. Cold start latency mitigation mechanisms in serverless computing: Taxonomy, review, and future directions. J. Syst. Archit. 2024, 151, 103115. [Google Scholar] [CrossRef]
  32. European Network of Transmission System Operators for Electricity (ENTSO-E). ENTSO-E Transparency Platform. Available online: https://transparency.entsoe.eu (accessed on 1 July 2024).
Figure 3. Analysis and prediction of time series data. (a) Observed time series data example (initial observed data points in black) and (b) predictions utilizing data-generating process (initial data points in black extended with predictions in red).
Figure 3. Analysis and prediction of time series data. (a) Observed time series data example (initial observed data points in black) and (b) predictions utilizing data-generating process (initial data points in black extended with predictions in red).
Mathematics 12 02682 g003
Figure 4. Probability of observed time series data X belonging to different candidate models.
Figure 4. Probability of observed time series data X belonging to different candidate models.
Mathematics 12 02682 g004
Figure 5. Generation of partitioned data. This diagram shows how data vectors pm are derived from Modelm, with arrows illustrating the process of partitioning data equal to the length of X. The arrows indicate the association between each partitioned vector pm,j and X, emphasizing the calculation of r(pm,j, X).
Figure 5. Generation of partitioned data. This diagram shows how data vectors pm are derived from Modelm, with arrows illustrating the process of partitioning data equal to the length of X. The arrows indicate the association between each partitioned vector pm,j and X, emphasizing the calculation of r(pm,j, X).
Mathematics 12 02682 g005
Figure 6. Distance between observed and partitioned data. This figure shows the observed data as a red line and a partitioned data series as a blue line. The gray arrows indicate the differences between corresponding points of the observed and partitioned data at the same time points, visually representing the discrepancies across the timeline.
Figure 6. Distance between observed and partitioned data. This figure shows the observed data as a red line and a partitioned data series as a blue line. The gray arrows indicate the differences between corresponding points of the observed and partitioned data at the same time points, visually representing the discrepancies across the timeline.
Mathematics 12 02682 g006
Figure 7. Diverse time series generated from a single data-generating process (DGP). This figure shows different time series, each represented by a different colored line, illustrating the range of results produced by the same DGP.
Figure 7. Diverse time series generated from a single data-generating process (DGP). This figure shows different time series, each represented by a different colored line, illustrating the range of results produced by the same DGP.
Mathematics 12 02682 g007
Figure 8. Comparative analysis of observed and partitioned data vector distributions. The figure shows the comparison between observed data f(X) and partitioned data vectors g(pm,j). The arrows represent the flow of data from the partitioned vectors to their respective distributions, with blue denoting partitioned data and red denoting observed data. This comparison helps to determine the likelihood that the observed data came from the same DGP.
Figure 8. Comparative analysis of observed and partitioned data vector distributions. The figure shows the comparison between observed data f(X) and partitioned data vectors g(pm,j). The arrows represent the flow of data from the partitioned vectors to their respective distributions, with blue denoting partitioned data and red denoting observed data. This comparison helps to determine the likelihood that the observed data came from the same DGP.
Mathematics 12 02682 g008
Figure 9. Seasonal and trend variations in ENTSO-E power data.
Figure 9. Seasonal and trend variations in ENTSO-E power data.
Mathematics 12 02682 g009
Figure 10. Visualization of the data splitting process for different time series models. The red vertical line separates the observed data on the left from the data used to generate the split datasets on the right. (a) Type A: AR(1); (b) Type B: AR(2); (c) Type C: MA(1); (d) Type D: MA(2); (e) Type E: ARMA(1); and (f) Type F: ARMA(2).
Figure 10. Visualization of the data splitting process for different time series models. The red vertical line separates the observed data on the left from the data used to generate the split datasets on the right. (a) Type A: AR(1); (b) Type B: AR(2); (c) Type C: MA(1); (d) Type D: MA(2); (e) Type E: ARMA(1); and (f) Type F: ARMA(2).
Mathematics 12 02682 g010aMathematics 12 02682 g010b
Figure 11. Histograms of L2 norms for different model types. (a) Type A: AR(1); (b) Type B: AR(2); (c) Type C: MA(1); (d) Type D: MA(2); (e) Type E: ARMA(1); and (f) Type F: ARMA(2).
Figure 11. Histograms of L2 norms for different model types. (a) Type A: AR(1); (b) Type B: AR(2); (c) Type C: MA(1); (d) Type D: MA(2); (e) Type E: ARMA(1); and (f) Type F: ARMA(2).
Mathematics 12 02682 g011aMathematics 12 02682 g011b
Figure 12. Components of the visualization framework for statistical analysis: (A) dataset selection via a dropdown menu; (B) time series graph of observed data and a summary of the observed data; (C) plot of selected model data; (D) summary statistics of the selected model data; (E) detailed view for the scaled model data, independent test results, and trend test results; (F) display of marginal likelihoods for the combined results of independent test and trend test, the independent test result, and the trend test result; (G) buttons for selecting either “Analysis” or “Comparison”; (H) options for selecting candidate models; and (I) settings for “Scale”, test methods, significance level adjustments, and a “Run” button to execute the analysis.
Figure 12. Components of the visualization framework for statistical analysis: (A) dataset selection via a dropdown menu; (B) time series graph of observed data and a summary of the observed data; (C) plot of selected model data; (D) summary statistics of the selected model data; (E) detailed view for the scaled model data, independent test results, and trend test results; (F) display of marginal likelihoods for the combined results of independent test and trend test, the independent test result, and the trend test result; (G) buttons for selecting either “Analysis” or “Comparison”; (H) options for selecting candidate models; and (I) settings for “Scale”, test methods, significance level adjustments, and a “Run” button to execute the analysis.
Mathematics 12 02682 g012
Figure 13. Visualization of marginal likelihood. The figure highlights sections where the relationship r(pm,j, X) ≤ SignIf.indep. is satisfied in the candidate model, as shown in “A”. For statistical tests, sections where p-valuetrend(pm,j) ≤ SignIf.trend are highlighted in “B”. The combined visualization in “A∩B” illustrates areas where both criteria are met. The arrow, letters, colors, and dotted line indicate key sections of the time series data where these conditions are met, helping to identify similarities with the observed data.
Figure 13. Visualization of marginal likelihood. The figure highlights sections where the relationship r(pm,j, X) ≤ SignIf.indep. is satisfied in the candidate model, as shown in “A”. For statistical tests, sections where p-valuetrend(pm,j) ≤ SignIf.trend are highlighted in “B”. The combined visualization in “A∩B” illustrates areas where both criteria are met. The arrow, letters, colors, and dotted line indicate key sections of the time series data where these conditions are met, helping to identify similarities with the observed data.
Mathematics 12 02682 g013
Figure 14. Adjusting the significance level. This figure illustrates the impact of adjusting the significance level on the analysis results. The color shading represents different ranges of comparison results, and the slide bar allows for interactive adjustments to the significance level.
Figure 14. Adjusting the significance level. This figure illustrates the impact of adjusting the significance level on the analysis results. The color shading represents different ranges of comparison results, and the slide bar allows for interactive adjustments to the significance level.
Mathematics 12 02682 g014
Figure 15. Response prediction. The figure illustrates how users can click on a specific index in the candidate model to generate a direct comparison between the observed data (black line) and the partitioned data vector (red line). The arrows indicate the process of selecting a data point for detailed predictive analysis.
Figure 15. Response prediction. The figure illustrates how users can click on a specific index in the candidate model to generate a direct comparison between the observed data (black line) and the partitioned data vector (red line). The arrows indicate the process of selecting a data point for detailed predictive analysis.
Mathematics 12 02682 g015
Table 1. Comparative overview of research techniques and outcomes.
Table 1. Comparative overview of research techniques and outcomes.
AuthorsKey TechniquesEvaluationDifference from This Research
Fatemi et al. [19]Integrates causal inference with deep learning, GNNs, LSTM, GMM, ErosOutperformed traditional methods on cold-start scenariosFocuses on deep learning and causal inference, not Bayesian model selection
Xu et al. [20]Probabilistic embeddings, variational inference, regularized priorsSignificant improvements in cold-start scenariosUses variational inference rather than Bayesian methods for a time series
AlRossais et al. [21]Item-based stereotypes, metadata independent of user item ratingsSuperior to traditional SVD-based approachesTargets recommendation systems, not TSF
Pirasteh et al. [22]Combines multiple similarity measures, integrates user and item similaritiesOutperformed conventional CF techniques in cold-start conditionsFocuses on collaborative filtering, not Bayesian methods or time series data
Rohani et al. [23]User preferences, hierarchical preference tree structureSignificantly improved recommendation accuracyEmploys social networking for recommendations, not applicable to TSF
Ni et al. [24]SGNNs, LLM embeddings, contrastive learning frameworkOutperformed existing baselinesFocuses on student performance prediction, different domains, and methods
Tey et al. [25]Indirect relations, user preferences, social media interactionsSignificant improvements in recommendation accuracyApplies social network data, unrelated to TSF
Kuznetsov and Kordík [26]Ontologies, knowledge graphs, semantic layer in text-based methodsEffective compared with state-of-the-art text feature extraction techniquesUses knowledge graphs and ontologies, unlike Bayesian inference
Li et al. [27]Reinforced active learning, human-in-the-loop, anomaly scoringOutperformed five state-of-the-art modelsFocuses on anomaly detection, not general TSF
Xie et al. [28]High-dimensional regression, matrix factorization, leveraging metadataRobust performance across multiple datasetsEmploys regression and matrix factorization, not Bayesian model selection
Ryu et al. [29]Invocation similarity, neighborhood similarity, location dataBetter performance in cold- and warm-start scenariosUses matrix factorization for web services, different applications and methods
Xie et al. [18]Repeated patterns, low-rank decompositions, metadata weightingsAccurate predictions and imputes missing valuesFocuses on missing data and long-range forecasts, different approaches
Chen et al. [31]Frequency domain data augmentation, frequency masking, frequency mixingEnhanced forecasting accuracy, mitigated performance degradationUses data augmentation techniques, not Bayesian methods
Ebrahimi et al. [32]Application-based, checkpoint-based, invocation time prediction-based, cache-basedDiscussed various methods and evaluated their effectivenessFocuses on serverless computing, not TSF
Notes: GNNs, graph neural networks; LSTM, long short-term memory; GMM, Gaussian mixture models; Eros, extended Frobenius norm; SGNNs, signed graph neural networks; LLM, large language model; CF, collaborative filtering; SVD, singular value decomposition; TSF, time series forecasting.
Table 2. Distance measures explored in this experiment.
Table 2. Distance measures explored in this experiment.
Namer(pm,j, X) FormulaFamily
Squared Euclidean∑[Xpm,j]2Squared L2 family (χ2 family)
Pearson 1∑[(Xpm,j)2/pm,j]Squared L2 family (χ2 family)
Neyman∑[(Xpm,j)2/X]Squared L2 family (χ2 family)
Squared chi∑[(Xpm,j)2/(X + pm,j)]Squared L2 family (χ2 family)
Prob. symmetric2 × ∑[(Xpm,j)2/(X + pm,j)]Squared L2 family (χ2 family)
Divergence2 × ∑[(Xpm,j)2/(X + pm,j)2]Squared L2 family (χ2 family)
Clark√(∑[|Xpm,j|/(X + pm,j)2])Squared L2 family (χ2 family)
L2 norm(|Xpm,j|)1/2Lp Minkowski family
Manhattan∑|Xpm,j|Lp Minkowski family
Chebyshevmax|Xpm,j|Lp Minkowski family
Sorensen(∑|Xpm,j|)/(∑(Xpm,j))L1 family
Gower1/d × ∑|Xpm,j|L1 family
Kulczynski’s D(∑|Xpm,j|)/(∑[max(X, pm,j)])L1 family
Canberra(∑|Xpm,j|)/(∑[min(X, pm,j)])L1 family
Lorentzian∑[log(1 + |Xpm,j|)]L1 family
Intersection∑[min(X, pm,j)]Intersection family
Nonintersection1 − ∑[min(X, pm,j)]Intersection family
Wage hedges(∑|Xpm,j|)/max(X, pm,j)Intersection family
Czeanowski(∑|Xpm,j|)/(∑|X + pm,j|)Intersection family
Motyka(∑[min|Xpm,j|])/(∑|X + pm,j|)Intersection family
Inner product∑(X × pm,j)Inner product family
Harmonic mean2 × ∑[(X × pm,j)/(X + pm,j)]Inner product family
Cosine(∑[X × pm,j])/√(∑X2) × √(∑(pm,j)2)Inner product family
Kumar–Hassebrook(∑[X × pm,j])/(∑X2 + ∑(pm,j)2 − ∑(X × pm,j))Inner product family
Dice(∑[Xpm,j]2)/(∑X2 + ∑(pm,j)2)Etc.
Wassersteininfγ∈(P_r,P_g)Ex,yγ[‖Xpm,j‖]Etc.
1 Pearson distance, as delineated in this text, differs from the Pearson correlation coefficient, which is more common.
Table 3. Parameters of synthetic candidate models.
Table 3. Parameters of synthetic candidate models.
CategoryModel LabelMean μ Standard Deviation σAutoregressive Coefficient α
Stationary Time SeriesA0.00.20.5
B0.00.40.5
C0.00.60.5
D0.00.80.5
E0.01.00.5
Unstable VarianceA0.01.20.5
B0.01.40.5
C0.01.60.5
D0.01.80.5
E0.02.00.5
Trend ChangesA0.51.01.0
B1.01.01.0
C2.01.01.0
D−0.51.01.0
E−1.01.01.0
F−2.01.01.0
Presence of a Unit RootA0.01.00.2
B0.01.00.4
C0.01.00.6
D0.01.00.8
E0.01.01.0
Table 4. Posterior probabilities in experiments involving stationary time series.
Table 4. Posterior probabilities in experiments involving stationary time series.
Distance MeasureSignIf.0.10.30.50.70.9Distance MeasureSignIf.0.10.30.50.70.9
Squared EuclideanA0.0250.0580.0960.1350.176LorentzianA0.0210.0560.0900.1310.175
B0.0220.0590.1030.1420.180B0.0220.0590.1030.1440.180
C0.0180.0670.1030.1410.179C0.0190.0670.1080.1470.183
D0.0150.0560.1020.1420.183D0.0190.0580.1010.1410.182
E0.0210.0590.0950.1410.182E0.0200.0590.0970.1370.180
PearsonA0.0190.0600.0950.1400.180IntersectionA0.0250.0690.1070.1420.181
B0.0200.0620.1060.1450.182B0.0210.0580.1000.1400.176
C0.0180.0500.0880.1310.180C0.0170.0550.0910.1360.182
D0.0210.0660.1080.1410.180D0.0190.0580.0980.1430.182
E0.0220.0620.1030.1430.179E0.0180.0610.1040.1400.180
NeymanA0.0190.0610.0990.1380.180NonintersectionA0.0190.0580.0930.1310.175
B0.0220.0600.0960.1380.178B0.0240.0600.1000.1420.179
C0.0200.0620.1040.1430.180C0.0180.0640.1090.1450.183
D0.0220.0580.1010.1430.179D0.0180.0570.1020.1420.181
E0.0180.0590.0990.1390.184E0.0200.0600.0960.1390.182
Squared chiA0.0200.0570.0970.1360.180Wave hedgesA0.0210.0630.1000.1360.179
B0.0180.0590.1010.1410.178B0.0190.0590.0990.1410.181
C0.0200.0610.1010.1430.182C0.0170.0500.0900.1350.180
D0.0190.0600.0990.1400.179D0.0210.0630.1070.1450.180
E0.0230.0630.1020.1410.182E0.0220.0650.1030.1430.180
Prob. symmetricA0.0200.0570.0970.1360.180CzekanowskiA0.0220.0570.1000.1440.181
B0.0180.0590.1010.1410.178B0.0180.0610.1030.1430.182
C0.0200.0610.1010.1430.182C0.0180.0560.1010.1390.180
D0.0190.0600.0990.1400.179D0.0240.0650.0990.1360.180
E0.0230.0630.1020.1410.182E0.0180.0610.0970.1380.177
DivergenceA0.0180.0570.1010.1400.180MotykaA0.0220.0570.1000.1440.181
B0.0200.0610.1000.1400.181B0.0180.0610.1030.1430.182
C0.0190.0620.1000.1430.180C0.0180.0560.1010.1390.180
D0.0220.0630.1010.1380.182D0.0240.0650.0990.1360.180
E0.0200.0560.0980.1390.177E0.0180.0610.0970.1380.177
ClarkA0.0180.0570.1010.1400.180Inner productA0.0230.0650.1030.1410.176
B0.0200.0610.1000.1400.181B0.0210.0600.0990.1410.177
C0.0190.0620.1000.1430.180C0.0220.0600.0950.1330.182
D0.0220.0630.1010.1380.182D0.0170.0560.0980.1430.185
E0.0200.0560.0980.1390.177E0.0180.0600.1050.1420.179
L2 normA0.0250.0580.0960.1350.176Harmonic meanA0.0200.0640.1030.1430.180
B0.0220.0590.1030.1420.180B0.0220.0590.0990.1410.182
C0.0180.0670.1030.1410.179C0.0180.0570.0990.1400.180
D0.0150.0560.1020.1420.183D0.0210.0600.1010.1400.181
E0.0210.0590.0950.1410.182E0.0180.0590.0970.1360.177
ManhattanA0.0220.0570.0920.1310.176CosineA0.0240.0650.1030.1410.176
B0.0220.0560.1040.1430.180B0.0200.0600.0990.1410.178
C0.0180.0670.1090.1430.181C0.0220.0590.0950.1330.182
D0.0170.0590.1000.1430.181D0.0170.0560.0980.1430.185
E0.0210.0600.0950.1400.181E0.0170.0600.1050.1420.180
ChebyshevA0.0280.0710.1140.1530.188Kumar–HassebrookA0.0240.0650.1030.1410.176
B0.0160.0590.0980.1370.180B0.0200.0600.0990.1410.178
C0.0160.0540.0930.1340.176C0.0220.0590.0950.1330.182
D0.0200.0570.1000.1400.180D0.0170.0560.0980.1430.185
E0.0220.0590.0960.1360.177E0.0170.0600.1050.1420.180
SorensenA0.0220.0570.1000.1440.181DiceA0.0240.0590.0970.1350.176
B0.0180.0610.1030.1430.182B0.0220.0590.1010.1400.180
C0.0180.0560.1010.1390.180C0.0180.0670.1050.1410.178
D0.0240.0650.0990.1360.180D0.0150.0570.1020.1440.183
E0.0180.0610.0970.1380.177E0.0200.0580.0950.1400.183
GowerA0.0220.0570.0920.1310.176WassersteinA0.0480.1040.1480.1750.194
B0.0220.0560.1040.1430.180B0.0230.0730.1090.1460.182
C0.0180.0670.1090.1430.181C0.0160.0490.0860.1270.159
D0.0170.0590.1000.1430.181D0.0120.0370.0690.1140.184
E0.0210.0600.0950.1400.181E0.0020.0380.0870.1380.180
Kulczynski‘s DA0.0120.0440.1090.1620.184Kolmogorov–SmirnovA0.0380.0780.1280.1620.181
B0.0510.1210.1460.1730.195B0.0290.0780.1070.1410.155
C0.0280.0690.1030.1350.191C0.0040.0340.0700.1080.156
D0.0020.0410.0940.1410.182D0.0080.0340.0550.0950.171
E0.0060.0240.0480.0890.149E0.0160.0530.1090.1480.178
CanberraA0.0220.0590.0990.1410.180RunsA0.0120.0520.0830.1250.161
B0.0180.0590.0990.1390.178B0.0090.0480.0840.1380.172
C0.0180.0610.1000.1410.180C0.0120.0530.0880.1310.171
D0.0220.0590.0990.1390.180D0.0100.0520.0920.1340.171
E0.0210.0610.1030.1400.182E0.0100.0530.0860.1280.167
Table 5. Posterior probabilities in experiments with unstable variance.
Table 5. Posterior probabilities in experiments with unstable variance.
Distance MeasureSignIf.0.10.30.50.70.9Distance MeasureSignIf.0.10.30.50.70.9
Squared EuclideanA0.0170.0590.0990.1400.179LorentzianA0.0190.0560.0950.1380.177
B0.0200.0620.1020.1400.181B0.0210.0640.1020.1400.181
C0.0220.0620.1000.1390.181C0.0210.0610.1060.1450.184
D0.0210.0600.0980.1410.178D0.0200.0590.0980.1370.178
E0.0200.0570.1010.1390.181E0.0190.0610.0990.1410.180
PearsonA0.0190.0580.1010.1390.179IntersectionA0.0220.0610.1040.1420.180
B0.0170.0520.0970.1440.182B0.0200.0610.1010.1390.180
C0.0210.0620.0970.1360.182C0.0170.0580.0970.1370.180
D0.0240.0650.1040.1400.179D0.0220.0630.1000.1400.180
E0.0190.0630.1010.1410.178E0.0200.0580.0980.1420.180
NeymanA0.0220.0620.1000.1420.182NonintersectionA0.0200.0580.0960.1390.178
B0.0200.0600.0990.1420.180B0.0200.0610.0990.1390.180
C0.0210.0570.0990.1370.180C0.0200.0630.1030.1420.183
D0.0170.0640.1020.1410.177D0.0200.0600.1000.1370.178
E0.0200.0570.0990.1390.181E0.0200.0580.1020.1420.180
Squared chiA0.0190.0600.1000.1410.181Wave hedgesA0.0210.0580.0990.1410.180
B0.0200.0610.1020.1410.180B0.0200.0600.0950.1410.181
C0.0130.0550.0990.1370.179C0.0190.0590.1010.1370.181
D0.0270.0660.1030.1390.179D0.0200.0630.1040.1420.178
E0.0200.0590.0970.1430.181E0.0200.0610.1000.1390.180
Prob. symmetricA0.0190.0600.1000.1410.181CzekanowskiA0.0200.0630.1010.1400.181
B0.0200.0610.1020.1410.180B0.0220.0640.1080.1440.180
C0.0130.0550.0990.1370.179C0.0190.0580.0950.1340.175
D0.0270.0660.1030.1390.179D0.0190.0590.0980.1410.181
E0.0200.0590.0970.1430.181E0.0210.0570.0970.1400.182
DivergenceA0.0190.0590.1010.1410.181MotykaA0.0200.0630.1010.1400.181
B0.0210.0570.0970.1370.178B0.0220.0640.1080.1440.180
C0.0230.0650.1050.1450.183C0.0190.0580.0950.1340.175
D0.0190.0570.0990.1350.176D0.0190.0590.0980.1410.181
E0.0190.0610.0980.1420.182E0.0210.0570.0970.1400.182
ClarkA0.0190.0590.1010.1410.181Inner productA0.0210.0590.1000.1430.183
B0.0210.0570.0970.1370.178B0.0200.0600.0980.1370.180
C0.0230.0650.1050.1450.183C0.0180.0610.1000.1390.179
D0.0190.0570.0990.1350.176D0.0220.0590.1020.1370.179
E0.0190.0610.0980.1420.182E0.0200.0610.1000.1450.179
L2 normA0.0170.0590.0990.1400.179Harmonic meanA0.0190.0590.1000.1400.181
B0.0200.0620.1020.1400.181B0.0200.0590.0980.1390.180
C0.0220.0620.1000.1390.181C0.0210.0630.1010.1450.187
D0.0210.0600.0980.1410.178D0.0210.0610.0970.1340.173
E0.0200.0570.1010.1390.181E0.0190.0560.1030.1410.180
ManhattanA0.0190.0580.0970.1360.177CosineA0.0200.0590.1000.1430.183
B0.0190.0620.1000.1400.182B0.0200.0600.0980.1370.180
C0.0200.0630.1030.1430.183C0.0180.0610.1000.1390.179
D0.0230.0590.0990.1390.178D0.0210.0590.1020.1370.179
E0.0190.0580.1000.1410.181E0.0200.0610.1000.1450.179
ChebyshevA0.0260.0710.1080.1490.186Kumar–HassebrookA0.0200.0590.1000.1430.183
B0.0200.0590.1020.1400.178B0.0200.0600.0980.1370.180
C0.0170.0530.0910.1320.176C0.0180.0610.1000.1390.179
D0.0200.0580.0990.1410.179D0.0210.0590.1020.1370.179
E0.0180.0590.0990.1380.181E0.0200.0610.1000.1450.179
SorensenA0.0200.0630.1010.1400.181DiceA0.0170.0570.1000.1410.180
B0.0220.0640.1080.1440.180B0.0200.0630.1020.1400.180
C0.0190.0580.0950.1340.175C0.0210.0610.1000.1390.182
D0.0190.0590.0980.1410.181D0.0210.0630.0980.1410.179
E0.0210.0570.0970.1400.182E0.0210.0550.1000.1390.180
GowerA0.0190.0580.0970.1360.177WassersteinA0.0240.0600.0950.1380.197
B0.0190.0620.1000.1400.182B0.0240.0710.1030.1360.176
C0.0200.0630.1030.1430.183C0.0090.0400.0900.1390.180
D0.0230.0590.0990.1390.178D0.0310.0720.1090.1460.175
E0.0190.0580.1000.1410.181E0.0120.0570.1030.1400.171
Kulczynski‘s DA0.0180.0560.1090.1540.193Kolmogorov–SmirnovA0.0260.0670.1190.1600.185
B0.0270.0650.1050.1560.188B0.0160.0490.0950.1330.182
C0.0160.0690.1150.1410.187C0.0080.0440.0850.1340.175
D0.0150.0520.0900.1350.188D0.0380.0760.0960.1350.172
E0.0240.0570.0810.1130.144E0.0100.0630.1010.1360.186
CanberraA0.0180.0590.0980.1410.181RunsA0.0190.0510.0910.1370.177
B0.0220.0630.1000.1390.178B0.0220.0600.1010.1390.178
C0.0140.0530.1000.1380.179C0.0210.0580.1040.1390.184
D0.0250.0670.1040.1400.180D0.0140.0490.0890.1280.176
E0.0200.0580.0980.1420.182E0.0230.0610.1030.1440.183
Table 6. Posterior probabilities in experiments on trend changes.
Table 6. Posterior probabilities in experiments on trend changes.
Distance MeasureSignIf.0.10.30.50.70.9Distance MeasureSignIf.0.10.30.50.70.9
Squared EuclideanA0.0170.0400.1670.1670.167LorentzianA0.0480.0830.1670.1670.167
B0.0400.1080.1670.1670.167B0.0390.1050.1670.1670.167
C0.0430.1520.1670.1670.167C0.0130.1120.1670.1670.167
D0.0000.0000.0000.0920.122D0.0000.0000.0000.0990.127
E0.0000.0000.0000.0620.132E0.0000.0000.0000.0610.129
F0.0000.0000.0000.0460.146F0.0000.0000.0000.0410.145
PearsonA0.0210.0700.1060.1320.153IntersectionA0.0000.0000.0000.0830.117
B0.0180.0600.1010.1250.147B0.0000.0000.0000.0620.129
C0.0200.0600.1040.1300.152C0.0000.0000.0000.0540.154
D0.0120.0360.0630.1020.151D0.0370.0730.1670.1670.167
E0.0150.0380.0640.1060.149E0.0390.1020.1670.1670.167
F0.0140.0360.0620.1040.148F0.0240.1240.1670.1670.167
NeymanA0.0300.0530.0720.0850.119NonintersectionA0.0500.0830.1670.1670.167
B0.0030.0330.0670.0990.147B0.0370.1040.1670.1670.167
C0.0000.0040.0350.1090.164C0.0130.1120.1670.1670.167
D0.0510.0830.1010.1170.139D0.0000.0000.0000.0930.129
E0.0160.0670.1040.1350.164E0.0000.0000.0000.0650.128
F0.0000.0600.1210.1550.167F0.0000.0000.0000.0420.143
Squared chiA0.0060.0140.0520.1050.156Wave hedgesA0.0170.0460.0750.1100.149
B0.0040.0140.0450.1070.157B0.0150.0370.0640.1030.145
C0.0060.0150.0510.1100.156C0.0150.0360.0620.1060.149
D0.0220.0710.1090.1210.143D0.0170.0600.0990.1270.152
E0.0290.0890.1200.1270.142E0.0180.0630.1010.1270.152
F0.0330.0980.1240.1300.145F0.0180.0600.1000.1270.152
Prob. symmetricA0.0060.0140.0520.1050.156CzekanowskiA0.0660.0960.1670.1670.167
B0.0040.0140.0450.1070.157B0.0270.1030.1670.1670.167
C0.0060.0150.0510.1100.156C0.0070.1010.1670.1670.167
D0.0220.0710.1090.1210.143D0.0000.0000.0000.0740.106
E0.0290.0890.1200.1270.142E0.0000.0000.0000.0650.138
F0.0330.0980.1240.1300.145F0.0000.0000.0000.0620.156
DivergenceA0.0390.0770.1070.1330.156MotykaA0.0660.0960.1670.1670.167
B0.0280.0770.1110.1360.157B0.0270.1030.1670.1670.167
C0.0140.0680.1070.1340.155C0.0070.1010.1670.1670.167
D0.0110.0350.0690.1080.146D0.0000.0000.0000.0740.106
E0.0060.0290.0570.0970.143E0.0000.0000.0000.0650.138
F0.0010.0140.0500.0920.143F0.0000.0000.0000.0620.156
ClarkA0.0390.0770.1070.1330.156Inner productA0.0000.0000.0000.1030.124
B0.0280.0770.1110.1360.157B0.0000.0000.0000.0640.127
C0.0140.0680.1070.1340.155C0.0000.0000.0000.0330.148
D0.0110.0350.0690.1080.146D0.0360.0630.1670.1670.167
E0.0060.0290.0570.0970.143E0.0350.1030.1670.1670.167
F0.0010.0140.0500.0920.143F0.0290.1340.1670.1670.167
L2 normA0.0170.0400.1670.1670.167Harmonic meanA0.0110.0620.1150.1530.161
B0.0400.1080.1670.1670.167B0.0100.0600.1220.1530.163
C0.0430.1520.1670.1670.167C0.0100.0570.1170.1510.161
D0.0000.0000.0000.0920.122D0.0230.0460.0570.0960.144
E0.0000.0000.0000.0620.132E0.0240.0390.0470.0780.137
F0.0000.0000.0000.0460.146F0.0210.0360.0430.0690.134
ManhattanA0.0450.0780.1670.1670.167CosineA0.0000.0000.0000.1120.134
B0.0410.1050.1670.1670.167B0.0000.0000.0000.0630.124
C0.0140.1170.1670.1670.167C0.0000.0000.0000.0250.141
D0.0000.0000.0000.1040.130D0.0260.0560.1670.1670.167
E0.0000.0000.0000.0600.128E0.0360.1040.1670.1670.167
F0.0000.0000.0000.0360.142F0.0370.1390.1670.1670.167
ChebyshevA0.0160.0400.1660.1670.167HassebrookA0.0000.0000.0000.1110.134
B0.0310.1020.1670.1670.167B0.0000.0000.0000.0630.124
C0.0530.1580.1670.1670.167C0.0000.0000.0000.0260.141
D0.0000.0000.0010.0790.113D0.0270.0570.1670.1670.167
E0.0000.0000.0000.0680.136E0.0360.1040.1670.1670.167
F0.0000.0000.0000.0530.151F0.0370.1390.1670.1670.167
SorensenA0.0660.0960.1670.1670.167DiceA0.0320.0550.1670.1670.167
B0.0270.1030.1670.1670.167B0.0420.1030.1670.1670.167
C0.0070.1010.1670.1670.167C0.0250.1410.1670.1670.167
D0.0000.0000.0000.0740.106D0.0000.0000.0000.1100.140
E0.0000.0000.0000.0650.138E0.0000.0000.0000.0630.130
F0.0000.0000.0000.0620.156F0.0000.0000.0000.0270.130
GowerA0.0450.0780.1670.1670.167WassersteinA0.0150.0310.0450.0670.127
B0.0410.1050.1670.1670.167B0.0190.0560.0860.1250.162
C0.0140.1170.1670.1670.167C0.0120.0550.1150.1550.167
D0.0000.0000.0000.1040.130D0.0210.0380.0560.0720.111
E0.0000.0000.0000.0600.128E0.0200.0560.0850.1280.167
F0.0000.0000.0000.0360.142F0.0140.0650.1120.1530.167
Kulczynski‘s DA0.0000.0000.0000.1100.144Kolmogorov–SmirnovA0.0170.0410.0500.0700.120
B0.0000.0000.0000.0570.160B0.0190.0510.0740.1190.162
C0.0000.0000.0000.0330.096C0.0070.0600.1180.1570.167
D0.0640.1670.1670.1670.167D0.0180.0380.0480.0700.115
E0.0260.0930.1670.1670.167E0.0200.0500.0700.1240.166
F0.0090.0400.1670.1670.167F0.0100.0600.1050.1550.167
CanberraA0.0100.0220.0580.1070.154RunsA0.0010.0060.0220.0490.117
B0.0070.0210.0530.1090.154B0.0030.0270.0740.1280.163
C0.0090.0250.0570.1100.154C0.0230.1020.1520.1660.167
D0.0210.0640.1020.1220.147D0.0020.0090.0280.0530.108
E0.0240.0790.1130.1250.144E0.0040.0260.0730.1220.163
F0.0290.0880.1170.1280.147F0.0220.0990.1500.1660.167
Table 7. Posterior probabilities in experiments determining the presence of a unit root.
Table 7. Posterior probabilities in experiments determining the presence of a unit root.
Distance MeasureSignIf.0.10.30.50.70.9Distance MeasureSignIf.0.10.30.50.70.9
Squared EuclideanA0.0260.0620.0980.1360.175LorentzianA0.0230.0610.0970.1370.177
B0.0280.0630.0960.1320.172B0.0250.0640.1000.1360.175
C0.0210.0600.1000.1370.179C0.0210.0650.1090.1450.180
D0.0160.0570.0990.1400.181D0.0220.0660.1070.1420.181
E0.0100.0570.1070.1560.193E0.0100.0450.0870.1390.187
PearsonA0.0220.0690.1090.1440.177IntersectionA0.0190.0600.1000.1420.183
B0.0170.0510.0880.1370.182B0.0240.0610.1010.1370.178
C0.0210.0590.0950.1330.181C0.0190.0550.0960.1370.177
D0.0210.0590.0990.1370.177D0.0220.0630.1000.1390.176
E0.0190.0620.1080.1490.182E0.0160.0610.1030.1450.186
NeymanA0.0190.0590.1010.1430.183NonintersectionA0.0170.0580.1000.1400.181
B0.0210.0600.1000.1450.181B0.0220.0630.0990.1390.176
C0.0220.0630.1010.1420.180C0.0230.0630.1040.1450.181
D0.0220.0600.1010.1420.180D0.0240.0610.1000.1370.178
E0.0160.0590.0960.1290.176E0.0140.0550.0970.1390.184
Squared chiA0.0200.0590.1010.1420.182Wave hedgesA0.0240.0630.1050.1410.183
B0.0210.0660.1050.1410.180B0.0180.0560.0950.1370.181
C0.0180.0550.1030.1440.181C0.0210.0620.1040.1480.182
D0.0170.0570.0980.1380.180D0.0190.0570.0950.1370.180
E0.0240.0630.0930.1350.178E0.0180.0610.1020.1370.174
Prob. symmetricA0.0200.0590.1010.1420.182CzekanowskiA0.0300.0780.1000.1240.174
B0.0210.0660.1050.1410.180B0.0220.0740.1050.1380.178
C0.0180.0550.1030.1440.181C0.0210.0580.0960.1300.178
D0.0170.0570.0980.1380.180D0.0150.0510.1000.1460.182
E0.0240.0630.0930.1350.178E0.0120.0380.0990.1610.187
DivergenceA0.0240.0630.1020.1460.180MotykaA0.0300.0780.1000.1240.174
B0.0180.0530.0930.1370.181B0.0220.0740.1050.1380.178
C0.0220.0660.1090.1450.180C0.0210.0580.0960.1300.178
D0.0190.0600.0990.1410.182D0.0150.0510.1000.1460.182
E0.0180.0590.0960.1320.177E0.0120.0380.0990.1610.187
ClarkA0.0240.0630.1020.1460.180Inner productA0.0250.0630.0990.1400.174
B0.0180.0530.0930.1370.181B0.0260.0650.1030.1350.172
C0.0220.0660.1090.1450.180C0.0220.0610.0990.1370.178
D0.0190.0600.0990.1410.182D0.0170.0580.1010.1430.185
E0.0180.0590.0960.1320.177E0.0100.0520.0990.1450.190
L2 normA0.0260.0620.0980.1360.175Harmonic meanA0.0180.0580.0980.1410.180
B0.0280.0630.0960.1320.172B0.0200.0590.0950.1350.179
C0.0210.0600.1000.1370.179C0.0190.0560.0980.1450.182
D0.0160.0570.0990.1400.181D0.0200.0620.1020.1430.183
E0.0100.0570.1070.1560.193E0.0220.0650.1070.1370.176
ManhattanA0.0240.0590.0970.1360.176CosineA0.0250.0640.0990.1400.174
B0.0250.0650.0980.1330.173B0.0260.0650.1030.1350.172
C0.0210.0650.1070.1410.180C0.0220.0610.0990.1360.179
D0.0210.0660.1050.1430.182D0.0170.0580.1010.1430.184
E0.0090.0440.0940.1480.189E0.0100.0520.0990.1460.190
ChebyshevA0.0170.0580.1020.1460.186Kumar–HassebrookA0.0250.0640.0990.1400.174
B0.0200.0600.0970.1380.181B0.0260.0650.1030.1350.172
C0.0170.0540.0940.1320.176C0.0220.0610.0990.1360.179
D0.0160.0500.0890.1290.170D0.0170.0580.1010.1430.184
E0.0290.0790.1190.1560.187E0.0100.0520.0990.1460.190
SorensenA0.0300.0780.1000.1240.174DiceA0.0260.0600.1010.1360.175
B0.0220.0740.1050.1380.178B0.0280.0650.0970.1350.174
C0.0210.0580.0960.1300.178C0.0210.0640.1010.1390.178
D0.0150.0510.1000.1460.182D0.0160.0570.0990.1420.183
E0.0120.0380.0990.1610.187E0.0100.0540.1010.1480.190
GowerA0.0240.0590.0970.1360.176WassersteinA0.0620.1360.1780.1930.200
B0.0250.0650.0980.1330.173B0.0160.0650.1200.1720.200
C0.0210.0650.1070.1410.180C0.0100.0550.1160.1840.200
D0.0210.0660.1050.1430.182D0.0100.0370.0710.1160.196
E0.0090.0440.0940.1480.189E0.0030.0080.0150.0340.104
Kulczynski‘s DA0.0180.0740.1450.2000.200Kolmogorov–SmirnovA0.0500.1140.1590.1990.200
B0.0070.0590.1110.1690.200B0.0150.0630.1020.1560.198
C0.0230.0830.1360.1930.200C0.0180.0710.1330.1860.200
D0.0520.0840.1080.1390.200D0.0140.0430.0810.1230.196
E0.0000.0000.0000.0000.100E0.0030.0090.0190.0350.105
CanberraA0.0190.0580.0980.1410.180RunsA0.0260.0670.1050.1470.181
B0.0210.0640.1050.1410.180B0.0160.0560.0880.1370.186
C0.0180.0600.1010.1430.180C0.0220.0650.1030.1550.191
D0.0180.0570.0990.1390.180D0.0190.0600.0980.1500.185
E0.0240.0610.0970.1370.179E0.0120.0380.0640.1010.142
Table 8. Detailed ISO country codes with corresponding national profiles.
Table 8. Detailed ISO country codes with corresponding national profiles.
Country CodeCountry NameCountry CodeCountry Name
ATAustriaIEIreland
BABosnia HerzegovinaISIceland
BEBelgiumITItaly
BGBulgariaLTLithuania
CHSwitzerlandLULuxembourg
CSSerbia and MontenegroLVLatvia
CYCyprusMEMontenegro
CZCzech RepublicMKNorth Macedonia
DEGermanyNINorthern Ireland
DKDenmarkNLNetherlands
DKWDenmark WestNONorway
EEEstoniaPLPoland
ESSpainPTPortugal
FRFranceRORomania
GBGreat BritainRSSerbia
GRGreeceSESweden
HRCroatiaSISlovenia
HUHungarySKSlovakia
Table 9. Analysis of posterior probabilities in ENTSO-E experiments, ranging from squared Euclidean to Canberra distances.
Table 9. Analysis of posterior probabilities in ENTSO-E experiments, ranging from squared Euclidean to Canberra distances.
Country CodeSquared EuclideanPearsonNeymanSquared ChiProb. SymmetricDivergenceClark
AT0.002680.002530.002010.002620.002620.002270.00227
BA0.002800.002870.002430.002600.002600.002570.00257
BE0.003150.003320.002320.002850.002850.002630.00263
BG0.003290.003000.002440.002810.002810.002810.00281
CH0.002420.002270.002120.002660.002660.002200.00220
CS0.001710.002910.003670.002500.002500.002180.00218
CY0.001710.002990.003510.002240.002240.001990.00199
CZ0.003360.002640.001770.002960.002960.002770.00277
DE0.002270.002760.002810.002470.002470.003030.00303
DK0.002820.002180.003090.002800.002800.002830.00283
DKW0.002730.002350.002130.002670.002670.002480.00248
EE0.004230.001800.002990.002960.002960.003110.00311
ES0.003840.002660.001730.003060.003060.002870.00287
FR0.002440.002390.001850.002600.002600.002260.00226
GB0.003930.003000.002130.003100.003100.002580.00258
GR0.001390.003250.003290.002230.002230.002260.00226
HR0.003720.002990.002260.002990.002990.002610.00261
HU0.002980.002930.003930.002440.002440.004150.00415
IE0.002810.002990.003540.002500.002500.003950.00395
IS0.004050.002290.001610.003180.003180.002690.00269
IT0.003280.002690.001850.002950.002950.002800.00280
LT0.002430.002570.002470.002600.002600.002160.00216
LU0.002190.003210.003910.002330.002330.002560.00256
LV0.002500.002220.001810.002690.002690.002200.00220
ME0.002630.002390.001600.002760.002760.002350.00235
MK0.001940.002320.002310.002600.002600.002040.00204
NI0.002550.002870.003940.002380.002380.003660.00366
NL0.002450.002900.003190.002470.002470.002530.00253
NO0.002520.002270.002220.002710.002710.002160.00216
PL0.002610.002400.002470.002640.002640.002250.00225
PT0.002180.002640.003200.002560.002560.003090.00309
RO0.002110.002950.003520.002470.002470.003370.00337
RS0.001250.003400.003770.002070.002070.002480.00248
SE0.002690.002750.004230.002480.002480.003460.00346
SI0.002390.002390.002430.002520.002520.002260.00226
SK0.002440.002060.001730.002630.002630.002040.00204
Country codeL2 normManhattanChebyshevSorensenGowerKulczynski‘s DCanberra
AT0.002680.002990.002510.002590.002990.002880.00257
BA0.002800.003190.002700.002820.003190.000000.00252
BE0.003150.003450.002790.002880.003450.001970.00272
BG0.003290.003510.002490.002570.003510.004600.00274
CH0.002420.002590.002120.002430.002590.000530.00265
CS0.001710.001850.000970.001600.001850.005290.00251
CY0.001710.001850.001140.001460.001850.006180.00232
CZ0.003360.003370.003640.003320.003370.001750.00291
DE0.002270.002210.002210.002560.002210.002430.00255
DK0.002820.002360.004010.003240.002360.001230.00279
DKW0.002730.002670.003420.002690.002670.001530.00270
EE0.004230.003590.005980.004760.003590.001900.00295
ES0.003840.003750.004130.003940.003750.000740.00297
FR0.002440.002770.002450.002310.002770.000030.00260
GB0.003930.003970.003640.003780.003970.000940.00295
GR0.001390.001450.000670.001910.001450.001250.00230
HR0.003720.003730.003150.003400.003730.002240.00289
HU0.002980.002360.004360.002460.002360.004020.00248
IE0.002810.002230.003930.002610.002230.004670.00251
IS0.004050.003970.003820.003990.003970.000230.00309
IT0.003280.003640.003090.003280.003640.000000.00285
LT0.002430.002650.001890.002150.002650.002510.00258
LU0.002190.002320.000850.001270.002320.006310.00236
LV0.002500.002670.002910.002590.002670.000670.00271
ME0.002630.002850.002730.002810.002850.000900.00274
MK0.001940.002020.001710.002330.002020.002290.00265
NI0.002550.001870.003930.003030.001870.006020.00247
NL0.002450.002660.001430.001700.002660.005290.00247
NO0.002520.002580.001810.002410.002580.000560.00272
PL0.002610.002680.002170.002330.002680.003470.00265
PT0.002180.001900.001920.002580.001900.004890.00262
RO0.002110.001760.002930.002430.001760.005460.00251
RS0.001250.001500.000480.001660.001500.004310.00220
SE0.002690.001930.004150.003580.001930.005280.00249
SI0.002390.002640.001760.002210.002640.003550.00252
SK0.002440.002730.002320.002360.002730.000340.00259
Table 10. Analysis of posterior probabilities in ENTSO-E experiments from Lorentzian distance to runs test.
Table 10. Analysis of posterior probabilities in ENTSO-E experiments from Lorentzian distance to runs test.
Country CodeLorentzianIntersectionNonintersectionWave HedgesCzekanowskiMotykaInner Product
AT0.003180.002580.003110.002310.002590.002590.00245
BA0.003450.003070.003290.002390.002820.002820.00290
BE0.003670.003920.003460.002980.002880.002880.00309
BG0.003610.003910.003560.002960.002570.002570.00309
CH0.002670.001970.002640.002310.002430.002430.00218
CS0.001940.001600.001860.002980.001600.001600.00105
CY0.001900.000930.001770.002640.001460.001460.00066
CZ0.003370.004440.003360.002800.003320.003320.00399
DE0.002190.002480.002210.002440.002560.002560.00280
DK0.002120.001660.002330.002440.003240.003240.00254
DKW0.002630.002210.002750.002750.002690.002690.00271
EE0.003170.003720.003560.002250.004760.004760.00400
ES0.003680.005110.003750.002630.003940.003940.00435
FR0.002950.002450.002850.002390.002310.002310.00227
GB0.004050.004760.003940.003050.003780.003780.00331
GR0.001480.000940.001340.002960.001910.001910.00071
HR0.003790.004310.003740.002900.003400.003400.00316
HU0.001960.002930.002290.002920.002460.002460.00407
IE0.001930.003470.002170.003020.002610.002610.00395
IS0.003910.005510.003940.002120.003990.003990.00442
IT0.003820.004140.003680.002420.003280.003280.00327
LT0.002750.001970.002650.002640.002150.002150.00213
LU0.002360.002020.002370.003030.001270.001270.00115
LV0.002730.002490.002710.002590.002590.002590.00261
ME0.002990.002850.003010.002420.002810.002810.00290
MK0.002050.001860.001930.002420.002330.002330.00217
NI0.001520.001980.001790.003000.003030.003030.00326
NL0.002760.002720.002650.002790.001700.001700.00221
NO0.002630.001840.002620.002510.002410.002410.00197
PL0.002700.002080.002760.002560.002330.002330.00219
PT0.001770.001870.001870.002550.002580.002580.00274
RO0.001620.002480.001680.003040.002430.002430.00317
RS0.001630.001120.001320.002780.001660.001660.00062
SE0.001440.001620.001810.002840.003580.003580.00299
SI0.002790.002320.002630.002560.002210.002210.00263
SK0.002950.002320.002820.002290.002360.002360.00245
Country codeHarmonic meanCosineKumar–HassebrookDiceWassersteinKolmogorov–SmirnovRuns
AT0.002880.002470.002470.002680.014040.014130.01195
BA0.002930.002870.002870.002780.012600.010560.00441
BE0.003020.003060.003060.003150.004170.002210.00233
BG0.002830.003050.003050.003290.001020.000180.00000
CH0.002810.002190.002190.002420.002820.003640.00707
CS0.002420.001030.001030.001710.000050.000120.00144
CY0.002440.000640.000640.001720.002110.002860.00438
CZ0.002980.003980.003980.003340.000430.000210.00056
DE0.002690.002770.002770.002250.000020.000290.00107
DK0.002540.002540.002540.002810.000000.000000.00003
DKW0.002500.002730.002730.002740.000000.000000.00055
EE0.002270.004010.004010.004230.000000.000000.00000
ES0.002800.004350.004350.003820.000000.000000.00004
FR0.002900.002270.002270.002450.008140.007270.00507
GB0.002910.003310.003310.003920.000000.000000.00000
GR0.002420.000730.000730.001420.000400.000800.00000
HR0.002900.003140.003150.003710.000030.000010.00000
HU0.002030.004060.004060.002990.000030.000020.00000
IE0.002200.003940.003930.002810.000000.000000.00000
IS0.003060.004430.004430.004030.000000.000000.00000
IT0.003050.003270.003270.003260.005180.003880.00318
LT0.002600.002140.002140.002440.006080.006280.00696
LU0.002390.001080.001080.002190.000070.000130.00175
LV0.002690.002640.002640.002510.001450.001090.00233
ME0.003080.002910.002910.002620.004220.004490.00353
MK0.002710.002200.002200.001950.001630.001950.00323
NI0.002020.003250.003250.002560.000000.000000.00002
NL0.002650.002160.002160.002450.004400.004260.00575
NO0.002710.001980.001980.002520.002190.002080.00005
PL0.002590.002210.002210.002610.000520.000490.00033
PT0.002560.002730.002730.002170.000000.000030.00000
RO0.002320.003160.003160.002110.000000.000080.00039
RS0.002360.000620.000620.001270.003650.005280.00314
SE0.001880.002940.002940.002680.000000.000000.00003
SI0.002600.002630.002630.002400.008540.006930.00620
SK0.002860.002490.002490.002450.010660.010550.01026
Table 11. Summary of L2 norm results across different model types and parameter settings.
Table 11. Summary of L2 norm results across different model types and parameter settings.
Type 1m = 1m = 2m = 3m = 4m = 5Type 2m = 1m = 2m = 3m = 4m = 5
Min.0.0000.0000.0000.0000.000Min.0.0000.0000.0000.0000.000
1st Qu.0.4910.5000.5240.7311.3241st Qu.0.5090.5220.5540.8011.492
Median1.0401.0581.1091.5482.803Median1.0771.1061.1741.6963.155
Mean1.2301.2511.3121.8313.314Mean1.2771.3111.3902.0053.732
3rd Qu.1.7721.8031.8912.6404.7793rd Qu.1.8391.8882.0032.8905.378
Max.8.5718.5979.38713.48020.753Max.9.1769.1339.94414.54924.916
Std. Dev.0.9290.9460.9911.3842.502Std. Dev.0.9680.9931.0531.5132.820
Type 3m = 1m = 2m = 3m = 4m = 5Type 4m = 1m = 2m = 3m = 4m = 5
Min.0.0000.0000.0000.0000.000Min.0.0000.0000.0000.0000.000
1st Qu.0.4810.4890.5050.6570.8261st Qu.0.5070.5220.5530.7460.979
Median1.0171.0341.0701.3911.749Median1.0741.1061.1701.5792.071
Mean1.2041.2241.2661.6452.067Mean1.2691.3081.3841.8672.447
3rd Qu.1.7351.7641.8252.3712.9813rd Qu.1.8301.8851.9942.6923.529
Max.9.3608.3988.88711.31714.086Max.9.5108.9099.65113.41417.055
Std. Dev.0.9100.9250.9571.2411.559Std. Dev.0.9590.9881.0461.4101.846
Type 5m = 1m = 2m = 3m = 4m = 5Type 6m = 1m = 2m = 3m = 4m = 5
Min.0.0000.0000.0000.0000.000Min.0.0000.0000.0000.0000.000
1st Qu.0.5910.6210.6940.8561.4451st Qu.0.6500.6930.7850.9881.663
Median1.2511.3131.4701.8123.060Median1.3761.4651.6612.0923.518
Mean1.4821.5551.7402.1453.619Mean1.6271.7341.9652.4724.161
3rd Qu.2.1342.2402.5073.0925.2193rd Qu.2.3442.4982.8323.5645.995
Max.10.23410.69612.26715.04323.159Max.10.67412.48414.56417.81127.340
Std. Dev.1.1231.1771.3161.6222.732Std. Dev.1.2301.3101.4861.8663.144
Table 12. L2 norm results for different model configurations.
Table 12. L2 norm results for different model configurations.
Ordered by Min.
p1p2q1q2σMin.1st Qu.MedianMean3rd Qu.Max.Rank of Min.
0.9−0.1000.51.46611.91772.03062.02972.13542.50831
0.9−0.10.1−0.10.51.50461.93682.03152.03322.13662.46372
0.9−0.10.9−0.10.51.50492.14602.30432.30512.46452.95833
0.500.1−0.90.51.73471.90581.95001.95231.99602.1738182
0.500.5011.73492.24102.38452.38652.52763.1189183
0.100.5−0.111.73522.06752.16392.16532.25852.8228184
0.9−0.90.9−0.938.030812.048314.139814.327116.133823.06271278
0.5−0.90.5−0.938.564811.283413.041513.158414.584520.13271279
0.1−0.90.1−0.938.725311.555912.943413.106714.688918.58141280
Ordered by 1st Qu.
p1p2q1q2σMin.1st Qu.MedianMean3rd Qu.Max.Rank of 1st Qu.
0.9−0.10.5−0.10.11.70321.81421.84731.84681.87802.01851
0.9−0.10.9−0.50.11.71831.81671.85041.85081.88222.00632
0.9−0.10.500.11.71511.81731.84681.84761.87811.98923
0.900.9−0.50.51.67882.23982.41392.43152.62083.2917566
0.500.5011.73492.24102.38452.38652.52763.1189567
0.5−0.10−0.911.88882.24132.32072.32332.40492.8344568
0.9−0.90.9−0.537.860411.990113.768514.132615.595023.62381278
0.9−0.90.9−0.938.030812.048314.139814.327116.133823.06271279
0.5−0.90.9−0.936.887812.560514.043314.361116.342224.01621280
Ordered by Median
p1p2q1q2σMin.1st Qu.MedianMean3rd Qu.Max.Rank of Median
0−0.10.100.11.80161.82801.83431.83461.84131.86941
0.10000.11.79111.82701.83431.83461.84181.88072
0−0.10.1−0.10.11.80041.82801.83441.83471.84161.87483
0−0.50.9−0.111.91802.28982.37982.38512.47722.8140583
0.500.5011.73492.24102.38452.38652.52763.1189584
0.5−0.10.5−0.911.92442.29572.38942.39112.48742.9116585
0.9−0.90.9−0.537.860411.990113.768514.132615.595023.62381278
0.5−0.90.9−0.936.887812.560514.043314.361116.342224.01621279
0.9−0.90.9−0.938.030812.048314.139814.327116.133823.06271280
Ordered by Mean
p1p2q1q2σMin.1st Qu.MedianMean3rd Qu.Max.Rank of Mean
0.100−0.10.11.80021.82761.83451.83451.84151.87691
0−0.10.100.11.80161.82801.83431.83461.84131.86942
0−0.1000.11.80341.82841.83461.83461.84061.86593
0−0.50.9−0.111.91802.28982.37982.38512.47722.8140581
0.500.5011.73492.24102.38452.38652.52763.1189582
0.5−0.10.5−0.911.92442.29572.38942.39112.48742.9116583
0.9−0.90.9−0.537.860411.990113.768514.132615.595023.62381278
0.9−0.90.9−0.938.030812.048314.139814.327116.133823.06271279
0.5−0.90.9−0.936.887812.560514.043314.361116.342224.01621280
Ordered by 3rd Qu.
p1p2q1q2σMin.1st Qu.MedianMean3rd Qu.Max.Rank of 3rd Qu.
0−0.10−0.10.11.80541.82881.83461.83461.84041.86451
0.1−0.10−0.50.11.80591.82971.83551.83561.84151.86572
0−0.1000.11.80341.82841.83461.83461.84061.86593
0−0.50.1−0.912.08272.45692.56802.57182.68403.1144593
0.500.5011.73492.24102.38452.38652.52763.1189594
0.1−0.50.5−0.511.95882.38202.48202.48802.58203.1269595
0.9−0.90.9−0.537.860411.990113.768514.132615.595023.62381278
0.5−0.90−0.936.847010.669611.922212.685414.446223.87601279
0.5−0.90.9−0.936.887812.560514.043314.361116.342224.01621280
Ordered by Max.
p1p2q1q2σMin.1st Qu.MedianMean3rd Qu.Max.Rank of Max.
0.9−0.1000.51.46611.91772.03062.02972.13542.50831
0.9−0.10.1−0.10.51.50461.93682.03152.03322.13662.46372
0.9−0.10.9−0.10.51.50492.14602.30432.30512.46452.95833
0.500.1−0.90.51.73471.90581.95001.95231.99602.1738182
0.500.5011.73492.24102.38452.38652.52763.1189183
0.100.5−0.111.73522.06752.16392.16532.25852.8228184
0.9−0.90.9−0.938.030812.048314.139814.327116.133823.06271278
0.5−0.90.5−0.938.564811.283413.041513.158414.584520.13271279
0.1−0.90.1−0.938.725311.555912.943413.106714.688918.58141280
Table 13. Ranking of Model0.5,0,0.5,0,1 across different summary statistics.
Table 13. Ranking of Model0.5,0,0.5,0,1 across different summary statistics.
Rank of Min.Rank of 1st Qu.Rank of MedianRank of MeanRank of 3rd Qu.Rank of Max.
183567584582594183
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yoo, J.; Moon, J. Bayesian Model Selection for Addressing Cold-Start Problems in Partitioned Time Series Prediction. Mathematics 2024, 12, 2682. https://doi.org/10.3390/math12172682

AMA Style

Yoo J, Moon J. Bayesian Model Selection for Addressing Cold-Start Problems in Partitioned Time Series Prediction. Mathematics. 2024; 12(17):2682. https://doi.org/10.3390/math12172682

Chicago/Turabian Style

Yoo, Jaeseong, and Jihoon Moon. 2024. "Bayesian Model Selection for Addressing Cold-Start Problems in Partitioned Time Series Prediction" Mathematics 12, no. 17: 2682. https://doi.org/10.3390/math12172682

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop