Next Article in Journal
Leakage Benchmarking for Universal Gate Sets
Previous Article in Journal
Electrodynamics of Superconductors: From Lorentz to Galilei at Zero Temperature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Exchange-Traded Fund Price Predictions: Insights from Information-Theoretic Networks and Node Embeddings

Department of Industrial and Systems Engineering, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(1), 70; https://doi.org/10.3390/e26010070
Submission received: 31 October 2023 / Revised: 2 December 2023 / Accepted: 22 December 2023 / Published: 12 January 2024
(This article belongs to the Special Issue Information Theory-Based Approach to Portfolio Optimization)

Abstract

:
This study presents a novel approach to predicting price fluctuations for U.S. sector index ETFs. By leveraging information-theoretic measures like mutual information and transfer entropy, we constructed threshold networks highlighting nonlinear dependencies between log returns and trading volume rate changes. We derived centrality measures and node embeddings from these networks, offering unique insights into the ETFs’ dynamics. By integrating these features into gradient-boosting algorithm-based models, we significantly enhanced the predictive accuracy. Our approach offers improved forecast performance for U.S. sector index futures and adds a layer of explainability to the existing literature.

1. Introduction

Classifying listed companies into sectors offers a comprehensive perspective that is pivotal for constructing financial portfolios. The significance of the sector index as a primary boundary in classification systems, encompassing industry groups and sub-industries, is well established in the literature. This sectorial approach grants investors unique advantages, allowing them to harness specific sector opportunities, balance underrepresented sectors, and adopt active investment strategies while considering macroeconomic shifts, momentum, and other crucial factors [1].
However, while many studies have delved into sector indices and ETFs, Leung and Zhao (2021) rightly observed a research gap: there is a limited exploration of the price dynamics comparison between sector indices and ETFs. Research examining the nonlinear dependencies and causalities within sector indices, especially from an information-theoretic viewpoint, remains sparse [2].
Our contribution thus stands at the intersection of these identified gaps. The novelty in our approach lies in two key aspects:
We provide a fresh perspective by delving into the nonlinear network topologies of sector indices in the U.S. market, a dominant global player. This exploration unravels intricate nonlinear information exchanges and flows, showcasing a departure from traditional linearity-based studies.
Also, our study harnesses the potential of these insights to craft predictive models. We tap into their unique predictive power by utilizing a range from short-term (20-day window) to long-term (240-day window) nonlinear dependency and causal networks, incorporating centrality-based measures and novel node embedding techniques.
The manuscript is methodically structured, offering readers a clear progression through our research. Section 2 begins by setting our study within the context of the existing literature and also presenting our dataset. Section 3 offers a thorough explanation of our research methodology, which is intricately linked to the statistical properties of our data. In Section 4, we embark on the construction and analysis of networks, focusing on U.S. sector indices’ returns across varied time frames, while diving deeply into the inherent nonlinear dependencies and causalities. Section 5 serves as a reflective space, discussing the real-world implications of our findings, especially in terms of node-level network measures, the innovative use of node embeddings, and the demonstrable success of our predictive models. Finally, Section 6 wraps up our discourse, providing concluding thoughts on our contributions and hinting at promising avenues for future research.

2. Data and Methodology

2.1. Prior Research

In finance, sector and industry research incorporates varied macroeconomic views. Krause and Tse (2013) identified a consistent pattern of price discovery moving from the U.S. to Canada in securities, with bidirectional volatility spillovers [3]. Hernandez et al. (2015) pointed out a one-way time lag relationship between the IT-computer and medicine-biotechnology sectors, with the former influencing the latter’s price and return values [4]. Dutta (2018) revealed a persistent correlation between oil and stock market volatility indices [5]. Shahzad et al. (2018) used the Q-Q method to demonstrate an asymmetric negative relationship between credit and markets across industries, dependent on the nature and extent of stock market shocks [6]. Khan et al. (2020) confirmed short- and medium-term causality between the Financial and Economic Attitudes Revealed by Search (FEARS) and stock returns, with a stronger correlation in specific sectors [7]. Matos et al. (2021) discovered that early pandemic death cycles in Italy, followed by similar patterns globally, were indicators of disruptions in the U.S. stock market, with the energy sector reacting first to the pandemic [8]. Wan et al. (2021) suggested that solid media sentiment towards a particular company in an industry could lead to a significant shift in media sentiment towards related companies [9]. Shahzad et al. (2021) found that network structures and spillovers greatly vary with market conditions [10]. Choi and Kim conducted an empirical analysis on politically-themed stocks in South Korea, creating networks based on these stocks influenced by political figures [11]. Jin and Guo (2021) showed that since 2013, specific sector indices like consumption, industry, and real estate have been leading corresponding macroeconomic variables in the U.S. stock market [12]. Finally, Mensi (2022) found that sectors like oil, gold, financials, utilities, communication services, consumer staples, and healthcare are net recipients of spillovers, while other sectors are net contributors, determined through methods like the time-frequency spillover method, wavelet method, and the DCC-GARCH model [13].
In financial product price and return prediction, Beer et al. (2020) delved into the innovative territory of deep quantum neural networks for financial forecasting [14]. Chen, Zhang, and Lou (2020) developed a sophisticated model for stock price prediction using a hybrid deep learning approach, combining an attention mechanism with a multilayer perceptron and a bidirectional long short-term memory neural network [15]. Chen and Zhou (2020) introduced a stock prediction model that synergizes genetic algorithm feature selection with a long short-term memory neural network [16]. Althelaya, Mohammed, and El-Alfy (2021) enhanced stock market forecasting by integrating deep learning with a multiresolution analysis, demonstrating the efficacy of combining varied analytical methods [17]. Aldhyani and Alzahrani (2022) crafted a novel framework using deep learning techniques, marking a significant stride in computational finance [18].
Building upon the foundation established by the research, our study carves out a distinctive niche in sector- and industry-based research. While previous studies have predominantly operated within a linear paradigm, our work diverges by incorporating a nonlinear perspective. We not only challenge the prevalent focus on linear correlations or causalities between sectors and industries, but also enrich the discourse by introducing and utilizing the network theory. By adopting this approach, we delve deeper into the intricate nonlinear connections and dependencies within and between sectors. Furthermore, our integration of the network theory offers a fresh lens through which we can decode the complex interplay and dynamics of these sectors. Thus, our research expands the analytical boundaries of sector-based financial research and provides a novel toolkit to better understand and predict sector behaviors and interactions in today’s intricate financial landscape.

2.2. Data

We used Standard and Poors’ U.S. sector index ETFs’ price and trading volume data as the experimental data. There are eleven sectors in the sector Standard and Poor’s Depository Receipt (SPDR) ETFs: Materials (XLB), Communications Services (XLC), Energy (XLE), Financials (XLF), Industrials (XLI), Technology (XLK), Consumer Staples (XLP), Real Estate (XLRE), Utilities (XLU), Health Care (XLV), and Consumer Discretionary (XLY). In this paper, we only used the data from nine sector indices (XLB, XLE, XLF, XLI, XLK, XLP, XLU, XLV, and XLY), because those nine sector indices, from the beginning, were listed on the New York Stock Exchange (NYSE) Arca Exchange on 16 December 1998, but the other two were not. XLRE was listed on 8 October 2015 at NYSE Arca Exchange, and XLC was listed on 19 June 2018 at NYSE Arca Exchange. Those two indices have relatively fewer data; thus, we cannot conduct a network analysis and predict their fluctuation at the same condition. The experimental period is from January 2010 to September 2022. This experimental period is twelve years and nine months long, including a total of 51 quarters and 153 months. In the case of the predictive experiment, we designated the period from January 2010 to December 2018 (about 70% of the whole dataset) as the training and validation set and designated the period from January 2019 to September 2022 (about 30% of the whole dataset) as the testing set. The target data are only the return data of the nine U.S. sector index ETFs. Table 1 provides an overview of the datasets employed throughout our research.

3. Methodology

Based on Shannon’s entropy [19] concept, mutual information and transfer entropy serves as a nonparametric methodology to verify information exchange between pairs of variables.
Data assumptions such as normality, stationarity, and linearity should be preceded by general dependencies and causal relationships represented like Granger causality [20,21]. However, it is known that the nature of stock return-based data usually only satisfies some of these properties. Therefore, we used theories of econopyhsics and information theories that can be used without the above assumptions. To use these theories, we can consider nonlinear relationships between objectives to measure dependencies and causal relationships. Accordingly, we used the concept of mutual information (MI), first proposed by Shannon, and transfer entropy (TE), proposed by Schreiber (2000) [22]. These are entropy-based measures. Specifically, in this study, we used normalized mutual information (NMI) and transfer entropy (TE) based on Shannon entropy with a permutation test for threshold network construction (Boba et al., 2015) [23].

3.1. Mutual Information

3.1.1. Mutual Information

Mutual Information (MI) is a measure that captures the shared information between two variables, indicating their statistical interdependence. In the field of information theory, the behavior of a system, say System X, is understood through its probability distribution p ( x ) and logarithm value of p ( x ) . Based on this idea, the Shannon entropy is as follows:
H ( X ) = x X p ( x ) l o g 2 p ( x ) .
Shannon entropy quantifies the information required to identify random values from a discrete distribution. When two subsystems, X and Y, are present in a state of the system, their combined probability distribution is represented by a joint probability.
H X , Y = x X , y Y p x , y l o g 2 p x , y .
Finally, we can define MI as the quantity of identifying the interaction between subsystems.
I ( X , Y ) = H ( X ) + H ( Y ) H ( X , Y )
Mutual Information (MI) has been widely utilized in finance for network analysis across various stock exchanges. It’s been instrumental in developing networks and selecting portfolios, mainly using short-term data in different markets, and examining market behaviors during significant changes or events. This approach provides insights into market dynamics and investor sentiments in diverse economic contexts [24,25,26,27,28,29,30].

3.1.2. Normalized Mutual Information (NMI)

One of MI’s disadvantages is that it is hard to compare results from the MI derived from different data. Because the domain of MI is always finite for the discrete random variables, the maximum value of the MI is not constant. In other words, this means that it is hard to compare the statistical dependence derived from different datasets. Therefore, we used NMI to compare politically themed stock networks within the same range [ 0 , 1 ] . Since there are several normalized variants of NMI, their properties are slightly different. In this study, we used NMI with a minimum of two entropies, as shown in (4), because the normalization version of mutual information measures should be based on the least upper bound, m i n ( H ( X ) , H ( Y ) ) . Using NMI with a minimum of two entropies ensures that the maximum attainable value of NMI is one. This version of NMI is irrespective of the dimensions of two discrete variables and the marginal probabilities [31,32,33,34,35] (Kvålseth, 1987; Banerjee et al., 2005; Kraskov et al., 2005; Vinh et al., 2010; Sarhrouni et al., 2012; Kvålseth, 2017).
N M I ( X , Y ) = I ( X , Y ) m i n ( H ( X ) , H ( Y ) )

3.2. Transfer Entropy (TE)

Transfer Entropy (TE), based on Shannon entropy and mutual information concepts, is a non-parametric metric for quantifying information transfer between two variables. Unlike Granger causality, which is prediction-oriented, TE focuses on reducing uncertainty, measuring how one variable clarifies the future of another beyond its own past contributions. TE stands out as a model-free approach for identifying causal links in dynamic systems, particularly useful in finance for analyzing connections between various financial entities and market dynamics. This method is renowned for its ability to efficiently pinpoint sources and targets in causal relationships [36].
Transfer Entropy has been a key tool in financial research to explore causal relationships. Studies have examined the connections between credit default swap and bond markets, the causal links among international financial firms, the interplay between exchange rates and stock prices in emerging economies, and the information flow in U.S. equity and commodity markets. This method has proven effective in understanding both internal and cross-market dynamics, demonstrating its versatility in different financial contexts [37,38,39,40].
Based on the concepts mentioned earlier related to entropy, conditional entropy quantifies the amount of information needed to describe the outcome of a random variable, X, given that the value of another random variable, Y, is known. Here, the conditional entropy of X given Y can be expressed as
H X | Y = x X , y Y p x , y l o g 2 p ( x , y ) p ( y ) .
It can be interpreted as the uncertainty about Y when X is known, or as the expected number of bits needed to describe X when Y is known to both the encoder and the decoder. Based on the above definition, we can define the general form of k , l -history TE between two time series, X t and Y t , for x t ( k ) = ( x t , , x t k + 1 ) and y t ( l ) = ( y t , , y t l + 1 ) . The general k , l -history transfer entropy can be expressed as follows (Bossomaier et al., 2016):
T E Y X ( k , l ) t = H X t + 1 X t , , X t k + 1 H X t + 1 | X t , , X t k + 1 , Y t , , Y t l + 1 = i p x t + 1 , x t k , y t l l o g 2 p x t + 1 x t k , y t l ) i p x t + 1 , x t k , y t l l o g 2 p x t + 1 x t k ) = i p ( x t + 1 , x t ( k ) , y t ( l ) ) l o g 2 p ( x t + 1 | x t ( k ) , y t ( l ) ) p ( x t + 1 | x t ( k ) ) ,
where i = x t + 1 , x t k , y t l . T E Y X ( k , l ) t is non-negative, and we can drop the time dependency argument, t , for stationary processes. T E Y X ( k , l ) t represents the information about the future state of X I , which can be obtained by subtracting information retrieved from only X t ( k ) from the information gathered from both X t ( k ) and Y t ( l ) . The schematic representation of transfer entropy is shown in Figure 1.
In this study, we focused on the TE under the following conditions of two lags, k = l = 1 . These settings for lags are typically chosen as they align with the principles of the weak form of the Efficient Market Hypothesis (EMH) and the notion that stock prices follow a random walk pattern. [39,40]. Then, we can express the equation of (1,1)-history TE as follows:
T E Y X ( 1,1 ) t = i p x t + 1 , x t , y t l o g 2 p x t + 1 x t , y t p x t + 1 x t = i p ( x t + 1 , x t , y t ) l o g 2 p ( x t + 1 , x t , y t ) p ( x t ) p ( x t + 1 , x t ) p ( x t , y t ) ,
where i = x t + 1 , x t , y t .

3.3. Test for Obtaining p-Values of MI and TE

In recent research, transfer entropy has often been explored without considering the finite-size effects arising from sample variations. In this study, we adopt a nuanced approach by integrating the permutation test alongside transfer entropy to effectively mitigate these finite-size effects [23,36]. The strength of permutation tests lies in their non-parametric nature, necessitating minimal assumptions and making them particularly adept at discerning statistical significance, especially when the data-generating mechanism remains elusive. To ensure robustness, we introduced a shuffling mechanism for each element in the time series, deriving mutual information and transfer entropy complemented by p-values. Our methodological choice of conducting 1000 shuffles (M = 1000) to compute ETE values underscores our commitment to precision. This novel integration, encapsulating both permutation tests and transfer entropy, is our distinct contribution to the literature, offering insights with heightened accuracy.

3.4. Network Analysis

Network analysis offers a unique lens through which we can gain insights into complex social phenomena by representing them as interconnected systems. This approach not only simplifies the depiction of interactions but also provides a structured framework for understanding the intricate dynamics between different entities. In the context of our research, we view the entropy-based causal relationships among exchange rates as a web of interactions between sector ETFs. By doing so, we unlock the potential to visualize and delve deeper into the intricate ties binding these sector ETFs.
For our study, we leveraged various network topology metrics, transforming them into features for our predictive models. These metrics serve as crucial indicators, helping us navigate the vast network landscape and understand our dataset’s underlying patterns.

3.4.1. Network Theory

Nodes and links are the fundamental components of network theory. As a subject component of a network, a node functions as an interactive agent. A link or connection between two subjects is also an interaction between them. The network type can be classified according to the connection’s characteristics. Networks can be classified as directional or non-directional depending on whether they have a direction. Moreover, if the network is weighted, it is known as a weighted network; otherwise, it is known as a binary network.
Graphs and matrices are used to represent networks. Both systems have advantages in terms of mathematical processing and visual explanation. Using a graph format is a means to depict a network and intuitively show its shape by giving nodes and links shapes, colors, sizes, labels, and arrows. A matrix format describes network attributes as a matrix, often called adjacency matrix.
We used the weighted-undirected graph for MI and the directed-weighted graph for TE. Then, we used the p-value from the permutation tests as a threshold for deciding connections between two U.S. sector index ETFs’ log-returns and trading volumes’ rates of changes. In other words, the p-values of MI and TE determine the linkages in the created directed weighted network. Our p-value-based threshold value was 0.1, one of the conventional statistics values.
In detail, each edge u , v E was attributed a w u , v calculated from one of our statistical dependency measures and an associated p-value, p u , v .
To construct a p-value Threshold Tree, we define a p-value criterion under which an edge is considered statistically significant. For a commonly used significance level of α = 0.05 , our construction rule might be articulated as
E α = { u , v E p u , v < α }

3.4.2. Centrality Measures

Our data were collected from observations within the same system over time. We analyzed topological measures of an evolving network at regular time intervals, yielding time-ordered sequences of topological observations. Our focus was primarily on centrality measures, which are quantified by applying a real-valued function to the vertices of a graph, aiming to rank nodes based on their significance.
The concept of “importance” in a network can be interpreted in various ways, leading to different centrality definitions. One approach conceptualizes importance in terms of network flow or transfer, categorizing centralities based on the type of flow they emphasize. Alternatively, importance can be seen as a node’s contribution to network cohesion, leading to centralities that assess cohesiveness.
Different centrality measures consider the number of paths passing through a node, varying in their definition and counting of relevant paths. This approach allows for a classification spectrum ranging from centralities concerned with short paths (like degree centrality) to those involving longer or infinite paths (such as eigenvector centrality). Other measures, like betweenness centrality, focus on a node’s role in overall network connectivity. The centrality measures used in our study are detailed in Table 2, with their respective citations ranging from [41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68].

3.4.3. Node Embedding Algorithm and Dimensionality Reduction

Understanding the intricate relationships between financial assets is of paramount importance. Traditional approaches, however, often miss the nuanced, nonlinear connections inherent within networks of assets, such as U.S. sector index ETFs. This is where node embeddings come to the fore, offering a fresh perspective that traditional methods may overlook.
Node embeddings convert the intricate attributes and relationships of nodes within a network into representative vectors. These vectors encapsulate essential structural details and node-specific features, furnishing us with enhanced capabilities for analysis and prediction. In our research, we leaned into these advantages, focusing specifically on understanding how features derived from node structures within our constructed network could aid in predicting the movement of nine U.S. sector index ETFs.
We harnessed the power of two notable node embedding techniques: Role2vec [69] and FEATHER [70]. While both techniques excel at capturing essential node information, they differ subtly in their focuses. Role2vec zeroes in on the structural characteristics of nodes, and FEATHER brings to light the attributes specific to each node. By employing both, we ensured a comprehensive grasp of the diverse facets of the network, from its broader architecture to the individual attributes of its constituents.
To refine our approach, we adopted 1024-dimensional embeddings and distilled them into more manageable 32-dimensional vectors. This refinement was achieved using UMAP [61], a technique renowned for preserving global structural information while having its roots firmly planted in Riemannian manifold and topological data analysis.
After deriving the node embeddings in our study, we further integrated them into our prediction framework. Specifically, we based our predictions on networks influenced by both mutual information and transfer entropy. This deliberate integration into the prediction problem enabled us to tap into the information-theoretic network structures.

Role2vec [69]

Ahmed et al. (2018) presented the Role2Vec framework, which employs the flexible concept of attributed random walks and serves as the foundation for leveraging random walks. Their proposed framework expands the applicability of these methods to transductive and inductive learning, as well as graphs with attributes (if available). This is accomplished by acquiring functions that are applicable to new nodes and graphs [69].
Role2vec focuses on learning role-based graph embeddings. The core idea is to learn a mapping of nodes to roles in the graph, and then learn embeddings for these roles.

FEATHER [70]

FEATHER, introduced by Rozemberczki and Sarkar (2020), is a flexible concept of characteristic functions defined on graph vertices to characterize the distribution of vertex features at multiple scales. FEATHER is a computationally efficient algorithm for calculating a particular variant of these characteristic functions in which the probability weights are defined as the transition probabilities of random walks [70].
Rozemberczki and Sarkar (2020) argued that the extracted features are useful for machine learning tasks at the node level. The pooling of these node representations yields compact graph descriptors that can serve as features for graph classification algorithms. They also demonstrated that FEATHER describes isomorphic graphs using the same representation and is resistant to data corruption analytically [70].

UMAP [71]

UMAP is a dimensionality reduction technique that is well suited for visualizing high-dimensional datasets. Developed by McInnes, Healy, and Melville in 2018, UMAP operates based on Riemannian geometry and algebraic topology principles [71].
At its core, UMAP constructs a high-dimensional graph representation of the data and subsequently optimizes a low-dimensional version of this graph to produce a dimension-reduced representation. The method starts by approximating the data’s manifold by using a fuzzy simplicial set. The next step involves finding a low-dimensional representation of this set.
UMAP’s foundational mathematics relies on three main aspects:
  • Topological Data Analysis: Used to understand the high-dimensional structure of the data.
  • Fuzzy Simplicial Sets: Used to approximate the manifold the data resides on, providing both local and global preservation.
  • Riemannian Geometry: Used to accurately measure distances and maintain data relationships.
Our research employed UMAP to condense the 1024-dimensional vectors obtained from our node embeddings down to a more manageable 32-dimensional space. This reduction was imperative not only for visualization but also for enhancing computational efficiency without significantly compromising the structural integrity of our dataset. The robust foundation of UMAP in topological data analysis ensured that the global structure of our data was retained, making the resulting low-dimensional embeddings particularly insightful for subsequent analyses.
Integrating node embeddings is not just a technical addition, but a revolutionary step in bridging the identified research gaps. As our abstract suggests, we aim to bring a layer of explainability that is absent in the existing literature. Node embeddings, especially using Role2vec and FEATHER, allow us to achieve this. We transform abstract financial relationships into tangible, quantifiable data points by converting nodes and their intricate relationships into vectors. This paves the way for integrating these insights into predictive models that forecast with higher accuracy and provide deeper insights into the underlying dynamics.
While many studies have delved into sector indices and ETFs, our adoption of node embeddings elevates our research by emphasizing prediction and understanding. The resulting models are not black boxes, but interpretable tools that shed light on the intricate web of relationships within the U.S. sector index ETFs, marking a significant advancement in financial network analysis.

3.5. Machine Learning Algorithms and xAI (eXplainable Artificial Intelligence) Techniques

We mainly used the most frequently used basic machine learning models to predict index sector ETFs’ returns despite there being many state-of-the-art models with good performance. For example, although many studies have shown that recurrent neural networks and gradient-boosting algorithms based on them perform very well in various areas, this study aims to extract features that can be used in all machine learning models through a nonlinear measure-based network analysis to see their effects. Accordingly, we tried to confirm the performance of well-known machine learning techniques. We used three tree-based machine learning algorithms: XGBoost, LightGBM, and CatBoost [72,73,74]. We used those three prominent models for the following reasons:
  • Interpretability: Tree-based models, at their core, make decisions based on certain conditions, making them more interpretable than many deep learning models. This interpretability is vital in financial sectors, where understanding the reasons behind predictions can be as critical as the predictions themselves.
  • Handling of Mixed Data Types: Financial datasets often consist of numerical and categorical data. Tree-based models like CatBoost are particularly effective at handling categorical variables without extensive preprocessing.
  • Automatic Feature Selection: These models inherently perform feature selection. As a result, they can identify and prioritize the most essential features, which is particularly useful in financial datasets with potentially redundant or less impactful variables.
  • Resistance to Overfitting: With techniques such as gradient boosting and regularization in models like XGBoost and LightGBM, tree models exhibit resistance to overfitting, especially when appropriately tuned.
  • Flexibility: These models can easily capture nonlinear relationships in the data, which is common in financial time series data. Traditional linear models might not capture this nonlinearity as quickly.
  • Efficiency and Scalability: Models like LightGBM and CatBoost have been designed with efficiency in mind. They can handle large datasets, making them suitable for comprehensive financial data.
  • Consistency in Results: While deep learning models like RNNs can be potent, they require more meticulous fine-tuning and can sometimes produce inconsistent results due to their complex architectures. In contrast, tree models, once well-tuned, can provide more consistent predictions.
  • End-to-End Modeling: These models do not necessarily require extensive data preprocessing or normalization, making the modeling process more straightforward and sometimes more accurate since no information is lost in preprocessing.

3.5.1. XGBoost

XGBoost (Chen and Guestrin, 2016) [72] is an algorithm that uses the boosting gradient technique proposed by Friedman (2001) [75]. XGBoost is an ensemble algorithm utilizing gradient tree boosting to enhance classifiers in a sequential manner. Its primary benefit lies in scalability across various situations, making it a highly popular choice for regression tasks.
y ^ i = k = 1 K f k ( x i ) , f k F
Z = i l ( y ^ i , y i ) + k Ω ( f k ) = i l ( y ^ i , y i ) + k ( γ T + 1 2 λ w 2 )
Equation (9) is an expression representing the ensemble model of the tree, F is the collective space of all possible classification and regression trees (CART). At this time, the final prediction is made by summing and comparing the scores of each leaf. Equation (10) is a normalized objective function of the XGBoost model. l ( y ^ i , y i ) is a differentiable convex loss function that measures the difference between the predicted and target values, and it is also a normalization term, and Ω ( f k ) is a CART function that prevents overfitting problems by smoothing the final learned weights by adjusting the complexity of the model. γ T represents the number of leaves in CART, and 1 2 λ w 2 represents the score assigned to the leaves in CART.

3.5.2. Light Gradient Boosting Machine (LightGBM)

LightGBM, developed by Ke et al. in 2017 [73], is a gradient-boosting machine learning model that incorporates gradient-based one-side sampling (GOSS) and exclusive feature bundling (EFB) to handle variables efficiently. Its unique vertical growth structure makes it more accurate and efficient than other machine learning approaches.
Y t = t = 1 T f t ( x )
In Equation (11), f t ( x ) is a tree, and its objective function is estimated using Newton’s method.

3.5.3. CatBoost

CatBoost, introduced by Prokhorenkova et al. in 2018, is a gradient boosting algorithm focused on categorizing data. It stands out in handling categorical features through sequential boosting and decision tree-based techniques. The trees in CatBoost are created by grouping similar instances within the learning dataset, contributing to its superior performance compared to other gradient boosting methods.

3.5.4. SHAP (Shapley Additive Explanation)

Machine learning shows promise in time series prediction but often lacks explanatory power. Addressing this, Lundberg and Lee (2017) introduced the SHAP method, enhancing interpretability across various machine learning models [76]. SHAP, based on the Shapley value from game theory [77], is a key approach in explainable AI (xAI), elucidating predictions by assessing the impact of individual features. This method calculates average Shapley values using game theory principles, clarifying predictions through the contribution of each data feature.
ϕ i = S N i S ! M S 1 ! M ! [ f x S i f x ( S ) ]
ϕ i is the Shapley value for the data, and N is the set of total input variables. S is the set of variables except for the i-th variable in the total input variable, and v ( S ) is the contribution that the remaining subset, except the i-th data, contributed to the result, and f x ( S i ) is the total contribution including the i-th data.
In this study, we also generated equally weighted soft-voting regressors to check the average overall performance and analyzed their mean absolute SHAP values.
I j = 1 N i = 1 N ϕ j ( i )

3.5.5. Performance Metrics of Classification Problem

In addition, we calculated the relationship between the prediction values and real values (fluctuation) using the confusion matrix derived from the classification results. The confusion matrix is typically used to ascertain whether the predicted value was derived appropriately compared with the actual value. In this experiment, the confusion matrix was used to determine the extent to which up or down predictions fall into the fluctuations of U.S. sector index ETFs’ prices.
Figure 2 shows a confusion matrix, and Table 3 further presents the evaluation metric used for the confusion matrix.

4. Results

4.1. Exploratory Data Analysis (EDA)

Table 4 and Table 5 present the descriptive statistics for our two datasets, focusing on log return and the rate of change in trading volume for various securities like XLB, XLE, XLF, and others. For the price data in Table 4, the mean values indicate the average return of each security over the studied timeframe. Most returns are proximate to zero, but XLK and XLY stand out with the highest mean returns of 0.0006. The standard deviation showcases the inherent volatility or risk, with XLE being the most volatile, having a standard deviation of 0.0183, and XLP being the least volatile at 0.0090. The minimum and maximum values capture the extreme returns; XLE saw the largest negative return at −0.2249, while XLF recorded the highest positive return at 0.1524. Quartiles, particularly the medians, often reveal positive returns aligned with the mean values. The skewness of most securities is negative, suggesting that the left tail, or the negative returns, extends more than the right. The kurtosis, which is consistently greater than 3 for all securities, points to a distribution with heavier tails than a normal distribution, implying a higher likelihood of extreme values or outliers.
In Table 5, which covers trading volume data, we observe that XLF, intriguingly, has a negative mean, hinting at a general decreasing trend in its trading volume. On the volatility front, XLP leads with the most volatile trading volume. The skewness values for volume differ across securities, suggesting varied asymmetry in their volume distributions, and the kurtosis indicates that spikes or dramatic drops in trading volumes can occasionally occur.
Several statistical tests were performed, such as the Shapiro–Wilk test, Kolmogorov–Smirnov test, and Jarque–Bera test for normality. The Ljung–Box test was employed to scrutinize the autocorrelation at different orders, and the Augmented Dickey–Fuller (ADF) test was used to check for stationarity. The symbols *, **, and *** denote statistical significance at the 0.1, 0.05, and 0.01 levels. The outcome of this rigorous testing reveals that none of the data columns adhere to normality across both sets of descriptive statistics. This non-compliance with normality furnishes a quantitative foundation for leveraging nonparametric methodologies that do not rest on assumptions like normality.
For the more detailed analysis of non-normality in Table 4 and Table 5, Table 6 and Table 7 delineate the ratios at which the null hypothesis was rejected during normality testing for all generated datasets. Seven normality tests were conducted, including the Shapiro–Wilk test, D’Agostino K-squared test, and others. A striking revelation from these tests is that more than 80% of the dataset used for calculations failed to meet the criteria for normality. This discovery underpins our methodology decision to employ nonlinear nonparametric measures like mutual information and transfer entropy, which remains unfazed by normality prerequisites. The central limit theorem (CLT) theoretically suggests that as the sample size swells, the distribution of sample means should approximate a Gaussian distribution. However, our empirical findings, through exploratory data analysis (EDA), demonstrate a significant trend: as the window length elongates, the null hypothesis becomes rejected more frequently across all normality tests.

4.2. Prediction Results

4.2.1. Prediction Performance

We conducted 100 iterations, changing their seeds to consider the robustness of our experiments. Table 8 shows the average values of our predictions’ performance metrics. The original dataset includes U.S. sector index ETFs’ price-related data.
In our comprehensive analysis, meticulous care was taken to ensure the robustness of our experimental results. To this end, a series of 100 iterations were conducted, each employing a distinct seed, thereby enhancing the reliability and generalizability of our findings.
Presented in Table 8 are the aggregate values derived from our prediction’s performance metrics. This table facilitates a nuanced comparative examination between the performance outcomes obtained from the original dataset, predominantly price-related data from U.S. sector index ETFs, and the outcomes following the integration of advanced features. Specifically, these newly incorporated features are grounded in mutual information (MI) and transfer entropy (TE)-based network embeddings coupled with intricate network topology measures.
The last three columns of Table 8 are dedicated to the Paired T-test results. In evaluating our models, we employed a paired t-test to statistically validate the observed differences between the mean values obtained using different methods. The paired t-test was executed on three machine learning techniques’ results. Depending on the specific category under consideration, we adjusted our hypothesis. For most categories, our alternative hypothesis posited that the mean values in columns like the performance metrics, after including MI and TE network-based columns (network embeddings and network measures), were greater than their corresponding values in the original datasets, respectively. However, for the “Hamming Loss” category, we reversed this hypothesis, testing if the mean values in the original datasets’ performance metrics were greater than those in the performance metrics after including MI and TE network-based columns.
Our results, formatted with significance levels, clearly indicated the differences between the paired columns. Significance levels were denoted with asterisks, where ‘***‘ indicates a 0.01 level significance, ‘**‘ indicates a 0.05 level, and ‘*‘ indicates a 0.1 level. The values with a p-value greater than 0.1 are presented without any asterisks.
In Table 8, a comprehensive evaluation of three machine learning models—XGBoost, LightGBM, and CatBoost—is presented across various metrics and datasets. The datasets under consideration include the original dataset and a refined dataset enriched with Proposed features, specifically the MI and TE-based network-driven features. The table also highlights the results of the paired t-test statistic, which provides insight into the statistical significance of the performance differences observed between the two datasets.
For each sector, such as XLB (Materials) and XLE (Energy), we observed the performance of the three models on both datasets. At a glance, it is evident that the dataset with proposed features often achieves better or comparable results than the original dataset across most sectors and metrics. This indicates that the added features provide valuable information that enhances the model’s performance.
The XGBoost model consistently demonstrates improved performance on the dataset with the proposed features across nearly all sectors. The improvements are especially noticeable in sectors like XLE (Energy) and XLK (Technology), with the significance of this observed improvement reinforced by the paired t-test statistic.
LightGBM, on the other hand, while benefiting from the proposed features in sectors like XLB (Materials) and XLV (Health Care), shows diminished performance in others, such as the XLI (Industrials) sector. This suggests that while the proposed features enhance model robustness, they might introduce noise or redundancy for specific models or sectors.
CatBoost’s performance, compared to the other two models, is generally superior on the proposed dataset, especially in sectors like XLB (Materials) and XLK (Technology). The t-test values further validate the significance of these observations.
Diving deeper into the metrics, accuracy, which provides a general sense of model performance, often shows noticeable improvement when models are trained on the dataset with proposed features. Similarly, balanced accuracy, which provides a more nuanced view, especially in imbalanced datasets, mirrors the trends of regular accuracy. Cohen’s Kappa Coefficient, which assesses the agreement between predicted and actual classifications, significantly improves models like XGBoost in sectors such as XLE (Energy).
Other metrics, such as precision, recall, F1, and F-Beta scores, provide a detailed view of model performance by considering false positives and false negatives. Across most sectors, the dataset with proposed features tends to enhance these metrics for all three models, especially for XGBoost and CatBoost. The Hamming Loss metric, which evaluates the fraction of incorrectly predicted labels, also indicates fewer incorrect predictions for many sectors when using the proposed features.
In conclusion, Table 8 underscores the potential of the proposed MI and TE-based network-driven features in enhancing machine learning model performance across various sectors. While all three models benefit from these features, XGBoost and CatBoost often show the most pronounced improvements. The paired t-test statistics emphasize the value and reliability of incorporating the proposed features into the dataset.

4.2.2. Feature Importance of Causal Network-Related Features

We used the mean absolute SHAP values from the prediction results to elucidate the impact of our causal network-derived measures. Table 9, Table 10, Table 11, Table 12, Table 13 and Table 14 detail the top 20 features determined by the mean absolute SHAP values from the prediction results. The decision to use the top 20 features was guided by an elbow method, revealing a notable difference around the twentieth feature. Therefore, for our nine target ETFs across three gradient boosting algorithms (XGBoost, LightGBM, and CatBoost), a total of 540 features are presented. In detail, 20 (top 20 important features derived from mean absolute SHAP values) × 3 (the number of gradient boosting algorithms) × 9 (the number of target ETFs) features were used for the post-xAI analysis. We conducted a comprehensive analysis of the 540 features in our dataset. These features were systematically categorized to understand their characteristics and relationships better. By grouping them, we could identify patterns, trends, and anomalies more clearly, ensuring a streamlined and efficient data assessment.
In Table 9, of the 540 features, our proposed features constitute 283 (52.41%), while the original features account for 257 (47.59%). Given that the 18 original features are composed of nine log-return columns and nine columns depicting the rate of change in the trading volume, it implies that one feature, on average, appears approximately 14.3 times across the 27 models. Notably, the trading volume contributes nearly as much as price, solidifying its relevance in our study.
Table 10 underscores that roughly 70% of node-derived features stem from the MI network, encapsulating over a third (36.67%) of the 540 features. This suggests the importance of nonlinear mutual dependencies in analyzing the U.S. sector index. Notably, though nonlinear causal relationships are not ubiquitous, they hold significance when they emerge, especially for fluctuation prediction.
Observations from Table 11 demonstrate that features derived from 20-day, 60-day, 120-day, and 240-day windows span short-term to long-term influences on the U.S. sector index ETFs’ fluctuations. Short-term (20-day) features particularly exert substantial influence, more so with trading volume data than price-based log-return data. This indicates that nonlinear dependencies and causality hold consistent importance in gradient-boosting algorithm-based predictions.
Our findings, as detailed in Table 12, indicate that centrality measures and node embeddings both play pivotal roles in predicting U.S. sector index ETFs. Most notably, of the 32 node embedding-based features, 130 make the top 20 based on the mean absolute SHAP values, implying that node embeddings are instrumental in enhancing prediction accuracy.
Table 13 highlights that nearly all centrality measures are among the top 20 in predicting fluctuations in U.S. sector index ETFs. Specifically, second-order centrality, PageRank, and HITS—which all consider relative connections as opposed to direct ones—are predominant. This supports our approach of emphasizing interconnectivity and causality through network analysis.
Conclusively, in Table 14, while Role2vec (a structural node embedding algorithm) represents 56.92% of node embeddings, FEATHER (an attributed node embedding algorithm) is also notable at 43.08%. This paves the way for exploring a broader range of node embeddings in subsequent studies. Additionally, refining the embedding vector dimensions merits further investigation, given its variability based on objectives.

5. Discussion

To understand the intricacies of U.S. sector index ETFs’ price fluctuations, this study embarked on a mission to validate the existence and implications of nonlinear dependencies and causal relationships emanating from log returns and the rate of changes in trading volumes. By incorporating sophisticated techniques such as mutual information, transfer entropy with permutation tests, and threshold networks, we sought insights from the intricate web of relationships that govern ETFs’ price dynamics. Our methodological approach led to the construction of undirected and directed weighted networks, illuminating the nonlinear dependencies and causal dynamics between the U.S. sector index ETFs.
A significant revelation of our research was the potential to enhance the predictive prowess of machine learning models using centrality measures and node embeddings derived from the topology of these constructed networks. By integrating these measures into our prediction models, not only did we achieve superior forecasting accuracy, but we also heightened the explanatory capabilities of our models. The visual representation of information-theoretic dependency and information transfer through the networks further underscored the relationships and dependencies between the ETFs, offering a richer understanding of their interconnected dynamics.
A cornerstone of our analysis was using SHAP, rooted in Shapley values, to quantify the effectiveness of our centrality measures and node embeddings in forecasting the returns of U.S. sector index ETFs. The insights gleaned from analyzing the mean of absolute SHAP values were revelatory. They affirmed that features derived from our information-theoretic networks, especially those grounded in various temporal windows ranging from short-term (20-day) to long-term (240-day) windows, were pivotal in forecasting the fluctuations of U.S. sector index futures. This not only bolstered the case for employing log returns and trading volumes’ rates of changes as reliable measures for capturing mutual information and transferring entropy across diverse temporal windows, but also solidified their utility in crafting a plethora of networks and subsequently harnessing their node-level properties.
Our study illuminated the path forward in forecasting U.S. sector index futures. By harnessing the power of nonlinear measure-based networks and their node-derived features, we not only refined the predictive accuracy of our models but also enriched their explainability. This study’s revelations underscore these techniques’ promises and set the stage for future explorations in this domain.

6. Conclusions

In our quest to understand and predict the dynamics of U.S. stock market sector indices, our study introduced a unique perspective by employing a nonlinear, nonparametric measure-based complex network approach. Harnessing the nuanced insights from information entropy-based measures, we shed light on the intricate information-theoretic relationships embedded within U.S. stock market sector indices’ price and trading volume data. By delving into these nonlinear dependencies and causal relationships, we ventured into an under-explored domain and offered a fresh lens for forecasting U.S. sector index ETF prices.
Some of the distinct contributions of our study include the following:
  • The utilization of information entropy-based measures to discern and showcase the underlying relationships in U.S. stock market sector indices.
  • The illustration of nonlinear dependencies and causal relationships in the U.S. market sector index networks, which is an area that has yet to be deeply probed.
  • The revelation that nonlinear dependencies and causal relationships can significantly contribute to predictive models, shedding light on new avenues in market forecasting.
  • Empirical evidence supports using return-based data to bolster prediction results by probing into the intricate webs formed by return and trading volume networks. This offers a promising direction in enhancing data efficiency by leveraging inter-sectoral relationships without additional external features.
Despite our contribution, we recognize that there is always room for growth and improvement. For starters, diving deeper into the vast world of machine learning and graph embedding techniques could sharpen our analysis. These advanced tools might offer a more transparent lens to view our predictions. We also acknowledge that perfecting our models and tweaking their internal settings or “hyperparameters” could potentially enhance the accuracy of our sector forecasts. Moreover, while our focus has been specific, bringing in data like macroeconomic variables could offer a fuller picture and add depth to our findings.
While our research has added value to the financial sector’s understanding, it is just one step in a longer journey. We are excited about future studies and the potential to delve even deeper into the intricacies of financial forecasting.

Author Contributions

W.C.K.: Conceptualization and project administration. I.C.: Methodology, software, formal analysis, and visualization. I.C.: Data curation. W.C.K.: Supervision. I.C.: Writing—original draft preparation and editing. W.C.K.: Writing—review. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://finance.yahoo.com/ (accessed on 1 October 2023).

Conflicts of Interest

All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or non-financial interest in the subject matter or materials discussed in this manuscript.

References

  1. MSCI. Sector Indexes—Expanding Investor’s Toolkit. 2022. Available online: https://www.msci.com/our-solutions/indexes/sector-indexes (accessed on 30 October 2023).
  2. Leung, T.; Zhao, T. Multiscale Decomposition and Spectral Analysis of Sector ETF Price Dynamics. J. Risk Financ. Manag. 2021, 14, 464. [Google Scholar] [CrossRef]
  3. Krause, T.; Tse, Y. Volatility and return spillovers in Canadian and US industry ETFs. Int. Rev. Econ. Financ. 2013, 25, 244–259. [Google Scholar] [CrossRef]
  4. Hernandez, J.A.; Al Janabi, M.A.; Hammoudeh, S.; Khuong Nguyen, D. Time lag dependence, cross-correlation and risk analysis of US energy and non-energy stock portfolios. J. Asset Manag. 2015, 16, 467–483. [Google Scholar] [CrossRef]
  5. Dutta, A. Oil and energy sector stock markets: An analysis of implied volatility indexes. J. Multinatl. Financ. Manag. 2018, 44, 61–68. [Google Scholar] [CrossRef]
  6. Shahzad, S.J.H.; Bouri, E.; Kristoufek, L.; Saeed, T. Impact of the COVID-19 outbreak on the US equity sectors: Evidence from quantile return spillovers. Financ. Innov. 2021, 7, 1–23. [Google Scholar] [CrossRef]
  7. Khan, M.A.; Hernandez, J.A.; Shahzad, S.J.H. Time and frequency relationship between household investors’ sentiment index and US industry stock returns. Financ. Res. Lett. 2020, 36, 101318. [Google Scholar] [CrossRef]
  8. Matos, P.; Costa, A.; da Silva, C. COVID-19, stock market and sectoral contagion in US: A time-frequency analysis. Res. Int. Bus. Financ. 2021, 57, 101400. [Google Scholar] [CrossRef] [PubMed]
  9. Wan, X.; Yang, J.; Marinov, S.; Calliess, J.P.; Zohren, S.; Dong, X. Sentiment correlation in financial news networks and associated market movements. Sci. Rep. 2021, 11, 1–12. [Google Scholar] [CrossRef]
  10. Shahzad, S.J.H.; Mensi, W.; Hammoudeh, S.; Balcilar, M.; Shahbaz, M. Distribution specific dependence and causality between industry-level US credit and stock markets. J. Int. Financ. Mark. Inst. Money 2018, 52, 114–133. [Google Scholar] [CrossRef]
  11. Choi, I.; Kim, W.C. Detecting and Analyzing Politically-Themed Stocks Using Text Mining Techniques and Transfer Entropy—Focus on the Republic of Korea’s Case. Entropy 2021, 23, 734. [Google Scholar] [CrossRef]
  12. Jin, Z.; Guo, K. The dynamic relationship between stock market and macroeconomy at sectoral level: Evidence from Chinese and US stock market. Complexity 2021, 2021, 1–16. [Google Scholar] [CrossRef]
  13. Mensi, W.; Al Rababa’a, A.R.; Alomari, M.; Vo, X.V.; Kang, S.H. Dynamic frequency volatility spillovers and connectedness between strategic commodity and stock markets: US-based sectoral analysis. Resour. Policy 2022, 79, 102976. [Google Scholar] [CrossRef]
  14. Beer, K.; Bondarenko, D.; Farrelly, T.; Osborne, T.J.; Salzmann, R.; Scheiermann, D.; Wolf, R. Training deep quantum neural networks. Nat. Commun. 2020, 11, 808. [Google Scholar] [CrossRef] [PubMed]
  15. Chen, Q.; Zhang, W.; Lou, Y. Forecasting stock prices using a hybrid deep learning model integrating attention mechanism multilayer perceptron and bidirectional long-short term memory neural network. IEEE Access 2020, 8, 117365–117376. [Google Scholar] [CrossRef]
  16. Chen, S.; Zhou, C. Stock prediction based on genetic algorithm feature selection and long short-term memory neural network. IEEE Access 2020, 9, 9066–9072. [Google Scholar] [CrossRef]
  17. Althelaya, K.A.; Mohammed, S.A.; El-Alfy, E.S.M. Combining deep learning and multiresolution analysis for stock market forecasting. IEEE Access 2021, 9, 13099–13111. [Google Scholar] [CrossRef]
  18. Aldhyani, T.H.; Alzahrani, A. Framework for predicting and modeling stock market prices based on deep learning algorithms. Electronics 2022, 11, 3149. [Google Scholar] [CrossRef]
  19. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  20. Granger, C.W. Investigating causal relations by econometric models and cross-spectral methods. Econom. J. Econom. Soc. 1969, 37, 424–438. [Google Scholar] [CrossRef]
  21. Geweke, J. Measurement of linear dependence and feedback between multiple time series. J. Am. Stat. Assoc. 1982, 77, 304–313. [Google Scholar] [CrossRef]
  22. Schreiber, T. Measuring information transfer. Phys. Rev. Lett. 2000, 85, 461. [Google Scholar] [CrossRef]
  23. Boba, P.; Bollmann, D.; Schoepe, D.; Wester, N.; Wiesel, J.; Hamacher, K. Efficient computation and statistical assessment of transfer entropy. Front. Phys. 2015, 3, 10. [Google Scholar] [CrossRef]
  24. Fiedor, P. Networks in Financial Markets Based on the Mutual Information Rate. Phys. Rev. E 2014, 89, 052801. [Google Scholar] [CrossRef]
  25. You, T.; Fiedor, P.; Holda, A. Network Analysis of the Shanghai Stock Exchange Based on Partial Mutual Information. J. Risk Financ. Manag. 2015, 8, 266–284. [Google Scholar] [CrossRef]
  26. Barbi, A.Q.; Prataviera, G.A. Nonlinear Dependencies on Brazilian Equity Network From Mutual Information Minimum Spanning Trees. Phys. A Stat. Mech. Its Appl. 2019, 523, 876–885. [Google Scholar] [CrossRef]
  27. Han, D. Network Analysis of the Chinese Stock Market during the Turbulence of 2015–2016 Using Log–Returns, Volumes and Mutual Information. Phys. A Stat. Mech. Its Appl. 2019, 523, 1091–1109. [Google Scholar]
  28. Jiang, J.; Shang, P.; Li, X. An Effective Stock Classification Method via MDS Based on Modified Mutual Information Distance. Fluct. Noise Lett. 2020, 19, 2050018. [Google Scholar] [CrossRef]
  29. Yan, Y.; Wu, B.; Tian, T.; Zhang, H. Development of Stock Networks Using Part Mutual Information and Australian Stock Market Data. Entropy 2020, 22, 773. [Google Scholar] [CrossRef]
  30. Lahmiri, S.; Bekiros, S. Rényi Entropy and Mutual Information Measurement of Market Expectations and Investor Fear during the COVID-19 Pandemic. Chaos Solitons Fractals 2020, 139, 110084. [Google Scholar] [CrossRef] [PubMed]
  31. Kvålseth, T.O. Entropy and Correlation: Some Comments. IEEE Trans. Syst. Man Cybern. 1987, 17, 517–519. [Google Scholar] [CrossRef]
  32. Banerjee, A.; Dhillon, I.S.; Ghosh, J.; Sra, S. Clustering on the Unit Hypersphere Using Von Mises-Fisher Distributions. J. Mach. Learn. Res. 2005, 6, 1345–1382. [Google Scholar]
  33. Kraskov, A.; Stögbauer, H.; Andrzejak, R.G.; Grassberger, P. Hierarchical Clustering Using Mutual Information. Europhys. Lett. 2005, 70, 278. [Google Scholar] [CrossRef]
  34. Vinh, N.X.; Epps, J.; Bailey, J. Information Theoretic Measures for Clusterings Comparison: Variants, Properties, Normalization and Correction for Chance. J. Mach. Learn. Res. 2010, 11, 2837–2854. [Google Scholar]
  35. Kvålseth, T.O. On Normalized Mutual Information: Measure Derivations and Properties. Entropy 2017, 19, 631. [Google Scholar] [CrossRef]
  36. Dimpfl, T.; Peter, F.J. Using Transfer Entropy to Measure Information Flows between Financial Markets. Stud. Nonlinear Dyn. Econom. 2013, 17, 85–102. [Google Scholar] [CrossRef]
  37. Sandoval, L. Structure of a Global Network of Financial Companies Based on Transfer Entropy. Entropy 2014, 16, 4443–4482. [Google Scholar] [CrossRef]
  38. Sensoy, A.; Sobaci, C.; Sensoy, S.; Alali, F. Effective Transfer Entropy Approach to Information Flow between Exchange Rates and Stock Markets. Chaos Solitons Fractals 2014, 68, 180–185. [Google Scholar] [CrossRef]
  39. Bekiros, S.; Nguyen, D.K.; Junior, L.S.; Uddin, G.S. Information Diffusion, Cluster Formation and Entropy-Based Network Dynamics in Equity and Commodity Markets. Eur. J. Oper. Res. 2017, 256, 945–961. [Google Scholar] [CrossRef]
  40. Lim, K.; Kim, S.; Kim, S.Y. Information Transfer Across Intra/Inter-Structure of CDS and Stock Markets. Phys. A Stat. Mech. Appl. 2017, 486, 118–126. [Google Scholar] [CrossRef]
  41. Bron, C.; Kerbosch, J. Algorithm 457: Finding all cliques of an undirected graph. Commun. ACM 1973, 16, 575–577. [Google Scholar] [CrossRef]
  42. Freeman, L.C. A set of measures of centrality based on betweenness. Sociometry 1977, 40, 35–41. [Google Scholar] [CrossRef]
  43. Freeman, L.C. Centrality in social networks conceptual clarification. Soc. Netw. 1978, 1, 215–239. [Google Scholar] [CrossRef]
  44. Bonacich, P. Power and centrality: A family of measures. Am. J. Sociol. 1987, 92, 1170–1182. [Google Scholar] [CrossRef]
  45. Stephenson, K.; Zelen, M. Rethinking centrality: Methods and examples. Soc. Netw. 1989, 11, 1–37. [Google Scholar] [CrossRef]
  46. Wasserman, S.; Faust, K. Social Network Analysis: Methods and Applications; Cambridge University Press: Cambridge, UK, 1994. [Google Scholar]
  47. Kleinberg, J.M. Authoritative sources in a hyperlinked environment. J. ACM 1999, 46, 604–632. [Google Scholar] [CrossRef]
  48. Page, L.; Brin, S.; Motwani, R.; Winograd, T. The PageRank Citation Ranking: Bringing Order to the Web. 1999. Available online: https://www.cis.upenn.edu/~mkearns/teaching/NetworkedLife/pagerank.pdf (accessed on 30 October 2023).
  49. Brandes, U. A faster algorithm for betweenness centrality. J. Math. Sociol. 2001, 25, 163–177. [Google Scholar] [CrossRef]
  50. Barrat, A.; Barthelemy, M.; Pastor-Satorras, R.; Vespignani, A. The architecture of complex weighted networks. Proc. Natl. Acad. Sci. USA 2004, 101, 3747–3752. [Google Scholar] [CrossRef] [PubMed]
  51. Tomita, E.; Tanaka, A.; Takahashi, H. The worst-case time complexity for generating all maximal cliques. In Proceedings of the 10th Annual International Conference, COCOON 2004, Jeju Island, Republic of Korea, 17–20 August 2004; Springer: Berlin/Heidelberg, Germany, 2004; pp. 161–170. [Google Scholar]
  52. Borgatti, S.P. Centrality and network flow. Soc. Netw. 2005, 27, 55–71. [Google Scholar] [CrossRef]
  53. Brandes, U.; Fleischer, D. Centrality measures based on current flow. In Annual Symposium on Theoretical Aspects of Computer Science; Springer: Berlin/Heidelberg, Germany, 2005; pp. 533–544. [Google Scholar]
  54. Langville, A.N.; Meyer, C.D. A survey of eigenvector methods for web information retrieval. SIAM Rev. 2005, 47, 135–161. [Google Scholar] [CrossRef]
  55. Newman, M.E. A measure of betweenness centrality based on random walks. Soc. Netw. 2005, 27, 39–54. [Google Scholar] [CrossRef]
  56. Borgatti, S.P.; Everett, M.G. A graph-theoretic perspective on centrality. Soc. Netw. 2006, 28, 466–484. [Google Scholar] [CrossRef]
  57. Brandes, U.; Pich, C. Centrality estimation in large networks. Int. J. Bifurc. Chaos 2007, 17, 2303–2318. [Google Scholar] [CrossRef]
  58. Brandes, U. On variants of shortest-path betweenness centrality and their generic computation. Soc. Netw. 2008, 30, 136–145. [Google Scholar] [CrossRef]
  59. Cazals, F.; Karande, C. A note on the problem of reporting maximal cliques. Theor. Comput. Sci. 2008, 407, 564–568. [Google Scholar] [CrossRef]
  60. Hagberg, A.; Swart, P.; S Chult, D. Exploring Network Structure, Dynamics, and Function Using NetworkX (No. LA-UR-08-05495; LA-UR-08-5495); Los Alamos National Lab (LANL): Los Alamos, NM, USA, 2008. [Google Scholar]
  61. Estrada, E.; Higham, D.J.; Hatano, N. Communicability betweenness in complex networks. Phys. A Stat. Mech. Its Appl. 2009, 388, 764–774. [Google Scholar] [CrossRef]
  62. Kermarrec, A.M.; Le Merrer, E.; Sericola, B.; Trédan, G. Second order centrality: Distributed assessment of nodes criticity in complex networks. Comput. Commun. 2011, 34, 619–628. [Google Scholar] [CrossRef]
  63. Benzi, M.; Klymko, C. A matrix analysis of different centrality measures. arXiv 2014, arXiv:1312.6722. [Google Scholar]
  64. Boldi, P.; Vigna, S. Axioms for centrality. Internet Math. 2014, 10, 222–262. [Google Scholar] [CrossRef]
  65. Brandes, U.; Borgatti, S.P.; Freeman, L.C. Maintaining the duality of closeness and betweenness centrality. Soc. Netw. 2016, 44, 153–159. [Google Scholar] [CrossRef]
  66. Zhang, J.X.; Chen, D.B.; Dong, Q.; Zhao, Z.D. Identifying a set of influential spreaders in complex networks. Sci. Rep. 2016, 6, 1–10. [Google Scholar] [CrossRef]
  67. Negre, C.F.; Morzan, U.N.; Hendrickson, H.P.; Pal, R.; Lisi, G.P.; Loria, J.P.; Rivalta, I.; Ho, J.; Batista, V.S. Eigenvector centrality for characterization of protein allosteric pathways. Proc. Natl. Acad. Sci. USA 2018, 115, E12201–E12208. [Google Scholar] [CrossRef]
  68. Newman, M. Networks; Oxford University Press: Oxford, UK, 2018. [Google Scholar]
  69. Ahmed, N.K.; Rossi, R.; Lee, J.B.; Willke, T.L.; Zhou, R.; Kong, X.; Eldardiry, H. Learning role-based graph embeddings. arXiv 2018, arXiv:1802.02896. [Google Scholar]
  70. Rozemberczki, B.; Sarkar, R. Characteristic functions on graphs: Birds of a feather, from statistical descriptors to parametric models. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, Virtual Event, 19–23 October 2020; pp. 1325–1334. [Google Scholar]
  71. McInnes, L.; Healy, J.; Melville, J. Umap: Uniform manifold approximation and projection for dimension reduction. arXiv 2018, arXiv:1802.03426. [Google Scholar]
  72. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  73. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process. Syst. 2017, 30, 3146–3154. [Google Scholar]
  74. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; p. 31. [Google Scholar]
  75. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  76. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; p. 30. [Google Scholar]
  77. Shapley, L. Quota Solutions of n-Person Games. 1953, p. 343. Available online: https://www.rand.org/content/dam/rand/pubs/papers/2021/P297.pdf (accessed on 30 October 2023).
Figure 1. Schematic representation of transfer entropy (I. Choi. 2021).
Figure 1. Schematic representation of transfer entropy (I. Choi. 2021).
Entropy 26 00070 g001
Figure 2. Confusion matrix.
Figure 2. Confusion matrix.
Entropy 26 00070 g002
Table 1. Sector ETFs and their descriptions.
Table 1. Sector ETFs and their descriptions.
Full NameAbbreviationTracking IndexSpecialized Description
MaterialsXLBS&P Materials IndexTracks companies in the materials sector, including chemicals, mining, and forestry.
EnergyXLES&P Energy IndexComprises companies in the energy sector, including oil, gas, and renewable energy.
FinancialsXLFS&P Financials IndexRepresents companies in the financial sector, including banks, insurance, and real estate services.
IndustrialsXLIS&P Industrials IndexCovers companies in the industrial sector, such as machinery, aerospace, and defense.
TechnologyXLKS&P Technology IndexRepresents the technology sector, including software, hardware, and electronics.
Consumer StaplesXLPS&P Consumer Staples IndexConsists of companies in the consumer staples sector, like food, beverages, and household goods.
UtilitiesXLUS&P Utilities IndexRepresents utilities sector companies, including electric, gas, and water utilities.
Health CareXLVS&P Health Care IndexCovers companies in the health care sector, including pharmaceuticals, biotech, and health care services.
Consumer DiscretionaryXLYS&P Consumer Discretionary IndexRepresents companies in the consumer discretionary sector, like entertainment, retail, and autos.
Table 2. Centrality measures and their definitions.
Table 2. Centrality measures and their definitions.
Centrality MeasuresCitation
Average Neighbor Degree[50,60]
Degree Centrality[60]
Eigenvector Centrality[44,60,68]
Closeness Centrality[43,46,60]
Information Centrality[45,60,65]
Betweenness Centrality[42,49,57,58,60]
Current Flow Betweenness Centrality[53,55,60]
Communicability Betweenness Centrality[60,61]
Harmonic Centrality[60,64]
Second-Order Centrality[60,62]
Voterank Importance[60,66]
Number of Maximal Cliques[41,51,59]
PageRank[48,60]
HITS[47,54,60]
Table 3. Performance metrics of classification problems.
Table 3. Performance metrics of classification problems.
MetricDefinition
Accuracy T P   +   T N P   +   N
Balanced Accuracy T P / P   +   T N / N 2
Cohen’s Kappa Coefficient 2   ×   ( T P   ×   T N F N   ×   F P ) T P   +   F P   ×   F P   +   T N   +   ( T P   +   F N )   ×   ( F N + T N )
Precision T P T P   +   F P
Recall T P T P + F N
F1 Score 2 T P 2 T P   +   F P   +   F N
F-Beta Score ( 1   +   β 2 )   ×   P r e c i s i o n   ×   R e c a l l β 2   ×   P r e c i s i o n + R e c a l l
Hamming Loss 1 n s a m p l e s × n l a b e l s i = 0 n s a m p l e s 1 j = 0 n l a b e l s 1 1 y ^ i , j y i , j
where y ^ i , j is the predicted value for the j-th label of a given sample i, y i , j is the corresponding true value, n s a m p l e s is the number of samples, n l a b e l s is the number of labels (in this study, n l a b e l s = 2 ) , and 1(x) is the indicator function.
Table 4. Descriptive statistics (price).
Table 4. Descriptive statistics (price).
StatisticsXLBXLEXLFXLIXLKXLPXLUXLVXLY
Mean0.00040.00020.00040.00040.00060.00040.00040.00050.0006
Standard Deviation0.01430.01830.01770.01340.01350.00900.01120.01070.0132
Min.−0.1166−0.2249−0.1807−0.1204−0.1487−0.0987−0.1206−0.1038−0.1355
Max.0.11120.14870.15240.11910.11090.08170.12040.07420.0897
Q1−0.0063−0.0078−0.0065−0.0054−0.0049−0.0036−0.0051−0.0044−0.0050
Median0.00090.00040.00070.00090.00110.00070.00090.00090.0013
Q30.00800.00900.00790.00690.00720.00510.00620.00620.0072
Skewness−0.4846−0.7722−0.0455−0.4728−0.5109−0.5013−0.3185−0.4312−0.6604
Kurtosis6.050213.952116.49348.96469.781713.028116.14777.45027.9877
Shapiro–Wilk Test (W)0.9421 ***0.9073 ***0.837 ***0.9144 ***0.9175 ***0.9033 ***0.8903 ***0.9321 ***0.9227 ***
Jarque–Bera Test (JB)5395.7801 ***28,327.5218 ***39,109.1785 ***11,680.1006 ***13,903.6934 ***24,544.2796 ***37,543.7296 ***8084.7764 ***9421.6981 ***
Augmented Dickey–Fuller ( D F τ )−14.9626 ***−11.5752 ***−10.9766 ***−11.6973 ***−12.9465 ***−18.7009 ***−14.6622 ***−16.2154 ***−11.9238 ***
Ljung–Box Test (Q(5))27.8087 ***21.2132 ***100.1074 ***22.9614 ***70.5875 ***59.3114 ***61.5096 ***42.4347 ***23.582 ***
Ljung–Box Test (Q(10))65.3377 ***60.4795 ***164.8877 ***83.7618 ***190.4879 ***142.5232 ***154.016 ***121.0344 ***78.3964 ***
Ljung–Box Test (Q(20))87.2127 ***86.7658 ***207.5788 ***107.8259 ***237.6269 ***162.8771 ***209.2207 ***161.026 ***100.1244 ***
Table 5. Descriptive statistics (volume).
Table 5. Descriptive statistics (volume).
DataXLBXLEXLFXLIXLKXLPXLUXLVXLY
Statistics
Mean0.00020.0000−0.00030.00010.00020.00050.00050.00050.0001
Standard Deviation0.39120.32710.36760.37300.41870.43960.39180.39580.4108
Min.−1.5326−1.5394−1.3862−1.8871−1.8236−1.7922−1.5858−2.1560−1.6051
Max.1.72351.30461.52831.55981.96782.49721.66301.93972.1229
Q1−0.2523−0.2091−0.2497−0.2495−0.2570−0.2753−0.2428−0.2586−0.2596
Median−0.0080−0.0093−0.00590.0026−0.0064−0.0130−0.0112−0.0114−0.0068
Q30.24710.20550.22930.23570.25040.25340.24520.23760.2450
Skewness0.13010.07910.18920.09800.06560.29730.17100.09940.1642
Kurtosis0.70030.61450.48710.72791.24281.62480.82670.98421.1379
Shapiro–Wilk Test (W)0.9956 ***0.996 ***0.9963 ***0.9956 ***0.9906 ***0.9857 ***0.9935 ***0.9935 ***0.9918 ***
Jarque–Bera Test (JB)79.9351 ***57.6037 ***54.5113 ***81.3655 ***223.9659 ***429.7603 ***114.7139 ***144.4836 ***201.1662 ***
Augmented Dickey–Fuller ( D F τ )−20.5462 ***−15.9105 ***−15.1273 ***−15.3356 ***−16.3168 ***−16.3461 ***−16.415 ***−19.7349 ***−20.9397 ***
Ljung–Box Test (Q(5))439.3993 ***438.0747 ***427.4039 ***442.3769 ***545.3189 ***598.9662 ***632.7569 ***483.8212 ***594.0558 ***
Ljung–Box Test (Q(10))442.1597 ***456.3206 ***431.9172 ***445.1095 ***547.6234 ***618.9684 ***639.1071 ***489.3571 ***599.5767 ***
Ljung–Box Test (Q(20))473.9014 ***482.8198 ***458.9153 ***464.1181 ***558.6023 ***647.9025 ***654.0895 ***494.3574 ***610.6616 ***
Table 6. Rejected ratios of all generated datasets derived from the normality test results (price).
Table 6. Rejected ratios of all generated datasets derived from the normality test results (price).
(Unit: %)Shapiro–Wilk Test (W)D’Agostino K-Squared Test (K-Squared)Lilliefors Test (T)Jarque–Bera Test (JB)Kolmogorov–Smirnov Test (KS)Anderson–Darling Test (A-Squared)Cramér–von Mises Test (U)
XLB
α = 0.191.6991.5489.3391.54100.0092.39100.00
α = 0.0590.2890.0587.0190.53100.0090.60100.00
α = 0.0187.3087.2382.4688.43100.0087.32100.00
XLE
α = 0.193.4192.7289.8992.81100.0092.98100.00
α = 0.0592.2791.6487.6692.10100.0091.22100.00
α = 0.0190.4689.9283.7490.94100.0087.88100.00
XLF
α = 0.194.1193.9991.9693.99100.0093.94100.00
α = 0.0593.0492.7590.3393.16100.0092.37100.00
α = 0.0190.4989.6987.3691.39100.0089.93100.00
XLI
α = 0.193.3493.3092.4893.43100.0094.14100.00
α = 0.0592.0191.9790.5892.58100.0092.86100.00
α = 0.0189.6989.5686.6990.93100.0089.91100.00
XLK
α = 0.195.8895.6593.8395.51100.0095.94100.00
α = 0.0594.8994.6991.8994.83100.0094.80100.00
α = 0.0192.6292.4488.3693.51100.0092.41100.00
XLP
α = 0.194.4794.8691.7494.80100.0094.47100.00
α = 0.0593.1393.8189.4094.15100.0092.89100.00
α = 0.0190.8691.8684.8292.85100.0089.68100.00
XLU
α = 0.190.8391.6588.1291.65100.0089.81100.00
α = 0.0589.1389.9885.6290.46100.0088.06100.00
α = 0.0186.4087.4981.0188.65100.0084.48100.00
XLV
α = 0.193.6193.9190.5193.85100.0093.76100.00
α = 0.0592.5593.0687.9393.3100.0092.30100.00
α = 0.0189.7590.5983.9891.91100.0088.77100.00
XLY
α = 0.195.2295.4892.5495.19100.0094.53100.00
α = 0.0594.0094.3790.4894.39100.0093.41100.00
α = 0.0191.8391.8386.0492.69100.0091.05100.00
Table 7. Rejected ratios of all generated datasets derived from the normality test results (volume).
Table 7. Rejected ratios of all generated datasets derived from the normality test results (volume).
(Unit: %)Shapiro–Wilk Test (W)D’Agostino K-Squared Test (K-Squared)Lilliefors Test (T)Jarque–Bera Test (JB)Kolmogorov–Smirnov Test (KS)Anderson–Darling Test (A-Squared)Cramér–von Mises Test (U)
XLB
α = 0.175.0477.2540.8177.1299.7868.8999.90
α = 0.0570.1673.5725.5574.0199.5461.1099.75
α = 0.0160.5765.451.8768.9798.7446.4199.12
XLE
α = 0.176.6579.4937.7079.4799.9564.8899.99
α = 0.0569.9273.7627.7575.6299.8454.9599.94
α = 0.0150.6450.8311.2762.3699.3435.5699.56
XLF
α = 0.178.1979.6151.5980.0199.8970.3299.96
α = 0.0573.3774.1042.8075.0899.7163.2399.86
α = 0.0161.0063.8625.2166.3799.0650.9399.31
XLI
α = 0.173.7176.7128.2177.5299.8849.8599.95
α = 0.0570.0773.6116.6775.2199.6638.1999.84
α = 0.0156.8367.412.0971.2699.0015.8199.28
XLK
α = 0.185.9986.0377.4186.3399.7385.9399.87
α = 0.0583.6583.9869.7685.2199.4681.8399.68
α = 0.0178.6480.2654.7982.9198.6274.9499.00
XLP
α = 0.187.3489.8679.0189.6399.5785.6499.75
α = 0.0584.4486.8874.7887.3499.2482.0499.52
α = 0.0178.9180.8966.9282.9398.2573.6398.74
XLU
α = 0.179.7483.5868.1084.6399.7877.1999.89
α = 0.0574.9778.1662.5580.9299.5672.4299.74
α = 0.0164.4564.7347.4371.4098.8062.0699.15
XLV
α = 0.183.7285.2164.7685.8699.8082.5399.91
α = 0.0579.2080.7655.6182.4999.5877.0499.75
α = 0.0168.3669.7343.0675.5798.8462.8099.13
XLY
α = 0.173.5477.9057.7478.3399.7470.2799.86
α = 0.0568.8073.2950.1674.7799.4863.3999.67
α = 0.0162.4166.2031.0369.5598.6453.5398.98
Table 8. Prediction results.
Table 8. Prediction results.
ModelOriginal DatasetDataset with Proposed Features
(MI and TE-Based Network-Driven Features)
Independent t-Test Statistic
XGBoostLightGBMCatBoostXGBoostLightGBMCatBoostXGBoostLightGBMCatBoost
XLB (Materials)
Accuracy0.53760.54180.53120.54600.54600.561914.17 ***7.03 ***63.38 ***
Balanced Accuracy0.53890.54310.53240.54540.54260.559415.41 ***−2.1164.49 ***
Cohen’s Kappa Coefficient0.07730.08570.06440.09060.08550.1190297.86 ***−2.31711.19 ***
Precision0.54080.54510.53430.54710.54460.561211.61 ***−1.03109.54 ***
Recall0.53760.54180.53120.54600.54600.561914.49 ***6.35 ***59.31 ***
F1 Score0.53790.54210.53150.54640.54500.561518.34 ***8.83 ***67.71 ***
F-Beta Score (0.5)0.53930.54350.53290.54680.54470.561314.90 ***1.78 **80.04 ***
F-Beta Score (2)0.53740.54160.53110.54640.54500.561522.20 ***6.79 ***65.44 ***
Hamming Loss0.46240.45820.46880.45400.45400.438119.40 ***8.59 ***72.07 ***
XLE (Energy)
Accuracy0.52380.52700.54390.56510.55030.5365103.61 ***38.95 ***−14.99
Balanced Accuracy0.52630.53070.54850.55780.54220.528652.75 ***36.89 ***−34.53
Cohen’s Kappa Coefficient0.05230.06090.09600.11730.08570.05801282.24 ***389.73 ***−717.6
Precision0.52750.53280.55270.56810.55200.535267.81 ***41.47 ***−52.66
Recall0.52380.52700.54390.56510.55030.536588.46 ***38.01 ***−9.78
F1 Score0.52190.52250.53720.54750.52770.513957.68 ***10.00 ***−70.39
F-Beta Score (0.5)0.52430.52680.54360.55450.53540.5203161.22 ***11.99 ***−38.32
F-Beta Score (2)0.52210.52350.53870.55390.53630.522866.91 ***20.64 ***−37.53
Hamming Loss0.47620.47300.45610.43490.44970.4635120.90 ***42.91 ***−19.04
XLF (Financials)
Accuracy0.52590.52380.54070.54070.52490.550343.08 ***−0.4214.55 ***
Balanced Accuracy0.52790.52460.54130.53620.52630.547718.78 ***5.92 ***9.85 ***
Cohen’s Kappa Coefficient0.05550.04910.08230.07290.05230.0957232.33 ***62.69 ***124.04 ***
Precision0.52940.52590.54250.53830.52760.549118.96 ***0.8913.99 ***
Recall0.52590.52380.54070.54070.52490.550332.48 ***0.8657.54 ***
F1 Score0.52550.52400.54100.53700.52480.549318.74 ***3.61 ***17.84 ***
F-Beta Score (0.5)0.52730.52490.54170.53720.52610.549116.38 ***10.47 ***6.55 ***
F-Beta Score (2)0.52530.52370.54070.53870.52450.549826.20 ***2.29 **24.94 ***
Hamming Loss0.47410.47620.45930.45930.47510.449736.54 ***4.50 ***34.39 ***
XLI (Industrials)
Accuracy0.53540.51640.53760.53020.49100.5450−13.62−90.4933.05 ***
Balanced Accuracy0.53450.51310.53340.53280.49950.5480−4.78−76.4538.29 ***
Cohen’s Kappa Coefficient0.06880.02620.06700.0649−0.00090.0949−83.71−2684.93296.78 ***
Precision0.53710.51590.53620.53570.50230.5511−4.06−30.5935.38 ***
Recall0.53540.51640.53760.53020.49100.5450−10.48−109.9814.31 ***
F1 Score0.53600.51610.53670.53030.48510.5450−11.23−81.9427.85 ***
F-Beta Score (0.5)0.53660.51600.53630.53290.49240.5479−7.1−106.5228.69 ***
F-Beta Score (2)0.53560.51630.53720.52960.48600.5443−8.79−50.1125.06 ***
Hamming Loss0.46460.48360.46240.46980.50900.4550−14.29−176.4418.14 ***
XLK (Technology)
Accuracy0.48990.53650.50900.55130.53330.5397164.25 ***−3.89158.33 ***
Balanced Accuracy0.48660.53040.50170.53420.52670.5211112.09 ***−5.9136.56 ***
Cohen’s Kappa Coefficient−0.02680.06110.00350.07030.05380.04341838.24 ***−136.5926.00 ***
Precision0.49070.53440.50580.54260.53080.5284179.15 ***−15.4278.61 ***
Recall0.48990.53650.50900.55130.53330.5397162.94 ***−9.3269.18 ***
F1 Score0.49030.53510.50670.53450.53160.519297.61 ***−5.9340.65 ***
F-Beta Score (0.5)0.49050.53460.50600.53600.53100.5206108.76 ***−6.1122.56 ***
F-Beta Score (2)0.49010.53590.50790.54190.53250.5283414.04 ***−9.2345.26 ***
Hamming Loss0.51010.46350.49100.44870.46670.4603171.61 ***−9.1379.72 ***
XLP (Consumer Staples)
Accuracy0.49840.51850.51960.53540.51850.5429179.05 ***−0.841.66 ***
Balanced Accuracy0.49410.51240.51170.53360.51300.5307178.37 ***2.05 **41.78 ***
Cohen’s Kappa Coefficient−0.01170.02500.02360.06700.02610.06245736.80 ***105.42 ***1548.05 ***
Precision0.49800.51630.51570.53720.51680.5361219.84 ***0.4139.55 ***
Recall0.49840.51850.51960.53540.51850.542967.65 ***−1.1858.84 ***
F1 Score0.49820.51700.51650.53610.51750.534383.44 ***3.18 ***40.08 ***
F-Beta Score (0.5)0.49810.51650.51570.53670.51700.534188.12 ***4.62 ***51.29 ***
F-Beta Score (2)0.49830.51780.51810.53570.51800.5383138.82 ***1.80 **36.87 ***
Hamming Loss0.50160.48150.48040.46460.48150.4571129.93 ***1.0494.83 ***
XLU (Utilities)
Accuracy0.50900.50900.51640.52910.54290.538644.90 ***85.79 ***52.85 ***
Balanced Accuracy0.50750.50600.51010.52400.52540.514133.85 ***37.36 ***7.19 ***
Cohen’s Kappa Coefficient0.01490.01200.02030.04800.05220.0295683.80 ***1142.35 ***985.14 ***
Precision0.51200.51060.51470.52840.53270.522725.40 ***47.61 ***16.81 ***
Recall0.50900.50900.51640.52910.54290.538651.71 ***52.67 ***68.31 ***
F1 Score0.51000.50970.51540.52870.52770.506639.26 ***40.01 ***−41.61
F-Beta Score (0.5)0.51110.51020.51490.52850.52800.508946.33 ***34.48 ***−11.1
F-Beta Score (2)0.50930.50920.51600.52890.53450.520987.16 ***69.83 ***13.87 ***
Hamming Loss0.49100.49100.48360.47090.45710.461489.99 ***94.39 ***41.61 ***
XLV (Health Care)
Accuracy0.48990.50370.50690.52590.52590.5450127.26 ***34.39 ***106.88 ***
Balanced Accuracy0.48890.49900.50440.51920.52110.534787.79 ***50.89 ***68.11 ***
Cohen’s Kappa Coefficient−0.0222−0.00200.00880.03870.04240.07052969.95 ***1817.86 ***6413.97 ***
Precision0.49150.50150.50690.52200.52370.539571.63 ***37.44 ***116.81 ***
Recall0.48990.50370.50690.52590.52590.545088.41 ***170.69 ***62.38 ***
F1 Score0.49050.50210.50690.52200.52420.535392.07 ***133.52 ***82.12 ***
F-Beta Score (0.5)0.49100.50170.50690.52150.52380.536075.33 ***119.90 ***56.07 ***
F-Beta Score (2)0.49010.50300.50690.52200.52420.535354.62 ***107.51 ***53.91 ***
Hamming Loss0.51010.49630.49310.47410.47410.455083.94 ***39.56 ***162.06 ***
XLY (Consumer Discretionary)
Accuracy0.51850.53760.53230.53650.55240.554544.96 ***45.45 ***108.49 ***
Balanced Accuracy0.51410.53190.52720.54210.55340.554650.47 ***27.77 ***55.50 ***
Cohen’s Kappa Coefficient0.02810.06380.05430.08230.10530.1078786.75 ***462.79 ***1066.78 ***
Precision0.52060.53810.53340.54900.55940.5604123.30 ***36.93 ***60.39 ***
Recall0.51850.53760.53230.53650.55240.554543.30 ***44.46 ***65.24 ***
F1 Score0.51940.53780.53280.53720.55390.556079.35 ***34.30 ***63.25 ***
F-Beta Score (0.5)0.52000.53800.53320.54290.55670.558345.20 ***112.82 ***53.28 ***
F-Beta Score (2)0.51880.53770.53250.53720.55390.556051.08 ***46.74 ***51.19 ***
Hamming Loss0.48150.46240.46770.46350.44760.445557.21 ***45.01 ***50.27 ***
Table 9. Comparison of the number of original features and proposed features.
Table 9. Comparison of the number of original features and proposed features.
CategoryNumber of Features
PriceVolume
Original Features168 (31.11%)89 (16.48%)
Proposed Features131 (24.26%)152 (28.15%)
Table 10. Comparison of the number of MI-based features and TE-based features.
Table 10. Comparison of the number of MI-based features and TE-based features.
CategoryNumber of Features
Mutual Information198 (69.96%)
Transfer Entropy85 (30.04%)
Table 11. Comparison of the number of features based on window length.
Table 11. Comparison of the number of features based on window length.
CategoryNumber of Features
PriceVolume
20-day window42 (14.84%)52 (18.37%)
60-day window33 (11.66%)62 (21.91%)
120-day window21 (7.42%)33 (11.66%)
240-day window35 (12.37%)5 (1.77%)
Table 12. Comparison of the number of centrality measures and node embeddings.
Table 12. Comparison of the number of centrality measures and node embeddings.
CategoryNumber of Features
Centrality Measures153 (54.06%)
Node embeddings130 (45.94%)
Table 13. Comparison of the number of features based on centrality measures.
Table 13. Comparison of the number of features based on centrality measures.
FeatureNumber of Features
Average Neighbor Degree9 (5.88%)
Degree Centrality1 (0.65%)
  • Out-Degree—1 (0.65%)
Eigenvector Centrality13 (8.50%)
Closeness Centrality0 (0.00%)
Information Centrality2 (1.31%)
Betweenness Centrality9 (5.88%)
Current Flow Betweenness Centrality7 (4.58%)
Communicability Betweenness Centrality5 (3.27%)
Harmonic Centrality3 (1.96%)
Second-Order Centrality34 (22.22%)
Voterank Importance0 (0.00%)
Number of Maximal Cliques5 (3.27%)
PageRank26 (16.99%)
HITS39 (25.49%)
  • Hub—25 (16.34%)
  • Authority—14 (9.15%)
Table 14. Comparison of the number of Role2Vec vectors and FEATHER vectors.
Table 14. Comparison of the number of Role2Vec vectors and FEATHER vectors.
CategoryNumber of Features
Role2Vec74 (56.92%)
FEATHER56 (43.08%)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Choi, I.; Kim, W.C. Enhancing Exchange-Traded Fund Price Predictions: Insights from Information-Theoretic Networks and Node Embeddings. Entropy 2024, 26, 70. https://doi.org/10.3390/e26010070

AMA Style

Choi I, Kim WC. Enhancing Exchange-Traded Fund Price Predictions: Insights from Information-Theoretic Networks and Node Embeddings. Entropy. 2024; 26(1):70. https://doi.org/10.3390/e26010070

Chicago/Turabian Style

Choi, Insu, and Woo Chang Kim. 2024. "Enhancing Exchange-Traded Fund Price Predictions: Insights from Information-Theoretic Networks and Node Embeddings" Entropy 26, no. 1: 70. https://doi.org/10.3390/e26010070

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop