Next Article in Journal
Multi-Scale Mixed Attention Network for CT and MRI Image Fusion
Next Article in Special Issue
Complex Network Analysis of Mass Violation, Specifically Mass Killing
Previous Article in Journal
Multi-User Measurement-Device-Independent Quantum Key Distribution Based on GHZ Entangled State
Previous Article in Special Issue
Complex Network Study of Solar Magnetograms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Examining Supervised Machine Learning Methods for Integer Link Weight Prediction Using Node Metadata

1
School of Industrial Engineering, Purdue University, West Lafayette, IN 47906, USA
2
RAND Corporation, Santa Monica, CA 90407-2138, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2022, 24(6), 842; https://doi.org/10.3390/e24060842
Submission received: 27 April 2022 / Revised: 8 June 2022 / Accepted: 14 June 2022 / Published: 18 June 2022

Abstract

:
With the goal of understanding if the information contained in node metadata can help in the task of link weight prediction, we investigate herein whether incorporating it as a similarity feature (referred to as metadata similarity) between end nodes of a link improves the prediction accuracy of common supervised machine learning methods. In contrast with previous works, instead of normalizing the link weights, we treat them as count variables representing the number of interactions between end nodes, as this is a natural representation for many datasets in the literature. In this preliminary study, we find no significant evidence that metadata similarity improved the prediction accuracy of the four empirical datasets studied. To further explore the role of node metadata in weight prediction, we synthesized weights to analyze the extreme case where the weights depend solely on the metadata of the end nodes, while encoding different relationships between them using logical operators in the generation process. Under these conditions, the random forest method performed significantly better than other methods in 99.07% of cases, though the prediction accuracy was significantly degraded for the methods analyzed in comparison to the experiments with the original weights.

1. Introduction

The term “node metadata” refers to “observed discrete features or descriptors of nodes in a network” [1]. For example, in a network of co-authorships, an author’s country of origin, area of research, etc., could be examples of node metadata. Similarly, in an air traffic network, where airports are nodes and links are flights connecting them, the size of the population of the city where an airport is located and whether or not it is an airline’s hub can be used as node metadata. As the amount and quality of information about networks is made available, understanding the role of node metadata (or node attributes) in the prediction of network properties becomes increasingly important, given that they can carry relevant information about how nodes interact in a network [2].
In that sense, we would like to understand whether incorporating the information provided by node metadata can aid in the task of link weight prediction. As a step towards this objective, we investigate herein whether including it as a similarity feature (referred to as metadata similarity) between the nodes of a link increases the prediction accuracy of link weights when applied with common supervised learning methods.
We focus solely on the problem of link weight prediction, assuming that the existence of links is known. In line with Zhu et al. [3], we consider that link prediction and weight prediction are different problems and should be tackled separately, according to the “no free lunch” theorem [4]. More specifically, in a supervised learning setting, link prediction is more akin to a classification problem with a binary output, whereas link weight prediction better resembles a regression problem, since, in general, link weights assume real values.
This is a relevant issue in cases where there is incomplete information due the nature of data acquisition, for example, through surveys or automated processes. In social network analysis, survey responders may have indicated their connections to other members of an organization (i.e., the existence of links), but failed to provide information regarding the frequency of contact with them (i.e., missing weights). In recommendation systems, users may have indicated that they visited a place or bought an item but not how they rated it, and it might be desirable to predict this for future recommendations. The reasons for this can also be due to data anonymization, temporal or bandwidth limitations, etc. In these contexts, the fact that a weight is predicted to be zero would not indicate the non-existence of an edge but could indicate a low level of interaction between nodes or a low rating given by a user, for example.
To the best of our knowledge, the analysis of whether incorporating the information provided by node metadata can improve prediction accuracy has not been studied in the link weight prediction literature. In this work, we analyze whether including metadata similarity features in a baseline set of local topological similarity features increases the prediction accuracy of common supervised machine learning methods. We also analyze the performance of the same methods and set of features on synthesized weights generated to represent the extreme case in which the weights depend solely on the metadata of the end nodes, while expressing different relationships between them using logical operators in the generation process.
The literature on link weight prediction has concentrated on predicting link weights mapped to the [ 0 , 1 ] interval. The reasoning provided for this is that link weights are “analogous to link-existence probabilities” [5]. However, the majority of datasets considered in the link weight prediction literature have link weights that are count variables indicating the number of interactions observed between end nodes, as in [6,7,8,9]. In these cases, normalizing weights to values in [ 0 , 1 ] can hurt prediction accuracy given the arbitrary nature of the normalization function and the occurrence of rounding errors from scaling and re-scaling original weights and predictions. Following these observations, we argue that, in most cases, link weights should be treated as positive integer values.
The rest of this paper is organized as follows: Section 2 discusses related work, Section 3 describes the problem formulation and the metrics used for prediction accuracy comparison, Section 4 outlines the supervised learning methods analyzed, Section 5 presents metadata similarity features and the set of baseline topological features, Section 6 and Section 7 describe and analyze the experimental results on real-world and synthesized datasets, respectively, and finally Section 8 discusses the results and future work.

2. Related Work

We briefly explain the main branches of research related to link weight prediction and the use of node metadata/attributes.  
  • Link prediction. The link prediction problem is a widely known and studied problem in the literature, having been introduced in pioneering works such as [10,11,12,13]. Most of the traditional approaches use similarity indices (e.g., local, global, quasi-local) or maximum likelihood methods (e.g., hierarchical structure and stochastic block models) [14]. More recently, other approaches have appeared, such as the use of network embeddings, matrix completion, ensemble learning, etc. For a more comprehensive view of traditional and more recent methods, we refer the interested reader to the reviews presented in [14,15], respectively.
  • Link weight prediction. Few studies have analyzed the task of link weight prediction. Aicher et al. [16] proposed a weighted generalization of the stochastic block model to generate weighted networks, which can be used for weight prediction. Zhao et al. [5] proposed weighted extensions of unweighted local similarity metrics to predict both links and link weights using reliable routes as a motivation. Zhu et al. [3] used the assumption that link weights are locally homogeneous to propose a method to predict link weights based on the weights of neighbor sets. Fu et al. [17] compared a set of supervised learning methods on a large set of topological features derived from local and global similarity metrics derived from the original network and the line graph, and learned features using deep learning methods. None of these studies analyzed the use of node metadata to improve weight predictions.
  • Link weight prediction in weighted signed networks. Two recent papers have tackled the problem of link weight prediction in signed social networks [18,19], with the weights taking any value on a [ 1 , 1 ] interval. These papers extend the problem of predicting the signs of edges on signed social networks, where agents rate others positively or negatively, with a focus on also predicting the extent of a “like” or “dislike”. This problem differs from the one studied here because the methods are highly tailored to social networks, where weights assume continuous values and represent the opinion of one node about another node in the network, whereas in our work weights are integer-valued and represent a notion of frequency, such as a number of interactions, etc.
  • Use of node metadata in link or link weight prediction: Leveraging node metadata can be found mostly in the link prediction literature. Zhao et al. [20] added a regularization term in the loss function that enforced link predictions to take into account similarity measures, which could be derived from node metadata. In a different approach, relational models also included node metadata information in link prediction. Initial works such as that of Popescul and Ungar [21] leveraged relational information between authors and venues in regression models for the prediction of co-authorships, and Taskar et al. [22] used probabilistic relational models for link prediction with relational data. More recently, a number of papers have shown improved link prediction accuracy using node metadata in conjunction with relational models, such as [23,24,25]. Moreover, in the graph representation learning literature, Zhang and Chen [26] have shown improvements when using node attributes for link prediction using graph neural networks (GNNs). To the best of our knowledge, there are no such results for link weight prediction but the improvements seen in the link prediction literature would lead us to believe that improvements in weight prediction could be obtained through the use of node metadata as well.

3. Problem Definition and Accuracy Metrics

3.1. Problem Definition

Suppose that G = ( V , E , W ) is an undirected, weighted network, with V, E and W representing the set of nodes, links and link weights, respectively. Furthermore, G has no self-loops or multi-edges. The problem is to predict the missing weights in the network, given that all links are known.
Formally, we assume that the weight w u v Z + on a link ( u , v ) is sampled i.i.d. from a probability distribution P ( w | x , θ ) , where P is any distribution, x R p are link features and θ R q are distribution parameters. The goal is to obtain θ ^ that solves the minimization of the loss function:
minimize θ L = E D dist w , f ( x , θ ) ,
where f : R p × R q R is any function that estimates P ( w | x , θ ) , dist : R × R R is any distance function and the expectation is computed over the samples D .
Since the expectation in Equation (1) cannot be directly computed, we estimate it as an average over the samples available. For this purpose, we randomly divide 90% of the weight samples W into a training set W train and 10% into a testing set W test , with the ratios in line with previous works [3,5,17]. As usual, W train W test = and W train W test = W . Then, we train the supervised learning methods on W train to learn the parameters θ ^ , and calculate the average loss over W test as an estimate of the expectation.

3.2. Accuracy Metrics

In this section, we present the metrics used to measure the quality of our predictions, i.e., that will fill the role of the dist function in Equation (1). The most common metrics used in the literature have been the Pearson correlation coefficient (PCC) and the root mean squared error (RMSE) [3,5]. Although we will use the PCC as one of the accuracy metrics, we replace the RMSE with the relative squared error (RSE). Given that we are not considering normalized weights to [ 0 , 1 ] , the relative squared error (RSE) is a more appropriate metric, because it normalizes the error by indicating how much better (or worse) a method performs compared to merely predicting each weight as the average of the observed weights.
To simplify the notation and without the loss of generality, we assume that the sample links have an order and thus we merge the indices u , v in w u v to k, with  k = 1 , , N , and N is the number of elements in W test . We denote w k as the weight value and w ^ k as is its predicted value.
  • Pearson Correlation Coefficient (PCC): Computes the degree of correlation between two sets of data, and outputs values in a [−1,1] interval. Positive values indicate positive correlation (opposite for negative values) and zero values indicate no correlation.
    P C C = 1 N 1 k = 1 N w k w ¯ σ w w ^ k w ^ ¯ σ w ^
    where w ¯ = k = 1 N w k N is the average of the observed weights w k W test , and  w ^ ¯ = k = 1 N w ^ k N is the average of the predicted weights w ^ k .
  • Relative Squared Error (RSE): Computes the sum of the squared error between the original and predicted weights normalized by the sum of the squared error between the original weights and the average of the weights in the test data.
    R S E = ( k = 1 N w k w ^ k ) 2 ( k = 1 N w k w ¯ ) 2 .
All the abbreviations introduced throughout the text (such as RMSE and RSE above), along with their definitions, can be found in Abbreviations.

4. Supervised Learning Methods

Because of the assumption that weights are integers that represent counts, the most natural distribution to model them is a Poisson distribution. Therefores, we use the Poisson regression as a baseline supervised learning method. To allow for more flexibility, we compare this baseline to the performance of Poisson mixture models, thus permitting subpopulations following different Poisson distributions within the broader population.
A caveat of the Poisson distribution is that it is associated with rigid assumptions, such that the mean equals the variance, which may not hold for weights in real-world networks. Thus, we encourage the investigation of the fitness of other distributions that model count variables, such as negative binomial regression, and leave this task for future works.
Regression-based methods have the main advantage of statistical interpretability. Thus, we address the question of whether regression-based methods, given the topological features analyzed, can make predictions that are at least as good as the predictions of other common methods. For this purpose, we compare the prediction performance of Poisson regression and its mixture models to the performance of random forest and support vector machine, two methods that have been previously used in the link weight prediction literature [17], and of neural networks, which have achieved state-of-the-art results in many complex tasks in machine learning.
We now formalize each of the methods, according to the notation defined in Section 3.

4.1. Regression-Based Methods

  • Poisson regression (Poi): The baseline constrains the model to a single Poisson distribution; thus it is the least flexible model. It provides an adequate baseline given the assumption that link weights are positive integers representing a frequency, such as the number of encounters, etc. In this case, P ( w | x , θ ) = Poisson ( w | x , λ ( x ) ) = e λ ( x ) λ ( x ) w w ! , where λ ( x ) = exp ( β T x ) and θ = β .
  • Poisson mixture models (Mix#): These methods generalize the baseline method by modeling subpopulations (components), where each follows a Poisson distribution with a different rate. In this work, the total number of components is known. Formally, P ( w | x , θ ) = c = 1 C π c Poisson ( w | x , λ c ( x ) ) , where the mixture weights π c satisfy 0 π c 1 , c = 1 , , C , and  c = 1 C π c = 1 , λ c ( x ) = exp ( β c T x ) and θ = ( π 1 , , π C , β 1 T , , β C T ) , where C is the total number of subpopulations. We will refer to this method as Mix#, where # is the number of components.

4.2. Comparison Methods

  • Random Forest (RF): This is an ensemble method that combines the outputs of many decision trees to avoid overfitting. In a previous work [17], it achieved the best performance in link weight prediction when compared to other supervised learning methods, though with a different set of features. Formally, the decision tree model is described as: E [ w | x ] = m = 1 M h m I ( x R m ) = m = 1 M h m ϕ ( x , y m ) , where h m is the mean response in the m th region R m , and  y m denotes the variable and its threshold value to split the tree into the m th leaf. Thus, θ = ( y 1 , , y m ).
  • Support Vector Machine (SVM): In regression tasks, SVM aims to estimate the line that best fits the data within an arbitrary boundary. Differently from the previous methods, it does not have a probabilistic perspective. The method solves the following minimization problem: minimize β 1 2 | | β 2 s . t . w i β T x i ε , i 1 , , | W train | , for some ε > 0 .
  • Artificial Neural Network (NN#): The structure of an artificial neural network with n hidden layers is given by:
    f ( x , θ ) = f ( n ) f ( n 1 ) ( f ( 2 ) ( f ( 1 ) ( x ; θ 1 ) ; θ 2 ) , ) ; θ n 1 ) ; θ n T θ n + 1 + β n + 1 ,
    where f ( i ) ( x , θ i ) f ( i ) ( x , θ i , β i ) = σ ( θ i T x + β i ) , σ is a nonlinear activation function (e.g., sigmoid), and  θ = ( θ 1 , , θ n + 1 , β 1 , , β n + 1 ). For more details, see Chapter 6 in [27]. We will refer to this method as NN#, where # is the number of hidden layers n.
   For the methods above, we use the R built-in function glm and packages flexmix, randomForest (with n t r e e = 500 ), e1071 (with the radial kernel and default hyperparameters), and H2o (with distribution = ‘poisson’, epochs = 500, with each hidden layer composed of ten neurons, and other hyperparameters as default), respectively.

5. Proposed Features

In this section, we describe the features used for link weight prediction, which are based on node metadata or the network topology. We describe how we incorporate metadata about the nodes as a feature for link weight prediction. We also present the similarity features derived from local topology.
The notation for this section is defined as follows: we denote the similarity between nodes as u and v according to a metric z by s u , v z , where the metric z can be based on the node metadata or network topology. Then, for a link k = ( u , v ) , the feature vectors used for prediction will be of the form x k = [ s u , v z 1 , , s u , v z p ] , where each z 1 , , z p is one of the similarity metrics described.

5.1. Metadata Similarity Features: Incorporating Node Metadata

The metadata available are assumed to convey extra information about the nodes in the network, not the links. Given a link ( u , v ) , let y u , y v R t denote vectors that store metadata information about end nodes u and v, respectively.
However, we want to predict the values of link weights, so we need to process the node metadata available for the end nodes into a feature of the links. This could be achieved in multiple ways. In this study, for any link ( u , v ) , we calculate the similarity between end nodes with respect to their metadata y u and y v , and utilize the output as one of the link’s features in the supervised methods considered. We will later refer to this feature as metadata similarity.
Furthermore, for the calculation of the similarity between vectors y u and y v , there are many metrics that could be used. We consider the cosine similarity and the Pearson correlation coefficient for that purpose, which are described below.
  • Cosine Similarity (CS): This takes two vectors as its input and outputs a value in the interval [−1, 1]. Its mathematical definition is given by:
    s u , v C S = n = 1 t y u ( n ) y v ( n ) n = 1 t y u ( n ) 2 n = 1 t y v ( n ) 2
    where t is the size of the metadata vector, and  y u ( n ) is the nth entry of y u .
    It can be interpreted as the cosine of the angle between two vectors y u and y v , where s u , v C S = 1 means that the vectors have the same orientation, s u , v C S = 1 that they are diametrically opposed and s u , v C S = 0 that they are orthogonal to each other, irrespective of their magnitude.
  • Pearson Correlation Coefficient (PCC): For completeness, we repeat the definition expressed as a similarity measure between the metadata vectors y u and y v :
    s u , v P C C = 1 t 1 n = 1 t y u ( n ) y ¯ u σ y u y v ( n ) y ¯ v σ y v
    where y ¯ u is the sample mean of y u , and  σ y u is the standard deviation of y u .
When the vector of metadata contains distinct variable types (categorical, numerical and/or ordinal), we first replace the categorical variables with their one-hot encoding, then metadata similarity is computed over this new vector. For example, if a patient metadata is given by (age, blood type, Rh) = (64, A, positive), then the transformed vector with one-hot encoding would be given by (age, A, B, AB, O, Rh+, Rh-) = (64, 1, 0, 0, 0, 0, 1, 0), where the continuous variable age remains unchanged. A limitation of this approach is that the number of categories assumed by a categorical value has to be comparatively lower than the sample size. There is no need to transform ordinal variables, such as university rank, into a one-hot encoding, as the difference between their original values already serves as a distance measure between two nodes.

5.2. Topological Features

We investigate if the prediction accuracy increases when we include metadata similarity features to the supervised learning methods analyzed. Previous works have studied many topological metrics for link weight prediction [5,17]. As a set of baseline features, we focus on a set of topological features based on neighbors in common and neighbor degrees, as described in Table 1.

6. Experimental Results on Empirical Data

6.1. Description of Empirical Datasets

For each empirical dataset, we describe the network in terms of the nodes, links and the metadata available for each of the nodes. All networks are weighted and undirected, where link weights indicate the number of interactions between two nodes. In addition, we provide some topological details of the datasets in Table 2.
  • BusFac [11]: A network of transfers of business faculty between university departments. Nodes are university departments (in this case, business departments), links exist when at least one transfer occurred from one university to another (independently of direction) and weights indicate the quantity of transfers between them. Node metadata consist of: region where university is located (Northeast, Southeast, etc.), university’s 2021 US News rank, incoming and outgoing assistant professor percentage and incoming and outgoing male percentage.
  • CompSciFac [11]: This network has the same interpretation of nodes and links as BusFac. In this dataset, faculty transfers are related to computer science departments.
  • HistFac [11]: This network has the same interpretation of nodes and links as BusFac and CompSciFac. In this dataset, faculty transfers are related to history departments.
  • BookCross [36]: This is a subset of the Book Crossing network of book reviews. Each node is a user, which is linked to another user if they have a book review in common. The weights denote the quantity of books reviewed in common. Node metadata consist of country and age.

6.2. Results Obtained with Empirical Datasets

We now describe in detail the experiments and present their results on the empirical datasets. For each dataset, we predicted the weights of each of the supervised learning methods on the set of baseline topological features and compared the performance between methods (Section 6.2.1). Then, we included the metadata similarity features in the set of baseline features, and assessed whether there was an increase in prediction accuracy for each of the methods (Section 6.2.2).
For each experiment, we present the mean of 30 trials for both PCC and RSE, where each trial consists of a random sample of 90% of the data for training and 10% for testing, according to Section 3.2. To statistically compare the prediction accuracy obtained in different scenarions, we performed two-sided t-tests with 5% significance, where the null hypothesis was that the averages of the two samples were the same.

6.2.1. Comparison between Methods

We analyzed the prediction accuracy of the supervised methods when using the baseline topological similarity features (described in Section 5.2). For the Poisson mixture models, we varied the number of components in our experiments from two to five, and then determined the optimal number of components by comparing their prediction accuracy. This may not be feasible for large datasets, where other approaches for assessing the number of modes could be more appropriate.
We display the results according to the accuracy metrics—the Pearson correlation coefficient and relative squared error—for each empirical dataset in Figure 1. Both metrics provide a consistent ranking of the methods, when ordered from highest to lowest accuracy.
In the networks of faculty transfers BusFac, CompSciFac and HistFac, the Poisson mixture models had, on average, the best performance (with statistical significance) when compared to other methods, but the number of components that achieved the best performance varied. On BusFac, the mixture model with three components outperformed (with 0.818 for PCC and 0.342 for RSE), on average, the baseline Poisson regression (0.547 for PCC and 0.716 for RSE), random forest (0.565 for PCC and 0.712 for RSE) and SVM (0.609 for PCC and 0.696 for RSE). It also outperformed the neural networks with one (0.605 for PCC and 0.682 for RSE) and two hidden layers (0.597 for PCC and 0.674 for RSE). In the CompSciFac dataset, the mixture model with four and five components outperformed (0.804 for PCC and 0.372 for RSE, and 0.805 for PCC and 0.368, respectively), on average, the Poisson regression (0.554 for PCC and 0.705 for RSE), random forest (0.574 for PCC and 0.724 for RSE) and SVM (0.609 for PCC and 0.696 for RSE). The mixture models also outperformed both neural networks with one and two hidden layers, which had 0.565 for PCC and 0.724 for RSE, and 0.552 and 0.793 for RSE, respectively. Similarly, for HistFac, the mixture model with five components (0.774 for PCC and 0.415 for RSE) outperformed, on average, the Poisson regression (0.65 for PCC and 0.599 for RSE), random forest (0.602 for PCC and 0.653 for RSE) and SVM (0.627 for PCC and 0.642 for RSE). It also outperformed the neural networks with both one hidden layer (0.659 for PCC and 0.619 for RSE) and two hidden layers (0.649 for PCC and 0.633 for RSE). When comparing the mixture models among themselves, for most cases, we failed to reject the null hypothesis that the mean of the sample errors was different for pairs of mixture models with different numbers of components.
In the BookCross dataset, SVM, NN1 and NN2 had the best prediction accuracy, on average, with a PCC of 0.918, 0.918 and 0.916 and RSE of 0.16, 0.162 and 0.17, respectively. All methods performed, on average, better than the baseline Poisson regression (0.906 for PCC and 0.182 for RSE).
Based on the results in Figure 1, we can see that mixture models performed comparably to the comparison methods, achieving the best performance in many cases, for both metrics. Thus, their use should be considered especially when statistical interpretability is a valuable property. Even for the BookCross dataset, the mixture model with three components had a very close prediction accuracy to SVM, with PCCs of 0.912 and 0.918 and RSEs of 0.17 and 0.12, respectively. Still, this comparison was limited to four datasets, but further investigation into the comparison of mixture models with other methods on a larger and more diverse set of datasets is highly encouraged.

6.2.2. Comparison Using Metadata Similarity

In this section, we discuss the results from the inclusion of metadata similarity features (as described in Section 5.1) in the set of features used for prediction. We analyzed whether the prediction accuracy increased when we expanded the set of features to include the metadata similarity features in addition to the baseline topological features.
We show the outputs of the regression-based methods in Table 3 and the comparison methods in Table 4. For each method “X”, we show the average accuracy with both topological and metadata similarity features (denoted by “mX”) side by side to the average accuracy on the set of topological features (denoted as “tX”). The best-performing set of features is indicated in bold, and the pairs for which we rejected the null hypothesis that means were equal with 5% statistical significance are highlighted in gray.
For the regression-based methods (Table 3), it is not possible to draw the conclusion that including metadata similarity features resulted in accuracy gains. For some combinations of datasets and methods, the addition of metadata significantly improved accuracy. For example, Poisson regression on BusFac improved to 0.58 from 0.55 for PCC, and to 0.64 from 0.72 for RSE. On the other hand, mixture models with three components exhibited a decrease in accuracy with the inclusion of metadata similarity on CompSciFac from 0.775 to 0.758 for PCC, and 0.381 to 0.471 for RSE.
For the random forest and SVM methods (Table 4), it was not possible to draw the conclusion that including metadata similarity features improved the prediction accuracy. The results were mixed for all the methods and datasets. For example, the prediction accuracy improved for random forest on the BusFac dataset, from 0.566 to 0.672 for PCC and from 0.712 to 0.568 for RSE, and on the CompSciFac dataset from 0.575 to 0.652 for PCC and from 0.724 to 0.61 for RSE. On the other hand, the accuracy decreased for SVM on BookCross after including metadata features from 0.918 to 0.916 for PCC, and 0.156 to 0.163 for RSE.
For the neural networks with one and two hidden layers (NN1 and NN2, respectively), most of the comparisons were not statistically significant. In the cases where the comparison was statistically significant, the accuracy improved with the inclusion of metadata similarity features. On the BusFac dataset, the prediction accuracy improved both for NN1 (from 0.605 to 0.661 for PCC, and from 0.682 to 0.605 for RSE) and for NN2 (from 0.597 to 0.68 for PCC, and from 0.674 to 0.582 for RSE). Furthermore, on BookCross, the prediction with NN2 significantly improved from 0.916 to 0.919 for PCC and 0.1697 to 1.1615 with RSE.
Moreover, the null hypothesis—that the average accuracy for the two sets of features, for each method and dataset, was the same—was not rejected in the majority of cases. Due to the restricted number of public datasets of weighted networks that contain node metadata, this work was restricted to four real-world datasets. We encourage further studies with a large number of networks to obtain more generalizable conclusions.

7. Experimental Results Obtained with Synthesized Datasets

We aimed to investigate the degree to which the performance of supervised learning methods with the same set of topological and metadata similarity features as before was affected when the link weights were generated using different weight generation processes. In the following, we describe in detail how we generated the synthesized weights and the experimental results for these datasets.

7.1. Description of Synthesized Datasets

Given that our focus was on the role of node metadata in weight prediction, we generated weights that were determined by the application of a logical operator (AND, OR, XOR) to the pair of vector metadata of the end nodes of a link. More specifically, given a link ( u , v ) , where y u , y v R t denote the vectors that store metadata information about end nodes i and j, respectively, the generated weights are given by:
w ˜ u v = n = 1 t ( y u ( n ) y v ( n ) ) ,
where ⊕ is a placeholder for a specific logical operator (either AND, OR or XOR).
In the synthesized datasets, the same node and edge structure from the empirical datasets described in Section 6.1 is kept. For each empirical dataset (BusFac, CompSciFac, HistFac, BookCross), we create a set of synthesized weights using each of the logical operators (AND, OR and XOR), according to Equation (6). We note that the metadata for the incoming/outgoing male percentage and incoming/outgoing assistant professor percentage and region was not used for BusFac, CompSciFac and HistFac.Moreover, if the resulting sum between two nodes in the structure is zero, then the weight is defined as w ˜ u v = 0 .
Note that all metadata must be a binary digit to be appropriately used by a logical operator for weight generation. Thus, any continuous or ordinal variable needs to be transformed into binary values for the experiment. To accomplish this, we assign each continuous or ordinal variable to its quartile number and then create a one-hot encoding of the quartiles. For example, if the age of a user falls in the 2nd quartile, then the transformed age of this user would be equal to (0,1,0,0), with a “1” in the index corresponding to the second quartile. Any categorical variable is directly transformed via a one-hot encoding, as was performed in Section 5.1. Following this process, all metadata assume a binary value, which can then be compared using logical operators.
Given this weight generation process, the weights generated using the AND logical operator test whether the methods (with the features analyzed) are effective in predicting weights that are directly related to the number of identical metadata (for categorical variables) and similar metadata (for continuous variables) between these two nodes. Meanwhile, the weights generated using the XOR logical operator test the performance of the prediction methods on link weights that are directly related to the number of distinct metadata between them. Finally, the weights generated using the OR logical operator test whether the analyzed methods can predict weights between nodes of which the “proximity” regarding their metadata is somewhere between the AND and XOR operators, given that the generated weights are higher when their metadata are different and lower (but not zero) when their metadata are similar.
We note that there are many ways one could encode link weight dependencies on node metadata in a weight generation process. The logical operators preserve the assumption that weights are integer values, thus allowing direct comparisons with the original weights in the empirical datasets analyzed. A downside of this generation process is that some granularity in the metadata is lost due to the truncation of continuous variables into quartiles. However, we consider that general (dis)similarities between nodes are still captured and allow for a sufficient distinction between different relationships (obtained by applying different logical operators) between link weights and the metadata of end nodes.

7.2. Results Obtained with Synthesized Datasets

We present the results of the experiments using the synthesized weights based on the logical operators AND, OR and XOR (henceforth collectively referred to as logical groups). First, for the same logical operator, we compare the performance of each of the methods in each dataset (Section 7.2.1). Then, we compare the prediction accuracy over the original weights versus the synthesized weights (Section 7.2.2). Finally, we compare the performance between logical operators, highlighting how different relationships between weight and node metadata can affect prediction performance (Section 7.2.3).
As with the empirical datasets, we used both the metadata similarity features and topological features for all the predictions, as described in Section 5.1 and Section 5.2, respectively. Likewise, for each experiment with synthesized data, we show the average prediction accuracy of 30 trials for both PCC and RSE, where 90% of data were randomly sampled for training and 10% for testing, as explained in Section 3.1.
To make comparisons of the prediction accuracy obtained in different scenarions, we performed two-sided t-tests with a significance level of 5%, where the mean of the PCC and RSE for each of the weights generated using a logical operator (AND, OR and XOR) was compared to the mean of the PCC and RSE obtained for the original weights. We also performed a pairwise comparison of the logical groups AND, OR and XOR with respect to their mean PCC and RSE using two-sided t-tests with the same significance level. The null hypothesis in all cases was that the mean over any two given groups was equal.

7.2.1. Comparison between Methods on Synthesized Weights

In this section, we aimed to understand which methods performed best in each dataset, given a logical operator (AND, OR and XOR). The results for BusFac, CompSciFac, HistFac and BookCross are displayed in Figure 2, Figure 3, Figure 4 and Figure 5, respectively, using the PCC and RSE metrics. Furthermore, we analyzed whether there was any dataset in which a method exhibited a better performance across all logical operators. Notably, for each logical operator, we were able to identify the best and worst performing methods across all datasets, with very few exceptions. Moreover, those methods tended to be the same for all logical operators, as we detail below.
In terms of PCC, random forest was the best performing method over the logical groups, outperforming the Poisson regression, the mixture models with two to five components, SVM and neural networks with one and two hidden layers in 99.07% of combinations of logical groups (AND, OR, XOR) and all of the four datasets. For example, for the weights generated using the AND logical operator on the CompSciFac dataset, the random forest method had a PCC of 0.4922 and an RSE of 0.7633. Meanwhile, for the AND logical operator and dataset CompSciFac, Poisson regression had a PCC of 0.3138 and RSE of 0.9054, whereas the mixture model methods with two, three, four and five components had a PCC of 0.3175, 0.3300, 0.3278 and 0.3371, and an RSE of 0.9033, 0.8959, 0.8982 and 0.8911, respectively. SVM had a PCC of 0.3742 and an RSE of 1.0005. Neural Networks with one and two hidden layers had a PCC of 0.3562 and 0.3506 and an RSE of 0.8928 and 0.8938, respectively. The only instance where random forest did not outperform the other methods in the logical groups was for BookCross with the AND logical operator (see Figure 5), where the mixture model with two components performed best (0.8811 for PCC, 0.2369 for RSE); random forest had a lower PCC of 0.8443 and a higher RSE of 0.2875.
On the other hand, Poisson regression and the neural network with one hidden layer were the worst performing methods on the logical groups in terms of the PCC metric. Poisson regression was outperformed by the mixture model methods with two to five components, random forest, SVM and neural networks with one to two hidden layers in 74.07% of combinations of logical groups (AND, OR, XOR) and all four datasets. For example, for the weights generated using the XOR logical operator on the BookCross dataset, the Poisson regression method had a PCC of 0.3109. Meanwhile, in the same context, for the weights generated by the operator XOR on the BookCross datset, the mixture models with two, three, four and five components had PCC values of 0.5389, 0.5464, 0.5592, and 0.5781, respectively; the random forest method had a PCC of 0.8429; the SVM method had a PCC of 0.7253 and the neural networks with one to two hidden layers had PCC values of 0.4948 and 0.4850, respectively. Additionally, the neural network with one hidden layer method was outperformed by the Poisson regression, the mixture model methods with two to five components, random forest, SVM and the neural networks with two hidden layers in 79.63% of combinations of logical groups (AND, OR, XOR) and datasets. For instance, for the weights generated using the XOR logical operator on the BusFac dataset, the neural network with one hidden layer had a PCC of 0.0208; the Poisson regression method had a PCC of 0.0642; the mixture models with two, three, four and five components had PCC values of 0.0546, 0.0623, 0.0534 and 0.0561, respectively; the random forest method had a PCC of 0.3303; the SVM method a PCC of 0.0331 and the neural network with two hidden layers had a PCC of 0.0243.
When comparing the performance across datasets, each method on the logical groups performed better on the BookCross dataset, which had the highest PCC and lowest RSE in 100% (in terms of PCC) and 99.07% (in terms of RSE) of combinations of methods and weight generation process (with AND, OR and XOR operators). For example, the mixture model method with two components over the weights generated using the AND logical operator had a PCC of 0.8811 and an RSE of 0.2369 on the BookCross dataset. For the same generation process with the AND operator and the mixture model with two components, the predictions on the other datasets had a lower PCC and a higher RSE. For instance, on the BusFac dataset, this combination had a PCC of 0.0527 and an RSE of 1.0041, on CompSciFac a PCC of 0.3175 and an RSE of 0.9033 and, finally, on HistFac a PCC of 0.3091 and an RSE of 0.9154.
In contrast, the prediction accuracy of the synthesized weights was worse on the BusFac dataset than in other datasets in 100% of methods on the logical groups in terms of the PCC metric and in 96.3% of cases in terms of the RSE metric. For example, for the SVM method, the weights generated with the OR logical operator on the BusFac dataset had a PCC of 0.0310 and an RSE of 1.2303. For the same logical operator OR, the predictions of the SVM had a higher PCC and a lower RSE in all the other datasets. On CompSciFac, the combination had a PCC of 0.3734 and an RSE of 0.9995, on HistFac a PCC of 0.1493 and an RSE of 1.2084, and on BookCross a higher PCC of 0.7253 and a lower RSE of 0.5563.
Finally, for the methods and datasets analyzed, we also noted that the best-performing method on the weights generated based on the logical operators on the metadata were different from the ones generated based on the original weights from the empirical datasets. As described above, when the weights were determined using logical operators on the metadata, then random forest performed better in 99.07% of cases. On the other hand, mixture models or random forest may achieve better performance on the unknown real-world generation process, as discussed in Section 6.2.1.

7.2.2. Comparison between Original and Synthesized Weights

In this section, we analyzed whether the prediction accuracy of the tested methods based on the original weights was higher than that based on each of the sets of synthesized weights generated by logical operators AND, OR and XOR on node metadata, given the same method and dataset.
The results for the regression-based methods and comparison methods (defined in Section 4) with the PCC metric are displayed in Table 5 and Table 6, respectively. For each method X, we display the average PCC for the original, AND, OR and XOR weight generation processes (denoted as ORIG-X, AND-X, OR-X and XOR-X, respectively) side by side. The best-performing weight generation processes according to the metric used are indicated in bold, and in gray we indicate the rows for which we rejected the null hypothesis that the means were equal between the original weight generation process and each of the logical operator weight generation processes, with a statistical significance of 5%.
A side-by-side presentation of the results obtained with the RSE metric regarding the regression-based methods and the comparison methods is included in the Appendix A in Table A1 and Table A2, respectively, since the results obtained with the RSE metric were mostly consistent with the PCC metric. As with the PCC results, the average RSE for the original, AND, OR and XOR weight generation processes (denoted as ORIG-X, AND-X, OR-X and XOR-X, respectively) are presented side-by-side for each method X.
For both the regression-based methods (Table 5) and the comparison methods (Table 6), we concluded that, for each dataset, method and metric, the prediction accuracy based on the original weights was higher than for each of the logical operator groups (for both PCC and RSE metrics). In all cases, we rejected the null hypothesis at a significance level of 5%.
As an illustration, for the Poisson regression method, the prediction accuracy was worse than for the original weights for each of the generation processes with AND, OR and XOR operators, with the latter having lower PCC values. For example, on the BusFac dataset, the prediction accuracy over the original weights (ORIG) had a PCC of 0.5784. Meanwhile, for the same BusFac dataset and the Poisson regression method, the AND, OR and XOR weight generation methods had PCC values of 0.0409, 0.0618 and 0.0642, respectively, as shown in Table 5.
Additionally, for the mixture model with three components on the HistFac dataset, the prediction accuracy over the original weights (ORIG) had a PCC of 0.7252. Meanwhile, the prediction accuracy was worse for each of the other generation processes with AND, OR and XOR operators, with lower PCC values. More specifically, the AND, OR and XOR weight generation methods had lower PCCs of 0.3486, 0.1462 and 0.1435, respectively.
It is evident that the original weight generation process makes it significantly easier to predict weights than the logical operator weight generation processes for the methods and set of features investigated. This might be related to a higher explanatory power of the topological features in the original weight generation process as compared to the processes that are solely dependent on metadata. Even so, it is worth noting that some PCC values in the logical groups indicated a degree of correlation above 0.8. This was true of the BookCross dataset for the AND logical operator group for the mixture model method with two components (0.8811 for PCC, 0.2369 for RSE), for the AND logical operator group for the mixture model with three components (0.8151 for PCC, 0.3376 for RSE), for the AND logical operator group for the mixture model with four components (0.8047 for PCC, 0.3564 for RSE), for the AND logical operator group for the mixture model with five components (0.8031 for PCC, 0.3596 for RSE), for the AND logical operator group for random forest (0.8443 for PCC, 0.2875 for RSE), for the OR logical operator group for random forest (0.8442 for PCC, 0.2878 for RSE) and for the XOR logical operator group for random forest (0.8429 for PCC, 0.2899 for RSE).

7.2.3. Comparison between Weights Generated from Logical Groups

In this section, we compare the prediction accuracy forthe synthesized weights generated based on logical operators, given the same dataset and method. For the regression-based methods and comparison methods, we performed a pairwise comparison between the logical operators AND, OR and XOR with respect to the prediction accuracy of each method, as presented in Table 7 and Table 8, respectively. For each method X, we display its average PCC for each pair of logical operator AND, OR and XOR weight generation processes (denoted as AND-X, OR-X and XOR-X, respectively) side by side. For each dataset (row), the best-performing logical operator in each pair is indicated in bold, whereas the comparisons for which we rejected the hypothesis that the means were equal with a statistical significance of 5% are highlighted in gray.
As in the previous section, we leave the results regarding the RSE metric for the Appendix A, which can be found in Table A3 and Table A4. For each regression-based method or comparison method X, these tables display the average RSE for each of the pairs of weight generation processes, as for the PCC metric.
When comparing the performance of the AND and OR logical operators, the weight generation process with the AND logical operator had a significantly higher accuracy (in terms of PCC or RSE) at a significance level of 5% than the weight generation process with the OR logical operator in 61.11% (in terms of PCC and RSE) of all 36 dataset and method combinations, whereas the weight generation with the OR logical operator was significantly easier to predict in 5.56% (in terms of PCC) and 8.33% (in terms of RSE) of all 36 dataset and method combinations. For example, for Poisson regression on BookCross the weights generated using the AND logical operator had a PCC of 0.6381 and an RSE of 0.5977, whereas the weights generated using the OR logical operator had a lower PCC of 0.4476 and a higher RSE of 0.8480. For mixture models with two components on CompSciFac, the weights generated using the AND logical operator had a PCC of 0.3175 and an RSE of 0.9033, whereas the weights generated using the OR logical operator had a lower PCC of 0.2917 and a higher RSE of 0.9181. Moreover, for mixture models with three components on HistFac, the weights generated using the AND logical operator had a PCC of 0.3486 and an RSE of 0.9013, and the weights generated using the OR logical operator had a lower PCC of 0.1462 and a higher RSE of 0.9846.
As for the comparison between the AND and XOR operators, the weight generation process with the AND logical operator had a significantly higher accuracy (in terms of PCC or RSE) at a significance level of 5% than the weight generation process with the XOR logical operator in 47.22% (in terms of PCC) and in 52.78% (in terms of RSE) of combinations between all 36 datasets and methods, whereas the XOR had higher prediction accuracy in in 8.33% (in terms of PCC) and 5.56% (in terms of RSE) of cases. For example, for Poisson regression on BookCross, the weight generation method with the AND logical operator had significantly higher prediction accuracy with a PCC of 0.6381 and an RSE of 0.5977 than the weight generation with the XOR logical operators with a lower PCC of 0.3108 and a higher RSE of 1.3553. Furthermore, for mixture models with two components on HistFac the weights generated using the AND logical operator had a PCC of 0.3091 and an RSE of 0.9154, whereas the weights generated using the XOR logical operator had a lower PCC of 0.1427 and a higher RSE of 0.9853. For mixture models with four components on BookCross the weights generated using the AND logical operator had a PCC of 0.8047 and an RSE of 0.3564 and the weights generated using the XOR logical operator had a lower PCC of 0.5592 and a higher RSE of 0.6903. For mixture models with five components on CompSciFac the weights generated using the AND logical operator had a PCC of 0.3371 and an RSE of 0.8911, and the weights generated using the XOR logical operator had a lower PCC of 0.2819 and a higher RSE of 0.9248.
Finally, when comparing the OR and XOR logical operators, none of the two weight generation process presented higher prediction accuracy for most cases. The OR logical operator had a significantly higher accuracy (in terms of PCC or RSE) at a significance level of 5% than the weight generation process with the XOR logical operator in 11.11% (in terms of PCC) and 5.56% (in terms of RSE) of all three dataset and method combinations, whereas the XOR had significantly higher prediction accuracy in in 11.11% (in terms of PCC) and 22.22% (in terms of RSE) of cases. For example, the weight generation method with the OR logical operator had significantly higher prediction accuracy than the weight generation with the XOR logical operators for mixture model with two components on BusFac, where the prediction of weights generated using the OR logical operator had a PCC of 0.0780 and an RSE of 0.9979 and of the weights generated using the XOR logical operator had a lower PCC of 0.0546 and a higher RSE of 1.0029. On the other hand, for mixture models with five components on HistFac, the weights generated using the XOR logical operator had a PCC of 0.1546 and an RSE of 0.9837, whereas the weights generated using the OR logical operator had a lower PCC of 0.1115 and a higher RSE of 0.9980.
Overall, we noted that the regression-based methods had higher and significant prediction accuracy over the synthesized weights based on the AND logical operator when compared pairwise to the OR and XOR operators, as can be observed from the first two columns in Table 7. Notably, for all datasets, the neural networks with one and two hidden layers also had significantly higher prediction accuracy for the weights generated by the AND operator in comparison to the OR one, as shown in Table 8. The same pattern was repeated for the comparison methods random Forest and SVM but the differences in prediction accuracy were not significant. Given that some of the features used for prediction were based on the metadata similarity of end nodes, the fact that the weights generated by the AND operator were directly related to another measure of similarity between the metadata of the end nodes may have contributed to the better performance of the methods in these cases.

8. Discussion

Metadata are still not as widely available as the information about the node and edge structures of networks. However, as this information becomes increasingly available in real-world datasets, it is important to understand whether metadata have a role in improving the prediction of network properties, such as link weights. As a preliminary step, we performed experiments on four empirical weighted networks to address the question of whether there are accuracy gains with the addition of features based on node metadata when using common supervised machine learning methods.
Based on the set of experiments performed, we concluded that there exists a set of supervised methods and real-world networks for which the inclusion of metadata similarity does not improve prediction performance. Though many networks nowadays are far larger than the ones analyzed, the networks studied can provide insights about important properties of larger networks, such as representing community structures present in those networks. As more public datasets of weighted networks with node metadata become available, we highly encourage further investigations with a larger range of methods and datasets.
To further explore the role of node metadata in weight prediction, we analyzed the extreme case in which the weights depended solely on the metadata of the end nodes. Different relationships between synthesized weights and metadata were encoded using different logical operators for the weight generation process. We observed that, for all datasets and methods studied, the prediction accuracy over the original weights was significantly higher than for any of the synthesized weights. This might be related to a higher explanatory power of the topological features over the original weight generation process as compared to the processes that were solely dependent on metadata. Thus, in real-world networks similar to the ones studied, if the weights are solely dependent on the metadata of end nodes, the prediction accuracy using topological and metadata similarity features is likely to have a degraded performance.
In future work, it would be interesting to expand this analysis to real-world networks with different properties and characteristics (such as average weight, assortativity coefficient, etc.). Furthermore, it would be worthwhile to make comparisons with different supervised methods, such as the negative binomial regression (as discussed in Section 4) and end-to-end learning methods, instead of manually choosing and computing the input features.

Author Contributions

Conceptualization, M.V., L.M. and T.A.P.; methodology, M.V., L.M. and T.A.P.; software, L.M., K.O., T.A.P.; formal analysis, M.V., L.M., K.O.; investigation, L.M. and K.O.; data curation, L.M. and T.A.P.; writing—original draft preparation, L.M. and K.O.; writing—review and editing, L.M., K.O., M.V. and T.A.P.; visualization, L.M. and K.O.; supervision, M.V. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: http://tuvalu.santafe.edu/aaronc/facultyhiring/ (for BusFac, CompSciFac and HistFac datasets); and http://www2.informatik.uni-freiburg.de/cziegler/BX/ (for BookCross dataset), both accessed on 17 February 2022.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
RMSERoot Mean Squared Error
RSERelative Squared Error
PCCPearson correlation coefficient
PoiPoisson regression
Mix#Poisson mixture models
RFRandom forest
SVMSupport vector machines
NN#Artificial neural networks
CSCosine similarity
| V | number of nodes in the network
| E | number of links in the network
k average degree
S average link weight
rassortativity
mXmethod X with metadata and topological features, where X is Poi, Mix#, RF or SVM
tXmethod X with topological features, where X is Poi, Mix#, RF or SVM
ORIG-Xweights predicted using the original method, where X is Poi, Mix#, RF or SVM
AND-Xweights predicted using the AND logical operator method, where X is Poi, Mix#, RF or SVM
OR-Xweights predicted using the OR logical operator method, where X is Poi, Mix#, RF or SVM
XOR-Xweights predicted using the XOR logical operator method, where X is Poi, Mix#, RF or SVM

Appendix A. Tables with Results Obtained Using the RSE Metric

Consistent with the PCC values, regarding the RSE values for both the regression-based methods (Table A1) and the comparison methods (Table A2), we concluded that, for each dataset, method and metric, the prediction accuracy over original weights was higher than for each of the logical operator groups (for both PCC and RSE metrics). In all cases, we rejected the null hypothesis at a significance level of 5%.
Table A1. RSE comparison of original and generated weights usingregression-based methods. Notation: ORIG-Poi denotes the method Poi with the original weights, AND-Poi denotes the method Poi with weights generated using the AND logical operator, OR-Poi denotes the method Poi with weights generated using the OR logical operator and XOR-Poi denotes the method Poi with weights generated using the XOR logical operator. The null hypothesis was that the means were equal between the original weight generation process and each of the logical operator weight generation methods with a statistical significance of 5% in each comparison.
Table A1. RSE comparison of original and generated weights usingregression-based methods. Notation: ORIG-Poi denotes the method Poi with the original weights, AND-Poi denotes the method Poi with weights generated using the AND logical operator, OR-Poi denotes the method Poi with weights generated using the OR logical operator and XOR-Poi denotes the method Poi with weights generated using the XOR logical operator. The null hypothesis was that the means were equal between the original weight generation process and each of the logical operator weight generation methods with a statistical significance of 5% in each comparison.
ORIG-PoiAND-PoiOR-PoiXOR-PoiORIG-Mix2AND-Mix2OR-Mix2XOR-Mix2
BusFac0.6756 1.0098 1.0024 1.00310.41241.0041 0.9979 1.0029
CompSciFac0.7432 0.9054 0.9312 0.91630.46150.9033 0.9181 0.9306
HistFac0.5971 0.9806 0.9872 0.98140.5359 0.9154 0.9872 0.9853
BookCross0.1673 0.5977 0.8480 1.35530.16530.2369 0.7463 0.7195
ORIG-Mix3AND-Mix3OR-Mix3XOR-Mix3ORIG-Mix4AND-Mix4OR-Mix4XOR-Mix4
BusFac0.3750 0.9990 1.0064 1.0024 0.38271.0016 1.0019 1.0047
CompSciFac0.45470.8959 0.9092 0.9156 0.3608 0.8982 0.9158 0.9192
HistFac0.4823 0.9013 0.9846 0.9867 0.43520.9427 0.9789 0.9896
BookCross0.1622 0.3376 0.7262 0.7082 0.15930.3564 0.6895 0.6903
ORIG-Mix5AND-Mix5OR-Mix5XOR-Mix5
BusFac0.3641 1.0007 0.9999 1.0036
CompSciFac0.3656 0.8911 0.9241 0.9248
HistFac0.4254 0.9527 0.9980 0.9837
BookCross0.16400.3596 0.7126 0.6670
The results for the regression-based methods and comparison methods are displayed in Table A3 and Table A4, respectively. We display the average RSE for each method and each pair of logical operator weight generation processes side by side. As for the PCC table, the best-performing results for each pair is indicated in bold, and in gray we indicate the pairs for which we rejected the null hypothesis that the means were equal between the two weight generation processes with a statistical significance of 5%.
We refer the reader to the main text for the conclusions regarding the comparison between the original weights and the synthesized weights, as well as the pairwise comparison between synthesized weights, as those are consistent with the results obtained with the PCC metric.
Table A2. RSE comparison of original and generated weights usingcomparison methodsRF, SVM, NN1 and NN2.Notation: ORIG-NN1 denotes the method NN1 with original weights, AND-NN1 denotes the method NN1 with weights generated using the AND logical operator, OR-NN1 denotes the method NN1 with weights generated using the OR logical operator and XOR-NN1 denotes the method NN1 with weights generated using the XOR logical operator. The null hypothesis was that the means were equal between the original weight generation process and each of the logical operator weight generation methods with a statistical significance of 5% in each comparison.
Table A2. RSE comparison of original and generated weights usingcomparison methodsRF, SVM, NN1 and NN2.Notation: ORIG-NN1 denotes the method NN1 with original weights, AND-NN1 denotes the method NN1 with weights generated using the AND logical operator, OR-NN1 denotes the method NN1 with weights generated using the OR logical operator and XOR-NN1 denotes the method NN1 with weights generated using the XOR logical operator. The null hypothesis was that the means were equal between the original weight generation process and each of the logical operator weight generation methods with a statistical significance of 5% in each comparison.
ORIG-RFAND-RFOR-RFXOR-RFORIG-SVMAND-SVMOR-SVMXOR-SVM
BusFac0.59860.8927 0.9020 0.9010 0.6441 1.2274 1.2303 1.2218
CompSciFac0.6196 0.7633 0.7412 1.0005 0.7479 1.0005 0.9995 1.0128
HistFac0.61710.8790 0.8878 0.8899 0.6083 1.2110 1.2084 1.2061
BookCross0.1669 0.2875 0.2878 0.2899 0.1559 0.5550 0.5563 0.5560
ORIG-NN1AND-NN1OR-NN1XOR-NN1ORIG-NN2AND-NN2OR-NN2XOR-NN2
BusFac0.6050 1.01591.2355 1.0730 0.5823 1.0090 1.1301 1.1045
CompSciFac0.72350.8928 1.0302 0.9575 0.74340.8938 1.0544 0.9488
HistFac0.62180.9912 1.2028 1.0733 0.64370.9924 1.2373 1.0753
BookCross0.16590.82150.8978 0.7626 0.1615 0.8207 0.89270.7732
Table A3. RSE pairwise comparison between weights generated by logical operators usingregression-based methods. Notation: AND-Poi denotes the method Poi with weights generated using the AND logical operator, OR-Poi denotes the method Poi with weights generated using the OR logical operator and XOR-Poi denotes the method Poi with weights generated using the XOR logical operator.
Table A3. RSE pairwise comparison between weights generated by logical operators usingregression-based methods. Notation: AND-Poi denotes the method Poi with weights generated using the AND logical operator, OR-Poi denotes the method Poi with weights generated using the OR logical operator and XOR-Poi denotes the method Poi with weights generated using the XOR logical operator.
AND-PoiOR-PoiAND-PoiXOR-PoiOR-PoiXOR-Poi
BusFac 1.0098 1.00241.00981.00311.00241.0031
CompSciFac0.9054 0.9312 0.90540.91630.93120.9163
HistFac0.98060.98720.98060.98140.98720.9814
BookCross0.5977 0.8480 0.59771.35530.8480 1.3553
AND-Mix2OR-Mix2AND-Mix2XOR-Mix2OR-Mix2XOR-Mix2
BusFac 1.0041 0.99791.00411.00290.99791.0029
CompSciFac0.9033 0.91810.9033 0.9306 0.91810.9306
HistFac0.9154 0.9872 0.9154 0.9853 0.98720.9853
BookCross0.2369 0.74630.2369 0.7195 0.74630.7195
AND-Mix3OR-Mix3AND-Mix3XOR-Mix3OR-Mix3XOR-Mix3
BusFac0.9990 1.0064 0.99901.00241.00641.0024
CompSciFac0.89590.90920.8959 0.9156 0.90910.9156
HistFac0.90130.9846 0.9013 0.9867 0.98460.9867
BookCross0.3376 0.7262 0.33760.7082 0.72610.7082
AND-Mix4OR-Mix4AND-Mix4XOR-Mix4OR-Mix4XOR-Mix4
BusFac1.00161.00191.00161.00471.00191.0047
CompSciFac0.8982 0.9157 0.8982 0.91920.91570.9192
HistFac0.9427 0.97890.9427 0.9897 0.9789 0.9897
BookCross0.3564 0.6895 0.35640.6903 0.68950.6903
AND-Mix5OR-Mix5AND-Mix5XOR-Mix5OR-Mix5XOR-Mix5
BusFac1.00070.99991.00071.00361.00361.0036
CompSciFac0.8911 0.92410.89110.92480.92410.9248
HistFac0.9527 0.9980 0.9527 0.9837 0.9980 0.9837
BookCross0.3596 0.7126 0.3596 0.6670 0.71260.6670
Table A4. RSE pairwise comparison between weights generated by logical operators using comparison methods RF, SVM, NN1 and NN2.Notation: AND-RF denotes the method RF with weights generated using the AND logical operator, OR-RF denotes the method RF with weights generated using the OR logical operator and XOR-RF denotes the method RF with weights generated using the XOR logical operator.
Table A4. RSE pairwise comparison between weights generated by logical operators using comparison methods RF, SVM, NN1 and NN2.Notation: AND-RF denotes the method RF with weights generated using the AND logical operator, OR-RF denotes the method RF with weights generated using the OR logical operator and XOR-RF denotes the method RF with weights generated using the XOR logical operator.
AND-RFOR-RFAND-RFXOR-RFOR-RFXOR-RF
BusFac0.89270.90200.89270.90100.90200.9010
CompSciFac 0.7633 0.74120.76330.75820.74120.7582
HistFac0.87900.88780.87900.88990.88780.8899
BookCross0.28750.28780.28750.28980.28780.2898
AND-SVMOR-SVMAND-SVMXOR-SVMOR-SVMXOR-SVM
BusFac1.22741.23031.22741.22181.23031.2218
CompSciFac1.00050.99951.00051.01280.99951.0128
HistFac1.21101.20841.21101.20611.20841.2061
BookCross0.55500.55630.55500.55590.55630.5559
AND-NN1OR-NN1AND-NN1XOR-NN1OR-NN1XOR-NN1
BusFac1.0159 1.2355 1.0159 1.0730 1.23551.0730
CompSciFac0.8928 1.0302 0.8928 0.9575 1.0302 0.9575
HistFac0.9912 1.2028 0.9912 1.0733 1.20281.0733
BookCross0.8215 0.8978 0.82150.7626 0.89780.7626
AND-NN2OR-NN2AND-NN2XOR-NN2OR-NN2XOR-NN2
BusFac1.0090 1.1301 1.00901.1045 1.13011.1045
CompSciFac0.8938 1.0544 0.8938 0.94881.0544 0.9488
HistFac0.9924 1.2373 0.9924 1.0753 1.2373 1.0753
BookCross0.8207 0.8927 0.8207 0.7732 0.8927 0.7732

References

  1. Peel, L.; Larremore, D.B.; Clauset, A. The ground truth about metadata and community detection in networks. Sci. Adv. 2017, 3, e1602548. [Google Scholar] [CrossRef] [Green Version]
  2. Fajardo-Fontiveros, O.; Sales-Pardo, M.; Guimera, R. Node metadata can produce predictability transitions in network inference problems. arXiv 2021, arXiv:2103.14424. [Google Scholar]
  3. Zhu, B.; Xia, Y.; Zhang, X.J. Weight Prediction in Complex Networks Based on Neighbor Set. Sci. Rep. 2016, 6, 38080. [Google Scholar] [CrossRef] [Green Version]
  4. Wolpert, D.; Macready, W. No Free Lunch Theorems for Optimization. IEEE Trans. Evol. Comput. 1997, 1, 67–82. [Google Scholar] [CrossRef] [Green Version]
  5. Zhao, J.; Miao, L.; Yang, J.; Fang, H.; Zhang, Q.M.; Nie, M.; Holme, P.; Zhou, T. Prediction of Links and Weights in Networks by Reliable Routes. Sci. Rep. 2015, 5, 12261. [Google Scholar] [CrossRef] [Green Version]
  6. Watts, D.J.; Strogatz, S.H. Collective dynamics of ‘small-world’networks. Nature 1998, 393, 440–442. [Google Scholar] [CrossRef]
  7. Batagelj, V.; Mrvar, A. Pajek Datasets. 2014. Available online: https://vlado.fmf.uni-lj.si/pub/networks/data/ (accessed on 18 July 2021).
  8. Kunegis, J. Konect: The koblenz network collection. In Proceedings of the 22nd International Conference on World Wide Web, Rio de Janeiro, Brazil, 13–17 May 2013; pp. 1343–1350. Available online: https://konect.uni-koblenz.de/ (accessed on 5 April 2022).
  9. Havugimana, P.C.; Hart, G.T.; Nepusz, T.; Yang, H.; Turinsky, A.L.; Li, Z.; Wang, P.I.; Boutz, D.R.; Fong, V.; Phanse, S.; et al. A census of human soluble protein complexes. Cell 2012, 150, 1068–1081. [Google Scholar] [CrossRef] [Green Version]
  10. Liben-Nowell, D.; Kleinberg, J. The Link-Prediction Problem for Social Networks. J. Am. Soc. Inf. Sci. Technol. 2007, 58, 1019–1031. [Google Scholar] [CrossRef] [Green Version]
  11. Clauset, A.; Moore, C.; Newman, M.E.J. Hierarchical Structure and the Prediction of Missing Links in Networks. Nature 2008, 453, 98–101. [Google Scholar] [CrossRef]
  12. Zhou, T.; Lu, L.; Zhang, Y.C. Predicting Missing Links via Local Information. Eur. Phys. J. B 2009, 71, 623–630. [Google Scholar] [CrossRef] [Green Version]
  13. Guimerà, R.; Sales-Pardo, M. Missing and Spurious Interactions and the Reconstruction of Complex Networks. Proc. Natl. Acad. Sci. USA 2009, 106, 22073–22078. [Google Scholar] [CrossRef] [Green Version]
  14. Lu, L.; Zhou, T. Link Prediction in Complex Networks: A Survey. Phys. A Stat. Mech. Its Appl. 2011, 390, 1150–1170. [Google Scholar] [CrossRef] [Green Version]
  15. Zhou, T. Progresses and Challenges in Link Prediction. iScience 2021, 24, 103217. [Google Scholar] [CrossRef]
  16. Aicher, C.; Jacobs, A.Z.; Clauset, A. Learning Latent Block Structure in Weighted Networks. J. Complex Netw. 2015, 3, 221–248. [Google Scholar] [CrossRef] [Green Version]
  17. Fu, C.; Zhao, M.; Fan, L.; Chen, X.; Chen, J.; Wu, Z.; Xia, Y.; Xuan, Q. Link Weight Prediction Using Supervised Learning Methods and Its Application to Yelp Layered Network. IEEE Trans. Knowl. Data Eng. 2018, 30, 1507–1518. [Google Scholar] [CrossRef]
  18. Kumar, S.; Spezzano, F.; Subrahmanian, V.S.; Faloutsos, C. Edge Weight Prediction in Weighted Signed Networks. In Proceedings of the 2016 IEEE 16th International Conference on Data Mining (ICDM), Barcelona, Spain, 12–15 December 2016; IEEE: Barcelona, Spain, 2016; pp. 221–230. [Google Scholar] [CrossRef]
  19. Malla, R.; Durga Bhavani, S. Link Weight Prediction for Directed WSN Using Features from Network and Its Dual. In Pattern Recognition and Machine Intelligence; Deka, B., Maji, P., Mitra, S., Bhattacharyya, D.K., Bora, P.K., Pal, S.K., Eds.; Springer International Publishing: Cham, Switzerland, 2019; Volume 11941, pp. 56–64. [Google Scholar] [CrossRef]
  20. Zhao, Y.; Wu, Y.J.; Levina, E.; Zhu, J. Link Prediction for Partially Observed Networks. J. Comput. Graph. Stat. 2017, 26, 725–733. [Google Scholar] [CrossRef] [Green Version]
  21. Popescul, A.; Ungar, L.H. Statistical Relational Learning for Link Prediction. In Proceedings of the Workshop on Learning Statistical Models from Relational Data at IJCAI-2003, Acapulco, Mexico, 9–11 August 2003; p. 7. [Google Scholar]
  22. Taskar, B.; Wong, M.F.; Abbeel, P.; Koller, D. Link Prediction in Relational Data. Adv. Neural Inf. Processing Syst. 2003, 16, 8. [Google Scholar]
  23. Kim, D.I.; Hughes, M.C.; Sudderth, E.B. The Nonparametric Metadata Dependent Relational Model. arXiv 2012, arXiv:1206.6414. [Google Scholar]
  24. Fan, X.; Xu, R.Y.D.; Cao, L.; Song, Y. Learning Nonparametric Relational Models by Conjugately Incorporating Node Information in a Network. IEEE Trans. Cybern. 2017, 47, 589–599. [Google Scholar] [CrossRef] [Green Version]
  25. Zhao, H.; Du, L.; Buntine, W. Leveraging Node Attributes for Incomplete Relational Data. arXiv 2017, arXiv:1706.04289. [Google Scholar]
  26. Zhang, M.; Chen, Y. Link Prediction Based on Graph Neural Networks. Adv. Neural Inf. Processing Syst. 2018, 31, 11. [Google Scholar]
  27. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; Available online: http://www.deeplearningbook.org (accessed on 2 April 2022).
  28. Newman, M.E.J. Clustering and preferential attachment in growing networks. Phys. Rev. E 2001, 64, 025102. [Google Scholar] [CrossRef] [Green Version]
  29. Jaccard, P. Étude comparative de la distribution florale dans une portion des Alpes et des Jura. Bull. Soc. Vaudoise Sci. Nat. 1901, 37, 547–579. [Google Scholar]
  30. Adamic, L.A.; Adar, E. Friends and neighbors on the web. Soc. Netw. 2003, 25, 211–230. [Google Scholar] [CrossRef] [Green Version]
  31. Salton, G.; McGill, M.J. Introduction to Modern Information Retrieval; McGraw-Hill: New York, NY, USA, 1983. [Google Scholar]
  32. Sorensen, T.A. A method of establishing groups of equal amplitude in plant sociology based on similarity of species content and its application to analyses of the vegetation on Danish commons. Biol. Skar. 1948, 5, 1–34. [Google Scholar]
  33. Ravasz, E.; Somera, A.L.; Mongru, D.A.; Oltvai, Z.N.; Barabási, A. Hierarchical organization of modularity in metabolic networks. Science 2002, 297, 1551–1555. [Google Scholar] [CrossRef] [Green Version]
  34. Leicht, E.A.; Holme, P.; Newman, M.E.J. Vertex similarity in networks. Phys. Rev. E 2006, 73, 026120. [Google Scholar] [CrossRef] [Green Version]
  35. Barabási, A.L.; Albert, R. Emergence of scaling in random networks. Science 1999, 286, 509–512. [Google Scholar] [CrossRef] [Green Version]
  36. Ziegler, C.N.; McNee, S.M.; Konstan, J.A.; Lausen, G. Improving recommendation lists through topic diversification. In Proceedings of the 14th International Conference on World Wide Web, Chiba, Japan, 10–14 May 2005; pp. 22–32. [Google Scholar]
  37. Newman, M.E.J. Assortative mixing in networks. Phys. Rev. Lett. 2002, 89, 208701. [Google Scholar] [CrossRef] [Green Version]
Figure 1. PCC (top) and RSE (bottom) for methods with topological features. For each dataset and accuracy metric, the above figure presents a boxplot comparison of 30 iterations of the results. The best mean prediction accuracy is indicated with an asterisk. The mixture models with three, five and five components had the highest accuracy among the methods for both metrics for the BusFac, CompSciFac and HistFac datasets, respectively. Meanwhile, SVM had the highest accuracy among methods for both metrics for the BookCross dataset. Thus, with the benefit of statistical interpretability, mixture models can achieve the best performance in comparison to other common supervised learning methods for some datasets.
Figure 1. PCC (top) and RSE (bottom) for methods with topological features. For each dataset and accuracy metric, the above figure presents a boxplot comparison of 30 iterations of the results. The best mean prediction accuracy is indicated with an asterisk. The mixture models with three, five and five components had the highest accuracy among the methods for both metrics for the BusFac, CompSciFac and HistFac datasets, respectively. Meanwhile, SVM had the highest accuracy among methods for both metrics for the BookCross dataset. Thus, with the benefit of statistical interpretability, mixture models can achieve the best performance in comparison to other common supervised learning methods for some datasets.
Entropy 24 00842 g001
Figure 2. PCC (left) and RSE (right) comparison of weight generation processes for theBusFacdataset. Each column represents a comparison of the results for each accuracy metric and the BusFac dataset. For the BusFac dataset, the methods with highest accuracy for the PCC metrics were the mixture model with 5 components for the ORIG weights and random forest for the AND weights, OR weights and XOR weights. For the BusFac dataset, the methods with the highest accuracy for the RSE metrics were the mixture model with 5 components for the ORIG weights and random forest for the AND weights, OR weights and XOR weights.
Figure 2. PCC (left) and RSE (right) comparison of weight generation processes for theBusFacdataset. Each column represents a comparison of the results for each accuracy metric and the BusFac dataset. For the BusFac dataset, the methods with highest accuracy for the PCC metrics were the mixture model with 5 components for the ORIG weights and random forest for the AND weights, OR weights and XOR weights. For the BusFac dataset, the methods with the highest accuracy for the RSE metrics were the mixture model with 5 components for the ORIG weights and random forest for the AND weights, OR weights and XOR weights.
Entropy 24 00842 g002
Figure 3. PCC (left) and RSE (right) comparison of the weight generation processes for theCompSciFacdataset. Each column represents a comparison of the results for 30 iterations for each accuracy metric and the CompSciFac dataset. For the CompSciFac dataset, the methods with the highest accuracy for the PCC metrics were the mixture model with four components for the ORIG weights and random forest for the AND weights, OR weights and XOR weights. For the CompSciFac dataset, the methods with highest accuracy for the RSE metrics were the mixture model with four components for the ORIG weights and random forest for the AND weights, OR weights and XOR weights.
Figure 3. PCC (left) and RSE (right) comparison of the weight generation processes for theCompSciFacdataset. Each column represents a comparison of the results for 30 iterations for each accuracy metric and the CompSciFac dataset. For the CompSciFac dataset, the methods with the highest accuracy for the PCC metrics were the mixture model with four components for the ORIG weights and random forest for the AND weights, OR weights and XOR weights. For the CompSciFac dataset, the methods with highest accuracy for the RSE metrics were the mixture model with four components for the ORIG weights and random forest for the AND weights, OR weights and XOR weights.
Entropy 24 00842 g003
Figure 4. PCC (left) and RSE (right) comparison of the weight generation processes for theHistFacdataset. Each column represents a comparison of the results for 30 iterations for each accuracy metric and the HistFac dataset. For the HistFac dataset, the methods with the highest accuracy for the PCC were the mixture model with 5 components for the ORIG weights and random forest for the AND weights, OR weights and XOR weights. For the HistFac dataset, the methods with the highest accuracy for the RSE metrics were the mixture model with 5 components for the ORIG weights and random forest for the AND weights, OR weights and XOR weights.
Figure 4. PCC (left) and RSE (right) comparison of the weight generation processes for theHistFacdataset. Each column represents a comparison of the results for 30 iterations for each accuracy metric and the HistFac dataset. For the HistFac dataset, the methods with the highest accuracy for the PCC were the mixture model with 5 components for the ORIG weights and random forest for the AND weights, OR weights and XOR weights. For the HistFac dataset, the methods with the highest accuracy for the RSE metrics were the mixture model with 5 components for the ORIG weights and random forest for the AND weights, OR weights and XOR weights.
Entropy 24 00842 g004
Figure 5. PCC (left) and RSE (right) comparison of the weight generation processes for theBookCrossdataset. Each column represents a comparison of the results for 30 iterations for each accuracy metric and the BookCross dataset. For the BookCross dataset, the methods with the highest accuracy for the PCC metrics were the support vector machines for the ORIG weights, the mixture model with 2 components for the AND weights and random forest for the OR weights and XOR weights. For the BookCross dataset, the methods with the highest accuracy for the RSE metrics were the mixture model with 5 components for the ORIG weights, the mixture model with 2 components for the AND weights and random forest for the OR weights and XOR weights.
Figure 5. PCC (left) and RSE (right) comparison of the weight generation processes for theBookCrossdataset. Each column represents a comparison of the results for 30 iterations for each accuracy metric and the BookCross dataset. For the BookCross dataset, the methods with the highest accuracy for the PCC metrics were the support vector machines for the ORIG weights, the mixture model with 2 components for the AND weights and random forest for the OR weights and XOR weights. For the BookCross dataset, the methods with the highest accuracy for the RSE metrics were the mixture model with 5 components for the ORIG weights, the mixture model with 2 components for the AND weights and random forest for the OR weights and XOR weights.
Entropy 24 00842 g005
Table 1. Topological similarity measures used as baseline features.Notation: Γ ( v ) denotes the set of neighbors of a node v, k v the degree of a node v and | · | denotes a set’s cardinality.
Table 1. Topological similarity measures used as baseline features.Notation: Γ ( v ) denotes the set of neighbors of a node v, k v the degree of a node v and | · | denotes a set’s cardinality.
MeasureDefinitionDescription
Common neighbors [28] s u , v C N = | Γ ( u ) Γ ( v ) | Number of neighbors in common between two nodes.
Jaccard [29] s u , v J I = | Γ ( u ) Γ ( v ) | | Γ ( u ) Γ ( v ) | Number of neighbors in common divided by the total number of neighbors of the two nodes.
Adamic Adar [30] s u , v A A = z Γ ( u ) Γ ( v ) 1 log k z Number of common neighbors of the vertices, weighted by the inverse logarithm of their degrees.
Salton [31] s u , v S A = | Γ ( u ) Γ ( v ) | k u k v Measures the angle between two vectors.
Resource allocation [12] s u , v R A = z Γ ( u ) Γ ( v ) 1 k z Number of common neighbors of the vertices, inversely weighted by their degrees.
Sørensen [32] s u , v S I = 2 | Γ ( u ) Γ ( v ) | k u + k v Two times the number of common neighbors divided by the sum of the vertices’ degrees.
Hub depressed [33] s u , v H D I = | Γ ( u ) Γ ( v ) | max ( k u k v ) Links adjacent to hubs are have lower scores, since the denominator depends only on the higher degree.
Hub promoted [33] s u , v H P I = | Γ ( u ) Γ ( v ) | min ( k u , k v ) Links adjacent to hubs have higher scores, since the denominator depends only on the lower degree.
Leicht-Holme-Newman [34] s u , v L H N = | Γ ( u ) Γ ( v ) | k u k v Pairs of nodes with more common neighbors as a proportion of the expected number of common neighbors have higher scores.
Preferential attachment [35] s u , v P A = k u k v Product of the degrees of the two nodes.
Table 2. Topological information of real-world networks.Notation: | V | is the number of nodes, | E | the number of links, k the average degree, S the average link weight, and r the assortativity coefficient [37].
Table 2. Topological information of real-world networks.Notation: | V | is the number of nodes, | E | the number of links, k the average degree, S the average link weight, and r the assortativity coefficient [37].
| V | | E | k S r
BusFac113351562.2122.572−0.173
CompSciFac206286527.8161.741−0.109
HistFac145233432.1931.944−0.245
BookCross24026,380219.8333.656−0.050
Table 3. PCC (top) and RSE (bottom) comparisons for Poisson regression and mixture models with and without metadata similarity features.Notation: mPoi indicates the method Poi with metadata and topological features, and tPoi indicates the method Poi with only topological features (analogous notation for Mix#). For each method and dataset, the best-performing set of features is shown in bold, and the pairs for which we rejected the null hypothesis are highlighted in gray. Although not statistically significant most of the time, the regressions with only the topological features had higher accuracy than those obtained after the inclusion of metadata similarity in the majority of cases.
Table 3. PCC (top) and RSE (bottom) comparisons for Poisson regression and mixture models with and without metadata similarity features.Notation: mPoi indicates the method Poi with metadata and topological features, and tPoi indicates the method Poi with only topological features (analogous notation for Mix#). For each method and dataset, the best-performing set of features is shown in bold, and the pairs for which we rejected the null hypothesis are highlighted in gray. Although not statistically significant most of the time, the regressions with only the topological features had higher accuracy than those obtained after the inclusion of metadata similarity in the majority of cases.
mPoitPoimMix2tMix2mMix3tMix3mMix4tMix4mMix5tMix5
BusFac0.58040.54690.78890.80040.81310.81760.80380.81260.8070.8105
CompSciFac0.55420.53970.73220.77270.75810.77490.8030.8040.81790.805
HistFac0.62610.63580.69560.71680.70420.76230.7590.74310.74490.7576
BookCross0.90760.910.91060.9120.91160.91210.9130.91120.91260.9106
mPoitPoimMix2tMix2mMix3tMix3mMix4tMix4mMix5tMix5
BusFac0.6434 0.71620.39540.36950.39010.34210.37320.34940.35980.3501
CompSciFac0.72510.70490.48470.49060.47150.38120.38150.37240.32920.368
HistFac0.59440.59880.53620.58490.46010.48520.47670.48510.43850.4152
BookCross0.17860.1817 0.17270.16940.17050.17030.1680.17240.16910.1726
Table 4. PCC (top) and RSE (bottom) comparison for RF, SVM and NNs with and without metadata similarity features.Notation: mSVM indicates the method SVM with metadata and topological features, and tSVM indicates the method SVM with only topological features (analogous notation for RF). For each method and dataset, the best-performing set of features is shown in bold, and the pairs for which we rejected the null hypothesis are highlighted in gray. RF performed significantly better after the inclusion of metadata similarity features on BusFac and CompSciFac. On the other hand, SVM performed significantly worse with the inclusion of metadata features on BookCross. Most of the other comparisons did not have statistically significant results.
Table 4. PCC (top) and RSE (bottom) comparison for RF, SVM and NNs with and without metadata similarity features.Notation: mSVM indicates the method SVM with metadata and topological features, and tSVM indicates the method SVM with only topological features (analogous notation for RF). For each method and dataset, the best-performing set of features is shown in bold, and the pairs for which we rejected the null hypothesis are highlighted in gray. RF performed significantly better after the inclusion of metadata similarity features on BusFac and CompSciFac. On the other hand, SVM performed significantly worse with the inclusion of metadata features on BookCross. Most of the other comparisons did not have statistically significant results.
mRFtRFmSVMtSVMmNN1tNN1mNN2tNN2
BusFac0.6719 0.56550.63080.60860.66070.60490.68050.5968
CompSciFac0.65220.57450.53580.5610.57000.56530.56410.5520
HistFac0.61280.60240.66110.62740.640140.65940.64600.6493
BookCross0.90910.9097 0.91580.91780.91750.91830.91900.9157
mRFtRFmSVMtSVMmNN1tNN1mNN2tNN2
BusFac0.5677 0.71160.63950.69570.6050 0.68180.5823 0.6741
CompSciFac0.6103 0.72390.76840.74610.72350.72400.75600.7440
HistFac0.64810.65260.60440.64190.62180.61950.64370.6327
BookCross0.1740.173 0.16310.15970.16590.16180.1615 0.1697
Table 5. PCC comparison of original and generated weights usingregression-based methods.Notation: ORIG-Poi denotes the method Poi with weights generated using the original dataset, AND-Poi denotes the method Poi with weights generated using the AND logical operator, OR-Poi denotes the method Poi with weights generated using the OR logical operator and XOR-Poi denotes the method Poi with weights generated using the XOR logical operator. The null hypothesis was that the means were equal between the original weight generation process and each of the logical operator weight generation methods with a statistical significance of 5% in each comparison.
Table 5. PCC comparison of original and generated weights usingregression-based methods.Notation: ORIG-Poi denotes the method Poi with weights generated using the original dataset, AND-Poi denotes the method Poi with weights generated using the AND logical operator, OR-Poi denotes the method Poi with weights generated using the OR logical operator and XOR-Poi denotes the method Poi with weights generated using the XOR logical operator. The null hypothesis was that the means were equal between the original weight generation process and each of the logical operator weight generation methods with a statistical significance of 5% in each comparison.
ORIG-PoiAND-PoiOR-PoiXOR-PoiORIG-Mix2AND-Mix2OR-Mix2XOR-Mix2
BusFac0.5784 0.0409 0.06180.06420.77920.05270.0781 0.0546
CompSciFac0.5347 0.31380.2756 0.29830.74710.3175 0.29170.2756
HistFac0.6465 0.15480.1370 0.15140.6916 0.30910.14450.1427
BookCross0.91400.6381 0.44760.31090.91440.88110.5379 0.5389
ORIG-Mix3AND-Mix3OR-Mix3XOR-Mix3ORIG-Mix4AND-Mix4OR-Mix4XOR-Mix4
BusFac0.8030 0.07980.0470 0.06230.79510.0647 0.06010.0534
CompSciFac0.7572 0.3300 0.3091 0.29960.81610.3278 0.29600.2963
HistFac0.7252 0.3486 0.14620.14350.75930.25720.1613 0.1276
BookCross0.9160 0.81510.56820.54640.9177 0.8047 0.5711 0.5592
ORIG-Mix5AND-Mix5OR-Mix5XOR-Mix5
BusFac0.8032 0.0655 0.07550.0561
CompSciFac0.8022 0.33710.28190.2819
HistFac0.7697 0.2333 0.11150.1546
BookCross0.9157 0.80310.57600.5781
Table 6. PCC comparison of original and generated weights usingcomparison methodsRF, SVM, NN1 and NN2.Notation: ORIG-RF denotes the method RF with weights generated using the original dataset, AND-RF denotes the method RF with weights generated using the AND logical operator, OR-RF denotes the method RF with weights generated using the OR logical operator and XOR-RF denotes the method RF with weights generated using the XOR logical operator. The null hypothesis was that the means were equal between the original weight generation process and each of the logical operator weight generation methods with a statistical significance of 5% in each comparison.
Table 6. PCC comparison of original and generated weights usingcomparison methodsRF, SVM, NN1 and NN2.Notation: ORIG-RF denotes the method RF with weights generated using the original dataset, AND-RF denotes the method RF with weights generated using the AND logical operator, OR-RF denotes the method RF with weights generated using the OR logical operator and XOR-RF denotes the method RF with weights generated using the XOR logical operator. The null hypothesis was that the means were equal between the original weight generation process and each of the logical operator weight generation methods with a statistical significance of 5% in each comparison.
ORIG-RFAND-RFOR-RFXOR-RFORIG-SVMAND-SVMOR-SVMXOR-SVM
BusFac0.64730.3338 0.3254 0.33030.6376 0.03320.03100.0331
CompSciFac0.6510 0.49220.51200.49570.5554 0.37420.3734 0.3615
HistFac0.6306 0.35990.3513 0.35030.6511 0.1589 0.14930.1514
BookCross0.9133 0.8443 0.8442 0.84290.9197 0.72700.72530.7253
ORIG-NN1AND-NN1OR-NN1XOR-NN1ORIG-NN2AND-NN2OR-NN2XOR-NN2
BusFac0.6607 0.0431 -0.00320.02080.68050.0499 0.0109 0.0243
CompSciFac0.5700 0.3562 0.2960 0.31650.56410.3506 0.3051 0.3251
HistFac0.6401 0.17510.1233 0.13060.64600.1772 0.0950 0.1331
BookCross0.9175 0.42680.3329 0.49480.9190 0.4270 0.33780.4850
Table 7. PCC pairwise comparison between weights generated by logical operators usingregression-based methods. The regression-based method had higher and significant prediction accuracy with the weight generation process using the AND logical operator than the OR and XOR weight generation processes. The weight generation process using the AND logical operator was significantly easier to predict than the weight generation process using the OR logical operator in 61.11%, and than the weight generation process using the XOR logical operator in 47.22%, of the 28 dataset and method combinations. Notation: AND-Poi denotes the method Poi with weights generated using the AND logical operator, OR-Poi denotes the method Poi with weights generated using the OR logical operator and XOR-Poi denotes the method Poi with weights generated using the XOR logical operator.
Table 7. PCC pairwise comparison between weights generated by logical operators usingregression-based methods. The regression-based method had higher and significant prediction accuracy with the weight generation process using the AND logical operator than the OR and XOR weight generation processes. The weight generation process using the AND logical operator was significantly easier to predict than the weight generation process using the OR logical operator in 61.11%, and than the weight generation process using the XOR logical operator in 47.22%, of the 28 dataset and method combinations. Notation: AND-Poi denotes the method Poi with weights generated using the AND logical operator, OR-Poi denotes the method Poi with weights generated using the OR logical operator and XOR-Poi denotes the method Poi with weights generated using the XOR logical operator.
AND-PoiOR-PoiAND-PoiXOR-PoiOR-PoiXOR-Poi
BusFac0.04090.06180.04090.06420.06180.0642
CompSciFac0.3138 0.27560.31380.29830.27560.2983
HistFac0.15480.13700.15480.15140.13700.1514
BookCross0.6381 0.44760.6381 0.31080.4476 0.3108
AND-Mix2OR-Mix2AND-Mix2XOR-Mix2OR-Mix2XOR-Mix2
BusFac 0.05270.07810.05270.05460.0780 0.0546
CompSciFac0.3175 0.29170.3175 0.27560.29170.2756
HistFac0.3091 0.14450.3091 0.14270.14440.1427
BookCross0.8811 0.53790.8811 0.5389 0.53790.5389
AND-Mix3OR-Mix3AND-Mix3XOR-Mix3OR-Mix3XOR-Mix3
BusFac0.0798 0.04700.07970.06230.04700.0623
CompSciFac0.33000.30900.3300 0.26170.30900.2996
HistFac0.3486 0.14620.3486 0.14350.14620.1435
BookCross0.81510.56820.8151 0.54640.56810.5464
AND-Mix4OR-Mix4AND-Mix4XOR-Mix4OR-Mix4XOR-Mix4
BusFac0.06470.06010.06470.05340.06010.0534
CompSciFac0.3278 0.29600.3278 0.29630.29600.2963
HistFac0.2572 0.16130.2572 0.12760.1613 0.1276
BookCross0.8047 0.57110.8047 0.55920.57110.5592
AND-Mix5OR-Mix5AND-Mix5XOR-Mix5OR-Mix5XOR-Mix5
BusFac0.06550.07550.06550.05610.0755 0.0561
CompSciFac0.3371 0.28190.3371 0.28190.28190.2819
HistFac0.2333 0.11150.2333 0.1546 0.11150.1546
BookCross0.8031 0.57600.80310.5781 0.57600.5781
Table 8. PCC pairwise comparison between weights generated by logical operators usingcomparison methodsRF, SVM, NN1 and NN2.Notation: AND-RF denotes the method RF with weights generated using the AND logical operator, OR-RF denotes the method RF with weights generated using the OR logical operator and XOR-RF denotes the method RF with weights generated using the XOR logical operator.
Table 8. PCC pairwise comparison between weights generated by logical operators usingcomparison methodsRF, SVM, NN1 and NN2.Notation: AND-RF denotes the method RF with weights generated using the AND logical operator, OR-RF denotes the method RF with weights generated using the OR logical operator and XOR-RF denotes the method RF with weights generated using the XOR logical operator.
AND-RFOR-RFAND-RFXOR-RFOR-RFXOR-RF
BusFac0.33380.32540.33380.33030.32540.3303
CompSciFac 0.49220.51200.49220.49570.51200.4957
HistFac0.35990.35130.35990.35030.35130.3503
BookCross0.84430.84410.84430.84290.84410.8429
AND-SVMOR-SVMAND-SVMXOR-SVMOR-SVMXOR-SVM
BusFac0.03320.03100.03320.03310.03100.0331
CompSciFac0.37420.37340.37420.36150.37340.3615
HistFac0.15890.14930.15890.15140.14930.1514
BookCross0.72700.72530.72700.72530.72530.7253
AND-NN1OR-NN1AND-NN1XOR-NN1OR-NN1XOR-NN1
BusFac0.0431 −0.00320.04310.0208−0.00320.0208
CompSciFac0.3562 0.29600.3562 0.31650.29600.3165
HistFac0.17510.12330.1751 0.13060.12330.1306
BookCross0.4268 0.3329 0.42680.4948 0.33290.4948
AND-NN2OR-NN2AND-NN2XOR-NN2OR-NN2XOR-NN2
BusFac0.0499 0.01090.0499 0.02430.01090.0243
CompSciFac0.3506 0.30510.35060.32510.30510.3251
HistFac0.1772 0.09500.1772 0.1331 0.09500.1331
BookCross0.4270 0.3378 0.42700.4850 0.33780.4850
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mori, L.; O’Hara, K.; Pujol, T.A.; Ventresca, M. Examining Supervised Machine Learning Methods for Integer Link Weight Prediction Using Node Metadata. Entropy 2022, 24, 842. https://doi.org/10.3390/e24060842

AMA Style

Mori L, O’Hara K, Pujol TA, Ventresca M. Examining Supervised Machine Learning Methods for Integer Link Weight Prediction Using Node Metadata. Entropy. 2022; 24(6):842. https://doi.org/10.3390/e24060842

Chicago/Turabian Style

Mori, Larissa, Kaleigh O’Hara, Toyya A. Pujol, and Mario Ventresca. 2022. "Examining Supervised Machine Learning Methods for Integer Link Weight Prediction Using Node Metadata" Entropy 24, no. 6: 842. https://doi.org/10.3390/e24060842

APA Style

Mori, L., O’Hara, K., Pujol, T. A., & Ventresca, M. (2022). Examining Supervised Machine Learning Methods for Integer Link Weight Prediction Using Node Metadata. Entropy, 24(6), 842. https://doi.org/10.3390/e24060842

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop