3.1. NFT Transfer Network
NFT transactions rely on smart contracts, automating operations and storing transaction records transparently on the blockchain. The ERC-721 standard [
7] defines the smart contract functionality required for NFT transactions, ensuring consistency in how NFTs are created, transferred, and tracked. A typical NFT transaction involves two key actions: the seller invokes the smart contract to transfer NFT ownership to the buyer. In contrast, simultaneously, the buyer’s account balance is transferred to the seller.
Figure 2 illustrates this process. Each transaction can be modeled as a directed network edge, where the seller represents the source node, the buyer is the target node, and the transaction details (e.g., timestamp, price) are encoded as edge attributes. Modeling NFT transfers as a network enables researchers to analyze market trends, detect trading risks (e.g., wash trading, fraud), and optimize platform strategies.
Based on the transfer data of NFTs, we define the NFT transfer network. The NFT transfer network consists of N nodes and M-directed edges without weights. Here, represents the specific transfer record of an NFT. We use an adjacency matrix, , to represent the topological structure of the NFT transfer network, where indicates a connection from node i to node j (i.e., the NFT seller transferring the NFT to the NFT buyer). Otherwise, .
Unlike dynamic networks, where edges evolve continuously, our analysis focuses on a static representation of the NFT transfer network. This static network aggregates all observed transactions within a predefined time window, capturing long-term structural relationships between traders rather than short-term fluctuations. While edge weights could reflect transaction frequency, our initial core–periphery detection phase adopts an unweighted approach to emphasize topological structure rather than trading intensity. However, in the later clustering aggregation stage, we introduce local weighting mechanisms to refine the final core–periphery structures by adjusting for structural uncertainties and improving aggregation consistency.
3.2. Core–Periphery Stochastic Block Models
Borgatti and Everett first introduced the core–periphery structure concept in 1999. Subsequent research has further developed and refined it [
1]. Core nodes are typically necessary network connectors, often possessing higher degrees of betweenness centrality. In contrast, peripheral nodes have fewer connections and lower status. This structural concept is crucial for understanding two key network dynamics: information dissemination and influence propagation.
The stochastic block model (SBM) is a probabilistic graph model with broad applications. It is primarily used in community detection, social sciences, and bioinformatics. The model enables node partitioning into distinct blocks, allowing researchers to infer network structures through inter-group connection probabilities. The SBM proves particularly effective for identifying hierarchical structures like the core–periphery arrangement. This effectiveness stems from its ability to capture differential connection patterns between core and peripheral nodes. Our study uses the SBM to extract core–periphery structures in the NFT transfer network. We infer the probabilistic connections between core and peripheral nodes through this model.
Our NFT network analysis modeled the transfer network as a static directed graph. Nodes represent individual traders, while directed edges represent NFT transactions. Although NFT transactions inherently occur over time, our focus remains on the cumulative network structure. This approach better reveals the long-term hierarchical relationships between core and peripheral nodes. The stochastic block model aligns well with this static representation. Its strength lies in identifying probabilistic connections based on node roles (core or periphery).
The stochastic block model initially selects n nodes and randomly assigns these n nodes to several sets. Taking the example of using the stochastic block model in core–periphery partitioning, the selected nodes are randomly allocated to the core node set, , and the periphery node set, . Each node has a probability, , of being assigned to . Correspondingly, the likelihood of being assigned to is . Next, the model generates edges between nodes. An undirected edge is placed between each pair of nodes with a probability, . Here, and represent the two nodes. Thus, the connection probability depends entirely on their set assignments. The probability matrix pst is a 2 × 2 matrix in core–periphery partitioning. The matrix contains four probabilities, labeled as , , , and . Because undirected edges are placed, , so only the probabilities , , and need to be considered.
In traditional stochastic block models used for community detection, higher within-block connectivity and lower between-block connectivity are considered. These conditions can be expressed as
and
. However, when applying the stochastic block model to represent core–periphery structures, we can obtain
. In core–periphery structures, nodes assigned to
are considered core nodes, while peripheral nodes are those in
, i.e.,
. Additionally, the characteristics of core–periphery structures determine that connections between peripheral nodes are less likely than connections between peripheral nodes and core nodes, i.e.,
. The different representations of community structures and core–periphery structures by the stochastic block model are illustrated in
Figure 3.
The SBM is a widely used statistical model for network structure analysis. In this study, we apply the SBM to the NFT transfer network to extract its core–periphery structure. This approach forms the foundation of our proposed core–periphery stochastic block model, which can be a general statistical model for analyzing network structures.
The core–periphery stochastic block model operates on a key assumption: core and peripheral nodes are assigned to distinct blocks. In the context of the NFT transfer network, we assume there are
nodes with an adjacency matrix,
. These
nodes are randomly allocated to
blocks. The block allocation status of nodes in the NFT transfer network is represented by vector
of length
, where
, and node
assigned to block
is denoted as
. The collection probability between any two nodes is defined by an M × M matrix R. Here,
represents the probability of a node in block
connecting to another node in block
. Through the modeling process, we observed that the core–periphery stochastic block model determines node connections based on their assigned blocks. Matrix
serves as the block connectivity matrix. While block allocation vector
and block connectivity matrix
are initially unknown in the NFT transfer network, their joint posterior distribution,
, is essential for core–periphery partitioning. Based on Bayesian methods, we relate
to two prior probabilities, which leads to the following Equation (1). Here,
denotes proportionality, and in specific statistical inference processes, this proportionality will be adjusted based on prior knowledge and evidence factors.
The core–periphery stochastic block model examines the core–periphery structure through node connections. These connections are reflected in block connectivity probability . In the NFT transfer network, we only consider the block connectivity probability under specific block allocations. We integrate the two-block model and k-core decomposition method with the stochastic block model. This integration yields two types of core–periphery stochastic block models: the hub-and-spoke model and the layered model. The hub-and-spoke model divides network nodes into core and peripheral node sets. Core nodes are connected, and they are connected to some peripheral nodes, but peripheral nodes are not connected. The layered model employs k-core decomposition to partition nodes into hierarchical shells, where each layer reflects distinct structural and functional roles. In our model, layers are defined from the innermost core (Layer 1) to the outermost periphery (Layer L), with the probability of connections decreasing as we move outward. The roles of different layers are as follows:
Layer 1 (core) comprises high-value traders, major NFT platforms, and influential participants who drive market trends. These nodes are densely connected and handle the majority of transactions.
Layer 2 (secondary core) serves as a bridge between the core and outer layers. This layer includes active traders who frequently interact with core participants but have fewer direct connections.
Intermediate layers () represent traders with moderate transaction activity. These nodes contribute to market liquidity but do not exhibit strong influence individually.
Layer l (Periphery) consists of occasional traders, newcomers, and dormant accounts. These nodes have minimal connectivity and limited impact on the network’s structure.
The connectivity probability matrix R follows
, ensuring a structural hierarchy where core–layer interactions dominate. This framework captures multi-level market dynamics more effectively than binary core–periphery models. Schematic diagrams of these two models are shown in
Figure 4. By applying Bayesian methods to the stochastic block model, we update the prior
. This embeds different node allocation and core–periphery partitioning methods into the model, enabling statistical inference and model fitting.
- A:
Hub-and-spoke model
The hub-and-spoke model fits the two-block model. In this model, the network is partitioned into fixed blocks: the core block and the peripheral block. The core block is encoded as
, while the peripheral block is encoded as
. From the definition of the two-block model, the definition of the hub-and-spoke model can be derived, where
and
(
). The definition establishes a clear core–periphery structure. The core is moderately or fully connected to the periphery. In contrast, connections between periphery nodes are minimal or nonexistent. Therefore, we constrain all prior probabilities of block connectivity matrix
of the hub-and-spoke model according to Equations (2) and (3).
- B:
Layered model
We assume the layered model consists of
layers. Here, layers correspond to the number of blocks used for node allocation in the stochastic block model. According to the k-core decomposition method definition, the probability of node connections gradually decreases from the innermost layer to the outermost layer. In the layered model, nodes in the innermost first layer are highly likely to be connected to nodes in other layers. However, the probability of nodes connecting to more peripheral layers decreases as we move outward. The node connection probability is further reflected in block connections. The degree of block connectivity decreases from the first layer to more peripheral layers. We describe the connectivity of block
as
. Based on this structure, we constrain all prior probabilities of block connectivity matrix
for the layered model. These constraints are defined in Equations (4) and (5).
3.3. Inference of Core–Periphery Stochastic Block Structure
We propose two core–periphery stochastic block models. These models require statistical inference of two key components: block assignment vector
and block connection matrix
. The inference process for the core–periphery stochastic block models is illustrated in
Figure 5. We implement Gibbs sampling to infer the distributions of
and
. Specific sampling steps are designed to collect samples based on the joint posterior distribution
. First, following the Gibbs sampling approach, we alternately sample
and
. Initially, we fix
and update
. Then, we fix the updated
and further update
. In the Gibbs sampling process, we assign the block
, to which
is most frequently assigned, as the statistical block assignment, denoted as
Through this method, we can obtain the general joint posterior distribution
in various network structures. To demonstrate the process, we take the Bayesian inference of the layered model as an example. The steps are detailed as follows:
First, in the sampling process of
, we focus on two blocks:
and
. Then, we define two quantities: the actual number of edges
between
and
, and the maximum possible number of edges
between them. Let
and
be the numbers of nodes in
and
, respectively. Then,
can be represented by Equation (6):
Next, we derive two expected values from
and
. The first is
, representing the expected number of edges originating from
to other blocks. The second is
, representing the expected number of non-edges. These values are computed through the formulas
and
. Based on these values, the posterior distribution
for the layered model can be represented by Equation (7):
We represent the connections of blocks in other layers as
. This notation allows us to express the distribution
as
. Establishing the connections between
and other blocks,
can be represented by Equation (8). Because
, the proportionality in Equation (8) allows the term
to be eliminated. This elimination ultimately leads to Equation (9).
According to our proposed layered model definition, given that
and
, we can ultimately derive Equation (10). Equation (10) indicates that the block connection
depends on other parameters in the layered model and satisfies
.
For a fixed block assignment
, new
(
) values can be sequentially obtained using Gibbs sampling. This process relies on the density variation in the distribution
. In the
-th sampling, the range of
needs to be controlled. This control ensures two objectives: smooth parameter updates in the layered model and increased likelihood that updated parameters reflect the actual network structure. Because the block connections
follow a beta distribution,
in the (
)-th sampling should be controlled such that
. This restriction ensures that
has a slight difference from the
of the previous sampling round. The sampling process incorporates an acceptance probability mechanism. If newly sampled parameters differ significantly from prior values, their acceptance probability decreases, leading to likely rejection. We determine whether the sampling is within the restricted range by calculating the peak value
of the distribution
in the
-th sampling, where
. If the sampling is within the restricted range, i.e.,
can be directly sampled from the beta distribution. If the sample
meets the constraints, it is accepted; otherwise, the value is rejected, and sampling continues. For samples outside the restricted range, i.e.,
or
, we use rejection sampling. In this case, the beta distribution can be expressed as a function,
, of the sample
, with the specific form provided in Equation (11). Here,
represents the gamma function.
We define a uniform distribution,
, over the range
. Based on this distribution, the value
is calculated as
. To ensure validity, the beta distribution function
must satisfy
. This condition determines the acceptance probability for sample
drawn from the uniform distribution
, as shown in Equation (12). Thus, moderate sampling of
can be performed within the beta distribution.
The next step is to sample from the distribution
). We use the Markov chain Monte Carlo (MCMC) method to obtain the distribution
based on
. First, we randomly assign blocks to the nodes in the network. After a certain number of iterations, we randomly select a node,
, and update its block label
. We use random sampling to select the new block label
for
, i.e.,
. By inverting
, a new block assignment
is obtained. According to the Metropolis–Hastings criterion, the probability of acceptance is calculated as shown in Equation (13).
Combining the two sampling processes described above, we propose the inference algorithm for the layered model (the hub-and-spoke model follows a similar procedure, with two key modifications: the constraint
and adjusted connectivity rules). Algorithm 1 describes the inference process for the layered model.
Algorithm 1 Layered Model Inference |
1: procedure |
2: //Initialize Block Assignments 3: , //Reorder under constraints 4: for do 5: 6: end for 7: 8: for do 9: //Initialize Gibbs Sample 10: for do //Sample 11: //Random Choose Node i 12: s //Random Choose Block s 13: 14: 15: //Update Assignments 16: 17: if then 18: //Accept new Assignments 19: else 20: //Revert Change 21: end if 22: end for 23: for do 24: //Sample 25: end for 26: end for 27: end procedure |
3.4. Comparison and Evaluation of Core–Periphery Structures
When assessing the performance of different core–periphery partitioning methods, we need a metric to measure the differences between different partitioning results. Our goal is to evaluate the performance of methods by comparing partitioning results with a “ground truth” or “authoritative” partitioning that aligns with shared understanding.
Core–periphery partitioning outcomes are sensitive to initialization settings and parameter choices. To account for this variability, the same method must be executed multiple times to generate diverse partitioning results. These results may form a collection of outcomes rather than a single partition. In such cases, the optimal subset can be selected or multiple subsets can be randomly sampled for clustering aggregation.
We can use the variation in information (VI) metric to compare the distances between different partitioning results. Rooted in information entropy theory, VI calculates the distance between two partitions by measuring information exchange, loss, and gain. This dual perspective makes VI particularly suitable for evaluating similarities and differences in core–periphery structural partitions.
The process of comparing the core–periphery structures is as follows. Firstly, select a node, , from the network. The probability of this node belonging to the core node set is defined as , where and are the sizes of and the total network, respectively. Furthermore, we define a discrete random variable of length , which selects nodes corresponding to the number of sets in partition . In core–periphery structures, there is only a core node set and a periphery node set, so takes a value of 2. The entropy of the discrete random variable is denoted as . It is essential to note that is the entropy of the partition , and it is a non-negative value. This entropy depends not on the absolute node count but on the relative proportions of node sets in .
We assume two different partitions,
and
, where the node set
in
corresponds to the node set
in
. The joint probability distribution of
belonging to
in partition
and
in partition
is denoted as
, where
represents the number of nodes assigned to both
and
in partitions
and
, respectively. We use mutual information (MI) to describe the information about partition
provided by partition
. When selecting
in the network, the uncertainty of
in partition
is denoted as
. If it is found that
is assigned to any node set in
, the uncertainty
should decrease accordingly. The decrease in uncertainty is distributed among the
nodes, partially explaining the principle behind MI. We use Equation (14) to express the MI between partitions
and
.
Based on the above content, we consider
and
as measures of uncertainties for node sets in two distinct partitions. The mutual information
represents the shared knowledge between these partitions, which effectively reduces uncertainty about node set assignments. To compute the variation in VI information between partitions, we first calculate the total uncertainties by summing
and
. Next, we subtract the mutual information
to eliminate the influence of shared knowledge. Equation (15) provides the calculation formula for the VI distance.
Figure 6 intuitively describes the VI distance for core–periphery structures. The shaded areas represent the uncertainty components that contribute to the VI distance. The middle blank area represents the mutual information (MI) shared between the two partitions.
After determining the VI distance as a comparison metric for core–periphery structures, we found significant differences in the core–periphery structures obtained with different model parameters. Therefore, we need to evaluate each core–periphery structure further. Based on the principle of MDL, we decompose the evaluation into two components: the number of bits
required to describe the core–periphery stochastic block model itself and the number of bits needed for the model to describe the network data. This is formalized as
. The length of the model describing the network data can be approximated as
. This form can be further obtained by integrating over the block connection
, as shown in Equation (16).
Direct computation of this integral is computationally challenging. To address this, we use Monte Carlo simulation to sample
values of
from its prior distribution
, following the sampling process described in
Section 4.2. These samples approximate the integral for the model description length. The sum of intervals among the
sampled points is 1, and the distribution of intervals is consistent and randomly combined. Therefore, the distribution of these intervals follows a Dirichlet distribution. The samples
can be described by the intervals
(
) of the samples, i.e.,
. We simplify the calculation by applying a logarithmic transformation to the model description length. This yields the expression in Equation (17), where
.
By approximating
and combining with the estimated model encoding lengths, we can obtain the
values for different core–periphery structures. For core–periphery structures
and
, the MDL ratio is calculated as
. This ratio facilitates direct comparison between implemented structures. Since different core–periphery structures are assumed to be equiprobable, the MDL ratio depends solely on the relative likelihoods of their distributions. We apply a logarithmic transformation to the MDL ratio to quantify statistical differences. This process is formalized in Equation (18), converting the ratio into an interpretable metric.
Our comparison and evaluation framework provides two key capabilities: First, it estimates the quality of different core–periphery structures and assesses their ability to characterize network features. These evaluations inform weight settings during locally weighted core–periphery clustering, offering data-driven guidance for structural optimization.
3.5. Local Weighted Aggregation of Core–Periphery Structures
Ensemble learning is an essential method in machine learning that combines multiple models into a more effective one. Classifier aggregation and clustering aggregation belong to supervised and unsupervised learning, respectively. When analyzing the core–periphery structure in NFT transfer networks, we are more concerned with the network’s structural features. Therefore, we consider using clustering aggregation methods to enhance the reliability of core–periphery partitioning results and the expressive power of network features.
Clustering aggregation enhances the robustness and accuracy of clustering results by considering the diversity of global clustering. Weighted clustering aggregation is an extension that considers the weights of different clusters in the aggregation process, thereby improving the aggregation effect.
Figure 7 shows a schematic diagram of weighted clustering aggregation.
Most networks exhibit complex structures in practical applications, challenging accurate and comprehensive analysis. Clustering aggregation methods have been introduced to graph structures to address these challenges. The core idea combines multiple node partitioning results to generate improved consolidated partitioning. However, the field of graph clustering currently lacks a universally accepted definition. This conceptual ambiguity has led to the proliferation of diverse algorithms with varying aggregation processes.
Taking the analysis of core–periphery structures in networks as an example, different core–periphery partitioning methods yield distinct core–periphery structures. These different core–periphery partitioning results essentially partition the nodes in the network into different node clusters. We represent graph as a binary data structure, . Here, denotes the set of nodes (), and represents the set of edges (). The partitioning results obtained through a specific core–periphery partitioning method consist of two node clusters, namely, and . Here, and . Therefore, this type of core–periphery partitioning result can be represented as . For real networks, aggregation results from multiple partitioning methods generate a composite structure . Building upon graph aggregation theory, this study integrates weighted clustering aggregation techniques. Specifically, we unify core–periphery partitioning results within cluster PP to enhance structural characterization.
We introduce the locally weighted core–periphery structure aggregation (LWCSA) method, which integrates bipartite graph models to balance global diversity and local reliability. This method aims to improve the accuracy and robustness of consensus core–periphery structures.
Figure 8 illustrates the process of LWCSA. The approach consists of three steps: the uncertainty estimation of node sets, the reliability testing of node sets, and core–periphery structure aggregation based on local weighted graph partitioning. Firstly, we estimate the uncertainty of each core–periphery partition using the concept of information entropy. Given a node set,
, and a core–periphery partition,
, where
and
represent all the node sets and core–periphery partition sets,
and
, respectively. The uncertainty of
with respect to
can be calculated by considering how nodes in
are aggregated in
. Firstly, we compute the distribution
of each node in
across node sets in
. Through this distribution, we can further obtain the uncertainty
of
concerning
. In our uncertainty computation, we introduce the MDL of core–periphery structures, resulting in the uncertainty
of
concerning the core–periphery partition set
as shown in Equation (19).
After determining the uncertainty of each node set in the core–periphery partitions, we calculate their reliability. The ensemble-driven clustering index (ECI) is the reliability metric for each node set. Given the core–periphery partition set and , where there are core–periphery partitions in , the ECI of the node set is calculated as follows: . Here, due to the drastic influence of instability on the growth of ECI, we introduce a parameter, , in the denominator of the exponent to balance the effect of instability.
We propose LWCSA based on bipartite graphs, considering the core–periphery structures’ global diversity and local reliability. A bipartite graph is defined by two disjoint sets with no intra-set connections. In core–periphery partitions, the core and periphery nodes are also distinct. Therefore, bipartite graph methods are suitable for core–periphery structure aggregation.
We represent the NFT transfer network nodes as nodes within a bipartite graph. In this bipartite graph, the core and periphery nodes are sets from two distinct node groups. If there is an edge between nodes in the graph, it indicates that the nodes in the NFT transfer network belong to different sets of nodes in the bipartite graph. By integrating the bipartite graph framework with ECI metrics and MDL-based core–periphery segmentation, our method achieves two objectives: (1) capturing node set affiliations in the network and (2) synthesizing local reliability metrics during structural aggregation. We define the bipartite graph
, where
,
represents all nodes and
represents the weight matrix of edges between two different node sets in
,
. For example, the definition of edge weights
and
belonging to different sets is shown in Equation (20).
Based on the above, we define the bipartite graph
. The next step is to partition
into disjoint sets of nodes. We employ the spectral partitioning algorithm (SPEC) proposed by Ng et al. [
18] to achieve the partitioning of
. SPEC embeds the nodes of
into a
k-dimensional space and then performs clustering in the
k-dimensional space, where
k represents the number of clusters in
. The specific process is as follows: SPEC first computes the degree matrix
of the nodes in
, where the elements of the matrix are denoted as
. SPEC then calculates the normalized weighted matrix
based on the degree matrix
and the weight matrix
and identifies the top k eigenvectors to form the feature matrix
. Finally, SPEC normalizes each row of
to unit length, resulting in k-dimensional embeddings for each node in
, which are then clustered using K-means. Nodes clustered in the same segment in the clustering result can be regarded as belonging to the same set, thus obtaining the partitioning result of
. The partitioning result of
corresponds to the final aggregated core–periphery structure. Algorithm 2 outlines the LWCSA scheme.
Algorithm 2 Locally Weighted Core-periphery Structure Aggregation |
Input: the set of core periphery partitions . |
1: Compute the uncertainty of the sets of nodes in P. 2: Compute the model description length of each structure. 3: Combine model description length to compute the ECI index of sets of nodes in P. 4: Build the bipartite graph based on Citation network. 5: Partition the graph into different part. 6: Group the nodes in the same part into one set and get all sets of Citation network. 7: Get the consensus core periphery structure through the obtained sets. Output: the consensus core periphery structure . |