Next Article in Journal
The Wheel–Rail Contact Force for a Heavy-Load Train Can Be Measured Using a Collaborative Calibration Algorithm
Previous Article in Journal
Toward Ensuring Data Quality in Multi-Site Cancer Imaging Repositories
Previous Article in Special Issue
Construction of Knowledge Graphs: Current State and Challenges
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Knowledge-Aware Recommendation with Dual-Graph Contrastive Learning

1
China Unicom Research Institute, Beijing 100176, China
2
Shandong Artificial Intelligence Institute, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250014, China
3
Faculty of Computer & Information Sciences, Hosei University, Tokyo 184-8584, Japan
*
Author to whom correspondence should be addressed.
Information 2024, 15(9), 534; https://doi.org/10.3390/info15090534
Submission received: 30 June 2024 / Revised: 26 August 2024 / Accepted: 1 September 2024 / Published: 2 September 2024
(This article belongs to the Special Issue Knowledge Graph Technology and its Applications II)

Abstract

:
Incorporating knowledge graphs as auxiliary information to enhance recommendation systems can improve the representations learning of users and items. Recommendation methods based on knowledge graphs can introduce user–item interaction learning into the item graph, focusing only on learning the node vector representations within a single graph; alternatively, they can treat user–item interactions and item graphs as two separate graphs and learn from each graph individually. Learning from two graphs has natural advantages in exploring original information and interaction information, but faces two main challenges: (1) in complex graph connection scenarios, how to adequately mine the self-information of each graph, and (2) how to merge interaction information from the two graphs while ensuring that user–item interaction information predominates. Existing methods do not thoroughly explore the simultaneous mining of self-information from both graphs and effective interaction information, leading to the loss of valuable insights. Considering the success of contrastive learning in mining self-information and auxiliary information, this paper proposes a dual-graph contrastive learning recommendation method based on knowledge graphs (KGDC) to explore a more accurate representations of users and items in recommendation systems based on external knowledge graphs. In the learning process within the self-graph, KGDC strengthens and represents the information of different connecting edges in both graphs, and extracts the existing information more fully. In interactive information learning, KGDC reinforces the interaction relationship between users and items in the external knowledge graph, realizing the leading role of the main task. We conducted a series of experiments on three standard datasets, and the results show that the proposed method can achieve better results.

1. Introduction

In the era of information explosion, users are confronted with an overwhelming number of choices. Recommendation systems alleviate this issue by providing personalized recommendations [1,2,3], helping users quickly find content or products of interest, thus reducing information overload and enhancing user satisfaction and experience. For e-commerce and content platforms, recommendation systems can efficiently guide users to content they are more likely to purchase or consume, significantly boosting conversion rates and sales, thereby directly increasing economic benefits for businesses. Collaborative filtering algorithms, widely employed in recommendation systems, leverage users’ historical behavior data to provide personalized recommendations for each user, improving the relevance of recommendations and user satisfaction [4,5,6,7]. However, real-world data often present challenges such as cold start and sparse data, which greatly limit the performance of collaborative filtering-based recommendation algorithms [8,9,10]. To address the issue of sparse data, an effective approach is to integrate auxiliary information into the collaborative filtering recommendation model, and knowledge graph-based recommendation is a typical method in this regard [11,12,13,14,15].
In knowledge graph-based recommendations, in addition to the interaction information between users and items, there is also knowledge graph information about the items. Knowledge graphs contain rich entity relationship information about items, enabling the construction of comprehensive item features and helping to uncover hidden relationships between items, thereby improving the accuracy of recommendations. To make it easier to understand, we use collaborative filtering (CF) to represent the interaction graph between users and items, and use knowledge graph (KG) to represent the connectivity graph between items. CF is a heterogeneous graph composed of users and items, and KG is a graph composed of huge items and multiple types of connection relationships. Cf and KG are two important graphs in KG-based recommendation, and contain rich information.
There already exists much research on sufficiently utilizing information in CF and KG. Earlier studies [11,16,17] focused on independently learning from the two graphs, which mainly represent the triplet information from the item knowledge graph as embedding and its use as contextual information for enhancing recommendations in the CF graph. These methods typically employ knowledge graph embedding (KGE) models (such as TransE [18], TransH [19]) to learn representations of entities in the KG. However, these approaches have limitations in extracting meaningful information from entities, and they can only extract information from a single graph structure, failing to integrate information from the CF graph. Therefore, subsequent works [20,21,22] have increasingly focused on how to extract more relevant information for recommendations from the KG. One productive approach is to represent the interactions with multi-hop paths from users to items, which is hard to optimize because of the manually designed meta-paths. Recently, graph neural networks (GNNs) have demonstrated strong capabilities in representing structural knowledge in graphs [12,13,23,24]. They are widely used in recommendation methods based on knowledge graphs and have achieved excellent performance.
Although knowledge graph-based recommendation has achieved promising results, it still faces the following issues:
  • Insufficient mining of the two graphs’ own information: Existing methods often use the interaction data between users and items as supervision signals to derive user and item representation vectors from the entities in the KG for learning and training. However, these methods do not fully exploit the information inherent in the two graphs, especially the strong and effective features of user and item IDs in the recommendation domain. This oversight can lead to the loss of valuable information, adversely affecting the recommendation performance.
  • Unbalanced information between the two graphs: Unlike the sparse behavioral data between users and items, the connections in knowledge graphs are dense, containing a wealth of information. The difference in the amount of knowledge contained in the two graphs can cause issues in the subsequent utilization of the information. The supervision signals in CF are directly related to the predictions, whereas the abundant redundant information in the KG can weaken these CF supervision signals. If the dominance of CF information is not maintained, it can lead to a decline in recommendation accuracy.
Inspired by the success of contrastive learning (CL) methods in sparse data scenarios, this paper proposes a Dual-graph Contrastive learning recommendation model based on the Knowledge Graph (KGDC) to address the aforementioned issues. To fully exploit the effective information from each of the two graphs, KGDC utilizes information propagation and aggregation techniques from GNNs to learn the vector representations of IDs in the CF graph and the entities and relations in the KG, respectively. To integrate information from both graphs, the proposed method leverages the concept of contrastive learning to fuse information from two aspects. The first method treats the items interacted with by the same user in the CF graph as pseudo-positive item sets, using them as positive supervision signals in the KG while using other non-similar, non-connected items as negative samples to further learn and train the representation vectors of entities in the KG. The second method considers the corresponding items and entities in the CF graph and KG as similar samples, with other non-corresponding samples as negative samples, to further enhance the information fusion and transfer between the two graphs. Finally, we employ a multi-objective training mode, where the vector representations of different components are used to calculate the loss according to different objectives, and the losses are summed with different weights to optimize and train the model parameters. We conducted extensive experiments on two public datasets, and the results show that our proposed method outperforms some state-of-the-art methods.
The contributions of this paper are as follows:
  • We propose the dual-graph conception to fully exploit the information within each graph while effectively integrating information between the graphs. It firstly enhances the deep exploration of each graph’s inherent information, particularly strengthening the representation learning of user and item IDs in CF graph. Then, during the information fusion process, it ensures the strong dominance of target consistency information, preventing interference from redundant information in the KG graph.
  • We introduce the contrastive learning to both individual and integrated learning stages. On the one hand, sample construction in contrastive learning prevents overfitting caused by excessive emphasis on the target. On the other hand, it incorporates the loss calculation based on the comparative learning sequence relationship to improve the effectiveness of ranking orders.
  • We conduct extensive experiments on public datasets, further validating the superior performance of the proposed method.
The rest of this paper is organized as follows. Section 2 reviews the related work on graph-based recommendation and contrastive learning. Section 3 defines the problem formulation and presents a detailed introduction of KGDC. Section 4 presents experimental results, and subsequent discussions are provided in Section 5. Section 6 concludes the paper and looks forward to future work.

2. Related Work

The study in this paper is mainly relevant to two parts of works: knowledge graph-based recommendation and contrastive learning.

2.1. Knowledge Graph-Based Recommendation

Research into knowledge graph-based recommendation systems has explored various methodologies, focusing on leveraging the structure and semantics of knowledge graphs to enhance recommendation performance. This part reviews notable contributions from two aspects: non-GNN-based methods and GNN-based methods.

2.1.1. Non-GNN-Based Methods

Early work on integrating knowledge graphs with recommendation systems primarily focused on leveraging the rich semantic relationships inherent in knowledge graphs to improve recommendation accuracy [11,17,25]. Zhang [17] introduced the Collaborative Knowledge Base Embedding (CKE) framework, which combines collaborative filtering techniques with knowledge graph embeddings to enhance recommendation performance by incorporating additional contextual information. Similarly, KTUP [25] considers both explicit user–item interactions and implicit knowledge graph relationships to improve the recommendation process with the TransH [19] method. Wang [11] proposed the RippleNet model, which propagates user preferences over a knowledge graph to uncover multi-hop relational paths, thereby enriching the user’s preference representation.
Another significant contribution in this area is the use of path-based methods. These approaches often employ random walk techniques or path ranking algorithms to identify relevant connections between entities. PER [26] extracts meta-path or meta-graph latent features from the knowledge graph to exploit the connectivity between users and items along different types of relation paths or graphs. KPRN [22] goes beyond predefined meta-paths by dynamically extracting paths that are relevant to user–item interactions, thereby improving the ability to capture complex relationships.
Embedding-based methods and path-based methods are both useful and effective. However, embedding-based methods may not fully capture the nuances of user preferences in recommendation scenarios where user–item interactions are sparse or when personalized recommendation accuracy is critical. Path-based methods rely heavily on the quality of path extraction and the ability to capture relevant semantic relationships.

2.1.2. GNN-Based Methods

GNN-based methods in knowledge graph-based recommendation systems have evolved significantly, addressing challenges related to sparse interactions, unbalanced graph utilization, and knowledge extraction. KGCN [12] focuses on leveraging user preferences to aggregate neighborhood information of items in the KG. Similarly, KGAT [13] uses GNNs to recursively aggregate information across the unified heterogeneous graph, emphasizing the importance of both collaborative filtering signals and knowledge signals. CKAN [15] highlights the significance of collaborative filtering signals in addition to knowledge signals. In KGIN [24], GNNs are applied on the user-intent–item-entity graph to capture nuanced user preferences. CG-KGR [27] uses pre-trained collaborative signals to guide the aggregation process on the knowledge graph, focusing on extracting sufficient knowledge for recommendations. KGIC [28] employs GNNs to capture hierarchical and structural dependencies within the KG.
Most GNN-based methods operate under supervised learning paradigms, which can be challenging when dealing with sparse user–item interactions. This limitation affects recommendation accuracy, particularly in scenarios with limited data. Moreover, the connections in knowledge graphs are dense, containing a wealth of information. These methods often overly emphasize the knowledge graph, potentially neglecting the crucial collaborative filtering signals inherent in user–item interactions.

2.2. Contrastive Learning

Contrastive learning, as an effective unsupervised learning method, has been widely studied and applied in recent years across domains such as image and natural language processing. Its core idea is to compare positive examples (similar samples) and negative examples (dissimilar samples), aiming to bring similar samples closer together in the representation space while pushing dissimilar samples further apart. This approach helps alleviate issues related to insufficient learning caused by sparse data. SimCLR [29] utilizes data augmentation techniques to generate multiple views of the same instance and employs a contrastive loss function to maximize the similarity between different views of the same instance while minimizing the similarity between different instances.
In recent research, several approaches have applied contrastive learning to enhance knowledge graph (KG)-based recommendation systems. SGL introduces three operators to generate augmented views and employs a multi-task strategy, jointly optimizing contrastive loss and recommendation loss to improve the robustness of representation learning in KG-based recommendation systems. SEPT [30] proposes a socially aware contrastive learning framework, and mines additional social information of users to enrich the learning process, aiming to better capture user preferences and interactions in recommendation tasks. CKER [31] integrates a contrastive learning module to enhance the sharing of user preferences, and derives additional supervision signals to improve the collaborative filtering aspect of KG-based recommendation systems. KGCL [32] introduces a knowledge-guided contrastive learning paradigm to derive more robust node representations, enhancing the quality of recommendations. KGIC [28] applies contrastive learning to learn embeddings or representations that distinguish relevant (positive) interactions from irrelevant (negative) ones based on the KG. KACL [33] utilizes contrastive learning to discern meaningful patterns and relationships within the knowledge graph.
Existing recommendation methods based on knowledge and contrastive learning mostly directly learn the integrated information of the two graphs, which can result in insufficient exploration of the inherent information in each graph. Moreover, when there is an imbalance of information between the CF and KG graphs, it is easy to introduce too much redundant information from the KG, which can negatively impact the prediction results.
Our proposed KGDC utilizes the dual-graph concept, simultaneously addressing the exploration of inherent information in both graphs and the integration of information between them. During the information fusion process, we ensure that the supervised signals in CF, which is more relevant to the target, maintain a dominant position.

3. Methodology

In this section, we present our proposed KGDC recommendation method. The purpose of the proposed KGDC is to fully exploit the individual information and interactive information from CF and KG. Figure 1 presents the framework of KGDC, which consists of three parts: Individual Graph Constructing and Encoding, Interactive Graph Constructing and Encoding, and Multi-task Learning. It performs contrastive learning in both individual graph and interactive graph learning to extract more effective information. Then we introduce the details of the three components.

3.1. Problem Formulation

To facilitate understanding, we describe the problem formulation here. We first introduce the structural data: the collaborative filtering graph of user–item interactions and the knowledge graph, and then formulate the problem statement of KG-based recommendation.
Collaborative Filtering Graph: To maintain consistency with the settings in GNN-based recommendation systems, we construct a bipartite graph G c f = { ( u , y u v , v ) } based on the historical interactions between users and items, such as the consumption, viewing, and clicking. In the construction of this graph, there are connections only between item and user nodes, with edges existing between nodes of users and items that have had historical interactions, where u U and v V denote the user and item involved in the interaction, and U and V are the sets of users and items, respectively. y u v is an indicator as to whether there is a connection between user u and item v based on the historical interactions. y u v = 1 indicates an interaction, while y u v = 0 indicates no interaction.
Knowledge Graph: In the knowledge graph, the connections are denoted as a set of triplets G k g = { ( h , r , t ) } , and each triplet describes a connection from head entity h I to tail entity t I with the relationship r R . There, it usually denotes I and R as the sets of entities and relations in the knowledge graph, and I includes items V and non-item entities I V .
KG-based Recommendation: Based on the user–item interaction graph G c f and item knowledge graph G k g , the target of KG-based recommendation is to learn a function y ( u , v ) that predicts the probability how likely a user u would interact with an item v.

3.2. Individual Graph Constructing and Encoding

Different from most existing methods, KGDC learns node and entity embedding from KG and CF, respectively, with GNN methods. In KG, KGDC uses contrastive learning to overcome the problem of missing explicit labels. In CF, KGDC uses propagation and aggregation methods for learning ID embedding of users and items.

3.2.1. Graph Constructing and Encoding in KG

During the individual graph constructing and encoding in KG, only triplets G k g = { ( h , r , t ) } are valid here. Considering that there are no explicit signals, KGDC utilizes contrastive learning to construct samples for information aggregation, loss calculation, and parameter optimization.
In terms of sample construction, the training entities, which are denoted as E, at different stages are also different. For example, the goal of the pre-training stage is to adequately train all nodes in the graph, and all nodes will participate in the training, E = I . In the subsequent model fine-tuning process, the entities to be trained are those involved in user–item interactions E I , with the aim of performing better for the prediction target. For any entity e i E , the positive samples of e i are sampled from the connected tails and relations from G k g , denoted as ( r i , p , t i , p ) G k g . Then the negative samples are sampled from all entities that are not connected in G k g , denoted as ( r i , n , t i , n ) G k g . Combining the positive and negative samples, we can obtain the samples for KG, denoted as S k g = { ( r i , p , t i , p ) , ( r i , n , t i , n ) , } , e i E .
With the constructed samples, KGDC uses the attentive mechanism in GNNs to propagate and aggregate information in KG. Moreover, for any entity e i E , we first compute the representations of positive samples by linear summation,
h k g , i + = j ( r i , p , t i , p ) α i j h k g , j ,
where h k g , j is the representation of entity j in KG, and α i j denotes the attention score between entity e i and entity e j . α i j is calculated as follows:
α i j = exp ( h k g , i · r j ) k ( r i , p , t i , p ) exp ( h k g , k · r k ) ) .
Then, the positive representation of entity i is calculated as follows:
h k g , i + = j ( r i , p , t i , p ) exp ( h k g , i · r j ) k ( r i , p , t i , p ) exp ( h k g , k · r k ) ) h k g , j .
Similarly, the negative representation of entity i is calculated as follows:
h k g , i = j ( r i , n , t i , n ) exp ( h k g , i · r j ) k ( r i , n , t i , n ) exp ( h k g , k · r k ) ) h k g , j .
With these representations of entity e i , we calculate the similarity score between positive samples and the target sample through the inner product, p i + = h k g , i · h k g , i + . Similarly, the similarity score between negative samples and the target sample is denoted as p i = h k g , i · h k g , i . Then, we adopt a pair-wise loss BPR to encourage the positive samples to be more similar with the target sample than the negative samples:
L k g = e i E log σ ( p i + p i ) ,
where σ is the Sigmoid function.

3.2.2. Graph Constructing and Encoding in CF

In the individual graph constructing and encoding in CF, KGDC learns the embedding of users and items by user–item interactions. This does not differentiate between users and items, and KGDC performs the same neighbor sampling and embedding aggregation on call nodes in CF to achieve information propagation.
During representation learning, the main idea is to update the vector representation of each vertex based on its own representation and that of its neighbors using some aggregation function. This process is repeated for all vertices for T iterations to obtain the representation of each node. Before representation learning, the neighbors of each vertex in CF are stored globally, denoted as N ( s i ) , where the nodes s i can be either a user node or an item node. Assuming it is the t-th iteration, KGDC directly retrieves its neighbor set from global storage, denoted as N ( s i ) . Then, it randomly selects a fixed number K of neighbors from N ( s i ) , denoted as N ( s i ) t . If the number of neighbors is less than K, sampling with replacement is used to reach the desired number; if it is greater than K, sampling without replacement is used.
With the sampled neighbors of the node s i in the t-th iteration, we can obtain the embedding set of these nodes using the equation as follows:
h c f , N ( s i ) ( t ) = AGGREGATE t ( h c f , j ( t ) , j N ( s i ) t ) ,
where AGGREGATE t is the aggregation method at the t-th iteration, such as mean pooling, sum pooling, max pooling, RNN, transformer, and other methods. Obtaining the presentation of neighbors, we perform a linear transformation and enhance the representation of the original node s i by using a custom activation function. The vector representation of vertex s i at the t + 1 -th iteration is calculated as:
h c f , s i ( t + 1 ) = σ W t CONCAT ( h c f , s i ( t ) , h c f , N ( s i ) ( t ) ) + b t ,
where W t R 2 d d and b t R d are the trainable weights and the trainable bis at the ( t -th iteration.
After T iterations, the node representation h c f , s i ( T ) is L2-normalized to obtain the final representation vector:
h c f , s i = h c f , s i ( T ) h c f , s i ( T ) 2 .
With the nodes’ embedding in CF, we compute the individual loss of CF as follows:
L c f = ( s i , y i , j , s j ) G c f ( y i , j l o g ( y i , j ^ ) + ( 1 y i , j ) ( 1 l o g ( y i , j ^ ) ) ) ) ,
where y i , j ^ = Sigmoid ( h c f , s i T h c f , s j ) is the predicted score of node s i and node s j , y i , j is namely the edges in CF, and the value is the same as y u , v .

3.3. Interactive Graph Constructing and Encoding

KGDC also leverages contrastive learning to learn the interactive information from two graphs. To ensure the dominant position of the CF graph, KGDC takes the signals in CF to guide the learning of entities in the KG. To further enhance the information fusion between the two graphs, KGDC uses the entities corresponding to the items in both graphs as similar samples for alignment.

3.3.1. Graph Encoding in KG with Signals of CF

In this section, KGDC extracts the signals of user/item from CF to guide the training of nodes in KG. The supervised signals extracted from CF are first-order connections between users and items. For any user u in CF, all of the items that have interacted with u are collected. Therefore, the signals from CF are denoted as:
V c f ( u ) = { v ( u , v ) G c f , and v { v y u v = 1 } } .
With the signals above, the initial entities learned in the KG are those that correspond to the items in the node set V c f ( u ) , which are denoted as:
V k g , 0 + ( u ) = { e ( u , e ) G c f } .
In the collaborative filtering recommendation, each user has his own interests and preferences, and there are commonalities in preferences among different users, which means that the items that interact with the same user tend to be more similar. Therefore, for any sample in CF ( u i , v j ) and y u i , v j = 1 , the items in V k g , 0 ( u i ) are more similar with v j than others. Furthermore, KGDC takes the items randomly selected from non-connected items across the entire dataset as negatives, denoted as V k g , 0 ( v j ) . Taking the positive samples V k g , 0 ( u i ) and negative samples V k g , 0 ( v j ) as the initial entity sets of the KG, KGDC constructs the embedding through a natural propagation in the KG. The propagation is the same as in Section 3.2.2, and aggregation is similar to that in Section 3.2.1, which randomly selects a fixed number of connected (relation, tail) as neighbors and aggregates the information using the attentive method as in Equations (1) and (2). At the l-th layer, the embedding of any entity e (including the positive samples and negative samples) defined above is as follows:
h k g , e ( l + 1 ) = CONCAT h k g , e l AGGREGATE l j ( r e , t e ) α i j h k g , j ( l ) ,
where h k g , * is the embedding in KG, and the definition of AGGREGATE l is the same as in Equation (6).
Similarly, after L iterations and L2-normalization as in Equation (8), we can obtain the entity presentation, denoted as h k g , e . Then, we utilize the triplet loss function, which is usually used in contrastive learning to compute the loss. The distance score between two entities is calculated through the inner product. Therefore, the loss function is as follows:
L k i c = ( u , v ) G c f e i V k g , 0 + ( u ) e j V k g , 0 ( v ) Max ( h k g , v T h k g , e j h k g , v T h k g , e i + m , 0.0 )
where m is the margin.

3.3.2. Align Encoding with CF and KG

To further enhance the interaction between the two graphs, we propose to map the item in CF and the corresponding entity in KG into the same space. For better understanding, we use h c f ( v ) to represent the item embedding in CF, and use h k g ( e ) to represent the entity embedding in KG, where e and v correspond to the same item. Therefore, for any item v in CF, the positive sample is the corresponding item e in KG, and the negative sample is the randomly selected item j from batch-wise in the training process. With the batch-wise negative sampling strategy, we do not need to sample globally, addressing computational power constraints and reducing time consumption.
Before the alignment between the two graphs, KGDC utilizes two MLPs to map them into the same space ( z c f ( v ) , z k g ( e ) ) to achieve the information shared across interaction and knowledge views. Thus, the positive pair is denoted as { z c f ( v ) , z k g ( e ) } , and the negative pairs are denoted as { z c f ( v ) , z k g ( j ) } and { z c f ( j ) , z k g ( e ) } . With these samples, we adopt the InfoNCE function to calculate the loss, which is commonly used in contrastive learning. InfoNCE loss aims to encourage the consistency between the representations of its different views while enforcing the divergence of negative pairs. The calculation formula is as follows:
L ali ( v ) = log exp ( c o s ( z c f ( v ) , z k g ( e ) ) j V { v } exp ( c o s ( z c f ( v ) , z k g ( j ) ) + exp ( c o s ( z c f ( j ) , z k g ( e ) ) )
where c o s ( · ) measures the cosine similarity of two vectors, and V (namely the items in CF) is the set of negative samples.

3.4. Multi-Task Learning

Considering that some of the losses obtained above are not directly related to the target, we take a portion of the representations to reconstruct embedding for users and items to ensure the update is more directly to the target. For any data ( u , v ) in the interaction dataset, the representations of individual CF in Section 3.2.2 are h c f , u and h c f , v , and the representations of interactive learning in Section 3.3.1 are { h k g , e } e V k g , 0 + ( u ) and h k g , v . Specifically, for the set of embedding { h k g , e e V k g , 0 + ( u ) } , we utilize the mean pooling method to aggregate them to obtain the user’s presentation, denoted as h ¯ k g , V k g , 0 + ( u ) . By concatenating these vectors, we can obtain the final presentations of the users and items, which are also the model’s predictions. The formulations of reconstruction and the matching score are as follows:
h u = CONCAT ( h c f , u , h ¯ k g , V k g , 0 + ( u ) ) ,
h v = CONCAT ( h c f , v , h k g , v ) ,
Then, we obtain the predicted score using the inner product and Sigmoid function, denoted as y ^ ( u , v ) = Sigmoid ( h u T h v ) . With y ^ ( u , v ) and the real labels y ( u , v ) , we utilize the cross-entropy function to compute the loss as Equation (9), denoted as L t a r g e t .
To combine the individual learning as well as the interactive learning proposed above with the recommendation task, we proposed a multi-task learning framework to jointly train the multiple losses as in Algorithm 1. In the proposed framework, the individual learning of KG is conducted in the pre-train process, which enables all of the entities and relations in KG to be fully trained compared with learning it in the training process. Moreover, the equation of L t o t a l consists of four parts, where L t a r g e t and L c f are directly related to the final objective and the other two are indirectly related. This ensures that the user–item interaction signal maintains a dominant position.
Algorithm 1 KGDC Algorithm
1:
Input: user–item interaction graph G c f ; knowledge graph G k g ; trainable parameters: { h c f , u , h c f , v } u U , v V , { e } e E , { r } r R , { W i , b i } i = 1 L ; hyper-parameters and functions: L , d i m , σ ( · ) , CONCAT ( · ) , AGGREGATE ( · )
2:
Output: Prediction vectors
3:
for  h = 1 , , e p o c h ( pre train ) do
4:
    for  e E  do
5:
        Construct samples and presentations as Equations (1)–(4)
6:
        Calculate and obtain loss L k g as in Equation (5)
7:
        Update parameters by gradient descent
8:
    end for
9:
end for
10:
for  h = 1 , , e p o c h ( train ) do
11:
    for  ( u , v ) G c f  do
12:
        Individual learning as Equations (3), (4) and (6)–(8)
13:
        Calculate loss L k g as Equation (5), L c f as Equation (9)
14:
        Interactive learning as Equation (12) and Section 3.3.2
15:
        Calculate loss L k i c as Equation (13), L a l i as Equation (14)
16:
        Reconstruct h u , h v as Equation (15), Equation (16), and calculate target loss L t a r g e t as Equation (9)
17:
        Calculate L t o t a l = L t a r g e t + α L c f + β L k g _ + γ L C L
18:
        Update parameters by gradient descent
19:
    end for
20:
end for
21:
return  h u , h v

4. Experiments

In this section, we conduct extensive experiments on two real-world datasets to evaluate the proposed KGDC. Next, we will sequentially provide detailed introductions to the experimental datasets, comparison baselines, experiment settings, and experimental results.

4.1. Datasets

We evaluate KGDC on two distinct scenarios: books and music. These datasets are publicly available and vary significantly in size and sparsity, and they are often used in the field of knowledge graph-based recommendation. Below are the details of each dataset:
  • Amazon-book (http://jmcauley.ucsd.edu/data/amazon/, accessed on 1 April 2024): This dataset includes user ratings and extensive metadata for books, such as descriptions, category information, price, and brand, sourced from Amazon.
  • Last.FM (https://grouplens.org/datasets/hetrec-2011/, accessed on 1 April 2024): Collected from the Last.FM online platform, this music dataset consists of user-generated music listening events.
The basic statistics of the two datasets are presented in Table 1.
We adopt the method in [11] to convert explicit feedback across three datasets into implicit feedback, where a label of 1 denotes positive samples. In the Book-Crossing dataset, the ratings greater than or equal to 4 are labeled as 1, while the threshold is 0 in the Last.FM dataset. For every dataset, the ratio of training, evaluation, and test set is 6:2:2. Each experiment is repeated 3 times, and the average performance is reported. For constructing sub-KGs, we utilize Microsoft Satori4, which is similar to RippleNet [11] and KGCN [12]. Each sub-KG adheres to the triple format and constitutes a subset of the entire KG with confidence levels exceeding 0.9. From the sub-KG, we extract Satori IDs of all valid movies, books, or musicians by matching their names with the tail of triples. Subsequently, we match item IDs with the head of all triples and select well-matched triples from the sub-KG.

4.2. Baselines

To demonstrate the effectiveness of the proposed KGDC, we compare it with the recent state-of-the-art methods, including BPRMF, CKE, KGCN, KGNN-LS, KGAT, CKAN, KGIN, and KGIC. The descriptions of these methods are as follows:
  • BPRMF [34]: It aims to rank user–item interactions in a way that places higher preference scores on items that users have interacted with, which is a traditional CF-based method;
  • CKE [17]: It leverages structured information from knowledge graphs to enhance the quality of recommendations by embedding both user–item interactions and the rich semantic relationships present in the knowledge graph;
  • KGCN [12]: It captures both the structural and semantic information from the knowledge graph, enabling it to learn more comprehensive and rich representations of items;
  • KGNN-LS [14]: It integrates knowledge graph information with neural networks, incorporating the technique of label smoothing to improve model performance;
  • KGAT [13]: It integrates knowledge graph information using attention mechanisms;
  • CKAN [15]: It employs attention mechanisms to dynamically focus on the most relevant entities and relationships within the knowledge graph, enhancing the collaborative filtering process;
  • KGIN [24]: It applies a GNN to the user-intent–item-entity graph, allowing for a more granular and nuanced understanding of these interactions;
  • KGIC [28]: It enhances traditional collaborative filtering by leveraging rich semantic relationships from knowledge graphs with the contrastive learning method;
  • CG-KGR [27]: It employs a collaborative guidance mechanism, which encodes historical interactions as guidance for personalized knowledge extraction, making it particularly effective for tasks like Top-K recommendation.

4.3. Experiment Settings

The settings of hyper-parameters for baselines and the proposed KGDC are as follows. To ensure the fairness of the experimental results, some parameters are set to fixed values. The dimensions of nodes in CF and nodes (and relations) in KG are set as 64, and the batch size is set as 1024. Model parameters are initialized using Xavier, and optimization is performed using Adam.
For some parameters, we search for the optimal values within a certain range. We explore learning rates in the range {0.0001, 0.0005, 0.001, 0.0015}, and vary the number of neighbors between 2 and 50. The depth exploration of GNNs ranges from 1 to 4. Baselines retain their default hyper-parameter settings except for the aforementioned parameters. To mitigate overfitting, we ensure training convergence by limiting the number of steps to 50.
To evaluate the effectiveness of our proposed KGDC in the knowledge graph-based recommendations, we employ four widely used metrics: AUC, F1, Recall@K, and NDCG@K, where the K values of Recall@K and NDCG@K are set to [5, 10, 20, 50, 100]. AUC and F1 are two widely used metrics in click-through rate (CTR) prediction, while Recall@K and NDCG@K are two methods usually used in the Top-K recommendation.
The AUC can be calculated using the following formula:
AUC = i = 1 P rank i P ( P + 1 ) 2 P × N ,
where rank i is the rank of the i-th positive sample, P is the number of positive samples, and N is the number of negative samples.
The F1 score combines precision and recall and is calculated as:
F 1 = 2 × Precision × Recall Precision + Recall ,
where Precision = TP TP + FP , and Recall = TP TP + FN . The definitions of “TP, FP, FN” are as follows:
  • TP (True Positives) is the number of correctly predicted positive outcomes;
  • FP (False Positives) is the number of incorrectly predicted positive outcomes;
  • FN (False Negatives) is the number of incorrectly predicted negative outcomes.
The formula of Recall@K is as follows:
Recall @ K = Number of relevant items in Top K recommendations Total number of relevant items
NDCG@K is calculated as follows:
NDCG @ K = DCG @ K IDCG @ K ,
DCG @ K = i = 1 min ( K , n ) 2 r e l i 1 log 2 ( i + 1 ) ,
IDCG @ K = i = 1 min ( K , n ) 2 max ( r e l i ) 1 log 2 ( i + 1 ) ,
where n is the number of items in the recommendation list, and r e l i is the relevance score of the i-th item in the recommendation list.

4.4. Results

In this section, we present a comprehensive performance analysis between our proposed KGDC and all representative baselines in the tasks of CTR prediction and Top-K recommendation.

4.4.1. Results of CTR Prediction

We present the experimental results of the CTR prediction task across all datasets in Table 2. From these results, we draw the following observations and analyses.

4.4.2. Results of the Top-K Recommendation

Similarly, we conduct experiments of Top-K recommendations across all datasets, and vary K in the range of { 5, 10, 20, 50, 100}. For a better comparison between KGDC and these baselines, we firstly show the average results of the Top-K, K = 5, 10, 20, 50, 100, in Table 3, Table 4, Table 5, Table 6 and Table 7. Then, we make use of four subfigures to further illustrate the complete results for the Recall@K and NDCG@K metrics, shown in Figure 2.

5. Discussion

We next conduct a detailed analysis and discussion of the results in Section 4.

5.1. Discussion on Results of the CTR Prediction

For the CTR prediction task, our KGDC demonstrates significant improvements in performance. Particularly on the Book-Crossing dataset, KGDC improves by 2.45% and 1.15% in AUC and F1 compared to the best result of the baselines, respectively, while on the Last.FM dataset, KGDC only improves AUC by 0.7% and improves F1 by 0.23%.
This is likely due to the relatively limited impact of knowledge supplementation on the Music dataset. We use the ratio of the number of KG triplets to the number of items to measure the average knowledge contribution to enrich item embedding. A higher ratio typically indicates richer semantics in the KG, enhancing item backgrounds. For the Music dataset, this ratio is 4.03, while the ratio of the Book dataset is 10.12. Thus KGDC can perform significantly better on the Book-Crossing dataset by effectively utilizing the rich semantics in these KGs for accurate predictions.

5.2. Discussion on Results of the Top-K Recommendation

For the Top-K recommendation task, KGDC achieves better results compared with the CTR task. Taking K = 20 as an example, KGDC improves Recall@20 and NDCG@20 by 8.418% and 6.762% compared to the state-of-the-art method on the Book-Crossing dataset; on the Last.FM dataset, it improves Recall@20 and NDCG@20 by 5.856% and 6.525%, respectively. We believe that the significant improvements are attributed to the extensive integration of contrastive learning techniques and pair-wise loss formulations in the algorithm. The introduction of contrastive learning and pair-wise loss have significantly improved the effectiveness of Top-K recommendation. Compared to traditional methods, introducing contrastive learning and pair-wise loss can achieve better results. It also demonstrates that contrastive learning can help the model to fully and effectively mine information when the supervision signals are insufficient. Meanwhile, the pair-wise loss function enhances the model’s learning of local ranking, which is particularly beneficial for Top-K recommendation scenarios.
As the value of K ranges, KGDC shows better performance compared to baselines, except NDCG@50 in Last.FM, and it demonstrates competitive performance in the evaluation of NDCG@K. By explicitly propagating interaction information between users, items, and entities, KGDC effectively learns latent representations of user preferences and item attraction patterns from CF and KG graphs. Additionally, instead of directly integrating the individual graph information, KGDC employs a collaborative guidance mechanism and an alignment mechanism to enhance the interaction between CF and KG. Moreover, KGDC adopts a multi-task framework to ensure the dominance of the supervision signals in CF. These results prove that KGDC has a significant advantage in Top-K recommendation.
In the Top-K recommendation task, BPRMF, as a traditional CF-based method, always performs better than knowledge graph-based CKE on both the Book-Crossing and Last.FM datasets. It also demonstrates that simply integrating KGs into recommendation systems does not always guarantee improved performance. Both the CF and KG graphs contain rich information, while not all information within KGs may contribute effectively to recommendations. Therefore, optimizing recommendation effectiveness requires making comprehensive and coherent use of CF and KG. This also indicates that the proposed KGDC can fully leverage the individual information and the interactive information.
Similar to CTR-based recommendations, KGDC shows greater improvement on the Book-Crossing dataset compared to the Last.FM dataset when K = 10, 20, 50, and 100. This further demonstrates that the proposed method can extract more valuable information from semantically rich knowledge graphs while maintaining the dominant role of CF information and preventing interference from irrelevant redundancies of KG.

5.3. Limitation

KGDC achieved good results on both CTR and Top-K recommendation tasks, but there are still noticeable problems. KGDC does not show stable improvement in the NDCG@K evaluation metric for the Top-K recommendation problem, and there is even no improvement in some cases. This means the model is unable to recommend the items that users like the most at the top of the list. We think this is because KGDC does not distinguish the different ratings of users in CF, leading to fluctuations in the more stringent NDCG performance, which evaluates the ranking order more strictly.

6. Conclusions and Future Work

In this paper, we propose a new method, KGDC, to tackle the aforementioned challenges. KGDC leverages GNN-based information propagation and aggregation techniques to effectively utilize the individual information from CF and KG, and also enhance the interaction of the two graphs by leveraging contrastive learning and multi-task learning. The extensive experiments show that KGDC significantly improves the recommendation performance over state-of-the-art methods on both tasks of CTR recommendation and Top-K prediction.
In this paper, we do not distinguish the different types of interactions between users and items, meaning that all connections in the CF graph are considered to have the same weight. However, in real-world scenarios, different behaviors of users towards items always represent varying attitudes. For example, a user’s preference for a movie rated 5 stars is higher than for one rated 1 star. Therefore, representing and learning different types of behavior in the CF graph can help the model distinguish and explore more nuanced user preferences, thereby improving the accuracy of recommendations. In future work, we will introduce the learning of different user behavior types in the CF graph during both independent and integrated learning processes in the dual-graph framework, further improving the accuracy of recommendations.

Author Contributions

Conceptualization, J.H. and H.Z.; methodology, J.H. and Z.X.; software, J.H. and Z.X.; validation, J.H. and B.Y.; formal analysis, J.H. and B.Y.; investigation, J.H. and C.D.; resources, Z.X.; data curation, J.H.; writing—original draft preparation, J.H.; writing—review and editing, H.Z., Z.X., B.Y., and C.D.; visualization, J.H. and H.Z.; supervision, R.H.; project administration, R.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Covington, P.; Adams, J.; Sargin, E. Deep neural networks for youtube recommendations. In Proceedings of the 10th ACM Conference on Recommendation Systems, Boston, MA, USA, 15–19 September 2016; pp. 191–198. [Google Scholar]
  2. Wang, J.; Huang, P.; Zhao, H.; Zhang, Z.; Zhao, B.; Lee, D.L. Billion-scale commodity embedding for e-commerce recommendation in alibaba. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 839–848. [Google Scholar]
  3. Zheng, G.; Zhang, F.; Zheng, Z.; Xiang, Y.; Yuan, N.J.; Xie, X.; Li, Z. DRN: A deep reinforcement learning framework for news recommendation. In Proceedings of the 2018 World Wide Web Conference, Lyon, France, 23–27 April 2018; pp. 167–176. [Google Scholar]
  4. Lee, D.; Kang, S.; Ju, H.; Park, C.; Yu, H. Bootstrapping user and item representations for one-class collaborative filtering. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 11–15 July 2021; pp. 167–176. [Google Scholar]
  5. Wang, C.; Yu, Y.; Ma, W.; Zhang, M.; Chen, C.; Liu, Y.; Ma, S. Towards representation alignment and uniformity in collaborative filtering. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 14–18 August 2022; pp. 1816–1825. [Google Scholar]
  6. Wu, J.; Wang, X.; Feng, F.; He, X.; Chen, L.; Lian, J.; Xie, X. Self-supervised graph learning for recommendation. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 11–15 July 2021; pp. 726–735. [Google Scholar]
  7. Yu, J.; Yin, H.; Xia, X.; Chen, T.; Cui, L.; Nguyen, Q.V.H. Are graph augmentations necessary? Simple graph contrastive learning for recommendation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, 11–15 July 2022; pp. 1294–1303. [Google Scholar]
  8. Ma, H.; King, I.; Lyu, M.R. Effective missing data prediction for collaborative filtering. In Proceedings of the 30th annual international ACM SIGIR Conference on Research and Development in Information Retrieval, Amsterdam, The Netherlands, 23–27 July 2007; pp. 39–46. [Google Scholar]
  9. Volkovs, M.; Yu, G.; Poutanen, T. Dropoutnet: Addressing cold start in recommendation systems. Adv. Neural Inf. Process. Syst. 2017, 30, 4957–4966. [Google Scholar]
  10. Khawar, F.; Zhang, N.L. Modeling multidimensional user preferences for collaborative filtering. In Proceedings of the 2019 IEEE 35th International Conference on Data Engineering (ICDE), Macao, China, 8–11 April 2019; pp. 1618–1621. [Google Scholar]
  11. Wang, H.; Zhang, F.; Wang, J.; Zhao, M.; Li, W.; Xie, X.; Guo, M. Ripplenet: Propagating user preferences on the knowledge graph for recommendation systems. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management, Torino, Italy, 22–26 October 2018; pp. 417–426. [Google Scholar]
  12. Wang, H.; Zhao, M.; Xie, X.; Li, W.; Guo, M. Knowledge graph convolutional networks for recommendation systems. arXiv 2019, arXiv:1904.12575. [Google Scholar]
  13. Wang, X.; He, X.; Cao, Y.; Liu, M.; Chua, T.S. Kgat: Knowledge graph attention network for recommendation. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 950–958. [Google Scholar]
  14. Wang, H.; Zhang, F.; Zhang, M.; Leskovec, J.; Zhao, M.; Li, W.; Wang, Z. Knowledge-aware graph neural networks with label smoothness regularization for recommendation systems. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 968–977. [Google Scholar]
  15. Wang, Z.; Lin, G.; Tan, H.; Chen, Q.; Liu, X. CKAN: Collaborative knowledge-aware attentive network for recommendation systems. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 25–30 July 2020; pp. 219–228. [Google Scholar]
  16. Huang, J.; Zhao, W.X.; Dou, H.; Wen, J.R.; Chang, E.Y. Improving sequential recommendation with knowledge-enhanced memory networks. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 505–514. [Google Scholar]
  17. Zhang, F.; Yuan, N.J.; Lian, D.; Xie, X.; Ma, W.Y. Collaborative knowledge base embedding for recommendation systems. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 353–362. [Google Scholar]
  18. Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating embeddings for modeling multi-relational data. Adv. Neural Inf. Process. Syst. 2013, 2787–2795. [Google Scholar]
  19. Wang, Z.; Zhang, J.; Feng, J.; Chen, Z. Knowledge graph embedding by translating on hyperplanes. In Proceedings of the AAAI Conference on Artificial Intelligence, Québec, QC, Canada, 27–31 July 2014; Volume 28. [Google Scholar]
  20. Hu, B.; Shi, C.; Zhao, W.X.; Yu, P.S. Leveraging meta-path based context for top-n recommendation with a neural co-attention model. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 1531–1540. [Google Scholar]
  21. Shi, C.; Hu, B.; Zhao, W.X.; Philip, S.Y. Heterogeneous information network embedding for recommendation. IEEE Trans. Knowl. Data Eng. 2018, 31, 357–370. [Google Scholar] [CrossRef]
  22. Wang, X.; Wang, D.; Xu, C.; He, X.; Cao, Y.; Chua, T.S. Explainable reasoning over knowledge graphs for recommendation. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 5329–5336. [Google Scholar]
  23. Sha, X.; Sun, Z.; Zhang, J. Hierarchical attentive knowledge graph embedding for personalized recommendation. Electron. Commer. Res. Appl. 2021, 48, 101071. [Google Scholar] [CrossRef]
  24. Wang, X.; Huang, T.; Wang, D.; Yuan, Y.; Liu, Z.; He, X.; Chua, T.S. Learning intents behind interactions with knowledge graph for recommendation. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; pp. 878–887. [Google Scholar]
  25. Cao, Y.; Wang, X.; He, X.; Hu, Z.; Chua, T.S. Unifying knowledge graph learning and recommendation: Towards a better understanding of user preferences. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 151–161. [Google Scholar]
  26. Yu, X.; Ren, X.; Sun, Y.; Gu, Q.; Sturt, B.; Khandelwal, U.; Norick, B.; Han, J. Personalized entity recommendation: A heterogeneous information network approach. In Proceedings of the 7th ACM International Conference on Web Search and Data Mining, New York, NY, USA, 24–28 February 2014; pp. 283–292. [Google Scholar]
  27. Chen, Y.; Yang, Y.; Wang, Y.; Bai, J.; Song, X.; King, I. Attentive knowledge-aware graph convolutional networks with collaborative guidance for personalized recommendation. In Proceedings of the 2022 IEEE 38th International Conference on Data Engineering (ICDE), Virtual, 9–12 May 2022; pp. 299–311. [Google Scholar]
  28. Zou, D.; Wei, W.; Wang, Z.; Mao, X.L.; Zhu, F.; Fang, R.; Chen, D. Improving knowledge-aware recommendation with multi-level interactive contrastive learning. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, Atlanta, GA, USA, 17–21 October 2022; pp. 2817–2826. [Google Scholar]
  29. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the 37th International Conference on Machine Learning, Virtual, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  30. Yu, J.; Yin, H.; Gao, M.; Xia, X.; Zhang, X.; Viet Hung, N.Q. Socially-aware self-supervised tri-training for recommendation. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Singapore, 14–18 August 2021; pp. 2084–2092. [Google Scholar]
  31. Pan, Z.; Chen, H. Collaborative knowledge-enhanced recommendation with self-supervisions. Mathematics 2021, 9, 2129. [Google Scholar] [CrossRef]
  32. Yang, Y.; Huang, C.; Xia, L.; Li, C. Knowledge graph contrastive learning for recommendation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, 11–15 July 2021; pp. 1434–1443. [Google Scholar]
  33. Wang, H.; Xu, Y.; Yang, C.; Shi, C.; Li, X.; Guo, N.; Liu, Z. Knowledge-adaptive contrastive learning for recommendation. In Proceedings of the sixteenth ACM International Conference on Web Search and Data Mining, Singapore, 27 February–3 March 2023; pp. 535–543. [Google Scholar]
  34. Rendle, S.; Freudenthaler, C.; Gantner, Z.; Schmidt-Thieme, L. BPR: Bayesian personalized ranking from implicit feedback. arXiv 2012, arXiv:1205.2618. [Google Scholar]
Figure 1. The framework of KGDC. The red arrows in the figure indicate the higher weights of the related information.
Figure 1. The framework of KGDC. The red arrows in the figure indicate the higher weights of the related information.
Information 15 00534 g001
Figure 2. Average results of Recall@K and NDCG@K in Top-K Recommendation task.
Figure 2. Average results of Recall@K and NDCG@K in Top-K Recommendation task.
Information 15 00534 g002
Table 1. Basic statistics of the two datasets.
Table 1. Basic statistics of the two datasets.
Book-CrossingLast.FM
User–item Interaction# users17,8601872
# items14,9673846
# interactions139,74642,346
Knowledge Graph# entities77,9039366
# relations2560
# triplets151,50015,518
Table 2. Average results of AUC and F1 in CTR prediction task. The best results are in boldface and the second best results are underlined.
Table 2. Average results of AUC and F1 in CTR prediction task. The best results are in boldface and the second best results are underlined.
ModelBook-CrossingLast.FM
AUCF1AUCF1
BPRMF0.65830.61170.75630.7010
CKE0.67590.62350.74710.6740
KGCN0.68410.63130.80270.7086
KGNN-LS0.67620.63140.80520.7224
KGAT0.73140.65440.82930.7424
CKAN0.74200.66710.84180.7592
KGIN0.72730.66140.84860.7602
KGIC0.74730.66900.85920.7753
CG-KGR0.74720.67940.83680.7424
Our KGDC0.76560.68720.86520.7771
%Improv2.45%1.15%0.7%0.23%
Table 3. Average results of Recall@5 and NDCG@5 in Top-K recommendation task. The best results are in boldface and the second best results are underlined.
Table 3. Average results of Recall@5 and NDCG@5 in Top-K recommendation task. The best results are in boldface and the second best results are underlined.
ModelBook-CrossingLast.FM
RecallNDCGRecallNDCG
BPRMF0.02670.01950.05140.0461
CKE0.03040.01670.0590.0404
KGCN0.03160.04710.0780.0569
KGNN-LS0.03520.05180.0660.0512
KGAT0.02760.01810.0670.0511
CKAN0.02730.01790.09010.0724
KGIN0.02940.02340.09230.0812
KGIC0.03880.03160.09060.0906
CG-KGR0.05670.07470.09870.0711
Our KGDC0.06140.07580.11430.1018
%Improv8.29%0.15%15.81%12.36%
Table 4. Average results of Recall@10 and NDCG@10 in Top-K recommendation task. The best results are in boldface and the second best results are underlined.
Table 4. Average results of Recall@10 and NDCG@10 in Top-K recommendation task. The best results are in boldface and the second best results are underlined.
ModelBook-CrossingLast.FM
RecallNDCGRecallNDCG
BPRMF0.03410.02180.10140.0699
CKE0.03150.01710.07840.0431
KGCN0.05730.05110.12430.0823
KGNN-LS0.06940.05810.10470.0734
KGAT0.03590.02490.11110.0749
CKAN0.03420.02510.12120.0994
KGIN0.04010.02450.13510.0978
KGIC0.04500.03180.14260.1136
CG-KGR0.08110.07250.16750.1012
Our KGDC0.09630.08190.17730.1264
%Improv18.74%12.97%5.85%11.27%
Table 5. Average results of Recall@20 and NDCG@20 in Top-K recommendation task. The best results are in boldface and the second best results are underlined.
Table 5. Average results of Recall@20 and NDCG@20 in Top-K recommendation task. The best results are in boldface and the second best results are underlined.
ModelBook-CrossingLast.FM
RecallNDCGRecallNDCG
BPRMF0.04670.02800.16840.0875
CKE0.04380.02170.11510.0496
KGCN0.07850.05930.18250.0973
KGNN-LS0.08510.06060.17730.0911
KGAT0.05340.03010.18220.0931
CKAN0.06190.03470.20780.1194
KGIN0.06590.3540.22570.1347
KGIC0.06660.03970.21790.1405
CG-KGR0.10810.07690.24420.1410
Our KGDC0.11720.08210.25850.1502
%Improv8.418%6.762%5.856%6.525%
Table 6. Average results of Recall@50 and NDCG@50 in Top-K recommendation task. The best results are in boldface and the second best results are underlined.
Table 6. Average results of Recall@50 and NDCG@50 in Top-K recommendation task. The best results are in boldface and the second best results are underlined.
ModelBook-CrossingLast.FM
RecallNDCGRecallNDCG
BPRMF0.05970.03710.26930.1201
CKE0.05020.02590.24150.1026
KGCN0.11000.06130.27830.1097
KGNN-LS0.13470.07010.28340.1273
KGAT0.06710.03490.29120.1235
CKAN0.08960.04070.31010.1516
KGIN0.10630.04140.29420.1567
KGIC0.12270.04880.37690.1765
CG-KGR0.15270.09010.28940.137
Our KGDC0.17920.10480.39510.1705
%Improv17.35%16.32%5.856%-0.34%
Table 7. Average results of Recall@100 and NDCG@100 in Top-K recommendation task. The best results are in boldface and the second best results are underlined.
Table 7. Average results of Recall@100 and NDCG@100 in Top-K recommendation task. The best results are in boldface and the second best results are underlined.
ModelBook-CrossingLast.FM
RecallNDCGRecallNDCG
BPRMF0.06120.04230.37110.1348
CKE0.05170.02630.37930.1294
KGCN0.14510.06870.40010.1323
KGNN-LS0.15990.07870.37720.1371
KGAT0.09120.04100.39970.1426
CKAN0.13040.05100.42340.1785
KGIN0.13210.05790.41670.1815
KGIC0.17190.05440.48420.1981
CG-KGR0.21870.10730.38740.1544
Our KGDC0.24400.11540.49790.2112
%Improv10.37%7.55%2.83%6.612%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, J.; Xie, Z.; Zhang, H.; Yang, B.; Di, C.; Huang, R. Enhancing Knowledge-Aware Recommendation with Dual-Graph Contrastive Learning. Information 2024, 15, 534. https://doi.org/10.3390/info15090534

AMA Style

Huang J, Xie Z, Zhang H, Yang B, Di C, Huang R. Enhancing Knowledge-Aware Recommendation with Dual-Graph Contrastive Learning. Information. 2024; 15(9):534. https://doi.org/10.3390/info15090534

Chicago/Turabian Style

Huang, Jinchao, Zhipu Xie, Han Zhang, Bin Yang, Chong Di, and Runhe Huang. 2024. "Enhancing Knowledge-Aware Recommendation with Dual-Graph Contrastive Learning" Information 15, no. 9: 534. https://doi.org/10.3390/info15090534

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop