Next Article in Journal
Reservoir Optimization Scheduling Driven by Knowledge Graphs
Previous Article in Journal
Improving Single-Image Super-Resolution with Dilated Attention
Previous Article in Special Issue
Improving Object Detection Accuracy with Self-Training Based on Bi-Directional Pseudo Label Recovery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Embedding Enhancement Method for LightGCN in Recommendation Information Systems

Department of Computer Science, Kyonggi University, Suwon 16227, Republic of Korea
*
Authors to whom correspondence should be addressed.
Electronics 2024, 13(12), 2282; https://doi.org/10.3390/electronics13122282
Submission received: 30 April 2024 / Revised: 31 May 2024 / Accepted: 4 June 2024 / Published: 11 June 2024

Abstract

:
In the modern digital age, users are exposed to a vast amount of content and information, and the importance of recommendation systems is increasing accordingly. Traditional recommendation systems mainly use matrix factorization and collaborative filtering methods, but problems with scalability due to an increase in the amount of data and slow learning and inference speeds occur due to an increase in the amount of computation. To overcome these problems, this study focused on optimizing LightGCN, the basic structure of the graph-convolution-network-based recommendation system. To improve this, techniques and structures were proposed. We propose an embedding enhancement method to strengthen the robustness of embedding and a non-combination structure to overcome LightGCN’s weight sum structure through this method. To verify the proposed method, we have demonstrated its effectiveness through experiments using the SELFRec library on various datasets, such as Yelp2018, MovieLens-1M, FilmTrust, and Douban-book. Mainly, significant performance improvements were observed in key indicators, such as Precision, Recall, NDCG, and Hit Ratio in Yelp2018 and Douban-book datasets. These results suggest that the proposed methods effectively improved the recommendation performance and learning efficiency of the LightGCN model, and the improvement of LightGCN, which is most widely used as a backbone network, makes an important contribution to the entire field of GCN-based recommendation systems. Therefore, in this study, we improved the learning method of the existing LightGCN and changed the weight sum structure to surpass the existing accuracy.

1. Introduction

Today, with the rapid development of the digital age, a lot of content and information is created per second in the modern digital environment, and Internet users are exposed to such large amounts. This forces users to filter massive amounts of data to find the information or products that they want.
Recommendation systems act as a key tool to present the information or content desired by users in this information overload state, and are used in various areas such as e-commerce, entertainment, and social media; moreover, they are used to provide the content desired by users on platforms such as YouTube, Netflix, and Amazon. In order to recommend user-specific content, the most relevant information or products are effectively presented by analyzing the user’s behavioral patterns, such as the user’s past behavior, preferences, and interests. This provides opportunities for new discoveries and increases user loyalty to the content platform. In addition, by continuously providing relevant and interesting content to users, it allows users to spend more time on the platform, thereby increasing its profitability.
However, the existing recommendation system mainly uses the matrix factorization (MF) approach and the collaborative filtering (CF) approach, but the amount of data used for recommendation increases. In identifying similarities, a scalability problem occurs, in which the amount of computer calculations increases enormously, and the need for a method to solve problems such as a slow learning speed and an increased inference time has emerged due to this problem. As a way to effectively solve this problem, a recommendation system using deep learning based on the CF approach is being proposed. Typically, the graph convolution network (GCN) captures the correlation between users and items better than the existing methods, and, through this, studies are being proposed to improve the learning speed, inference speed, and accuracy of the existing methods [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18].
Therefore, in this study, we have conducted research to optimize and improve the accuracy of the LightGCN model, which is mainly used as a backbone network in the latest GCN-based recommendation system. We propose a weight forwarding technique to improve learning speed and accuracy and use this method to improve the existing LightGCN model. Since the LightGCN model has been widely used as a backbone network in recent research, it is possible to improve the latest model by improving that model. A study was conducted to improve the layer combination structure.

2. Related Work

The MF approach, which is mainly used in existing recommendation systems, is a method of decomposing a matrix representing interactions and potential factors between users and items into two low-dimensional matrices. Hidden features (latent factors) such as user preferences or item characteristics are reflected in this decomposed matrix. In the MF approach, the collaborative signal, which refers to the latent feature vector of users who consumed items similar to mine, is used as the learning data, and the learning is performed using an optimization algorithm, such as gradient descent. After completing the learning process, a specific item in the matrix, which is reconstructed by multiplying the two low-dimensional matrices, represents the rating or preference that the user is expected to have for that item. However, it is computationally very complex to perform matrix decomposition on large-scale datasets with this method, and this process requires a lot of time and resources. Therefore, it is difficult for the MF approach to reflect dynamic changes in the system as user preferences or item characteristics change over time. In addition, the CF approach, which is mainly used in existing recommendation systems, is a recommendation method that recommends items preferred by neighbors with similar tastes to those of the target user to the target user themself. This methodology views user–item interaction data as a matrix and makes recommendations by restoring the entire matrix, like the MF methodology. Through this process, the similarity between the users or between the items is calculated based on evaluations of the items performed by past users, and recommendations are provided between users and items, allowing ‘neighbors’ with similar tastes to interact with the target user. Items that the user has not interacted with can be recommended as high-ranking items [19,20,21,22].
However, since the same problems as those seen in the MF methodology still occur in the CF methodology, deep learning technology is being proposed as a way to improve, starting with the neural graph collaborative filtering (NGCF) [23] model and the graph convolution network (GCN). A variety of recommended systems are being studied.
The NGCF model is used in the GCN-based CF model by explicitly encoding collaborative signals using GCN to consider high-order connectivity through user–item interaction data in a graph converted to a bipartite graph (Figure 1a). Here, high-order connectivity refers to high-order connectivity that cannot be expressed as a bipartite graph, and high-order connectivity generally refers to a path that reaches u 1 from all nodes with a path length greater than one. These high-order connections contain rich semantics that convey collaborative signals. As shown in Figure 1b, for example, the path u 1 i 2 u 2 indicates the behavioral similarity between u 1 and u 2 , because both users have interacted with i 2 . Looking at the longer path u 1 i 2 u 2 i 4 , we can see that u1 is likely to adopt i 4 , because u 2 , a similar user, has used i 4 before. Since i 4 has two connected paths and i 5 has one connected path, from the holistic perspective of < i 4 , u 1 > < i 5 , u 1 > − l = 3, i 4 is more likely to be of interest to u 1 than i 5 . These high-order connections can be used to provide more accurate recommendations by capturing collaborative signals between the users and items. A collaborative signal reflecting high-order connections is a concept proposed in the NGCF model, which includes the target users and neighboring users whose interaction history overlaps with the item’s embedding vector, or the user who interacted with the neighbors but did not interact with the target. This is the embedding vector of the item. In addition, while the MF methodology and CF methodology implicitly reflected this collaborative signal, the NGCF explicitly reflected this, and, through this method, showed improvement in recommendation performance compared to the existing MF and CF methodologies.
Equations (1) and (2) are the propagation rule and prediction layer of the NGCF model, respectively, and learning is performed using only the initial layer (shallow embedding) of Equation (2), e u 0 and e i 0 . Here, hop refers to the distance between the nodes in the graph, and the layer of the GCN algorithm serves to aggregate the information of the surrounding nodes. Therefore, when proceeding with the propagation rule, when passing through the first layer, the information of the target node’s 1-hop neighbors is aggregated; moreover, in the case of layer 2, the information of the 1-hop’s neighbors, that is, up to the 2-hop, is aggregated. If it goes through N-layers, the information up to N-hops is aggregated. The author of NGCF first named this aggregated information a collaboration signal and proposed a process of disseminating and learning the collaboration signal containing information of such high-order connectivity to the GCN, leading to the use of deep learning in the recommendation system. N u and N i in Equation (1) refer to the connected nodes of the user u and the item i , respectively, and the embedding value when performing the convolution operation according to the symmetric normalization term (= 1 / N u N i ). It serves to normalize in order to prevent it from becoming large, which means dividing by the number of neighboring nodes (users, items). The reason for normalization is that the more items connected to user u , the larger the collaborative signal becomes. The corresponding term ( 1 N u N i W 1 e i k + W 2 e i k e u k ) shows that it is a process of performing message construction, and, as more messages are delivered in the process, the expression ability and recommendation performance increase.
e u k is the embedding value in the k th layer of user u , and e i k is the embedding value in the k th layer of item i . σ means non-linear activation, and W 1 e u k and W 1 e i k mean self-connection, where W 1 and W 2 are learnable weight matrices. To obtain the embedding value of the k + 1 th layer containing the k th hop in the graph, the previous embedding values of the neighbors of user u and item i are added to the normalized weight.
e u k + 1 = σ W 1 e u k + i N u 1 N u N i W 1 e i k + W 2 e i k e u k , e i k + 1 = σ W 1 e i k + i N i 1 N u N i W 1 e u k + W 2 e u k e i k
e u * = e u 0 e u L , e i * = e i 0 e i L
LightGCN [24], an improved NGCF model, simplifies complex weight learning by removing feature transformation, non-linear activation, and self-connection from the propagation layer (Equation (3)). The prediction layer is shown in Equation (4). In addition, by changing the existing NGCF method to a weighted sum, all layers are summed and the last embedding values of the user and item, e u and e i , are calculated through their average ( α k = 1 / k ). Here, the weighted sum process used to obtain the last embedding value includes the initial layer embedding value. Score y ^ u i , which represents the preference between user u and item i, is calculated by inner producing e u and e i , as shown in Equation (5). The matrix obtained through the inner product is used for recommendation. The basic learning process is shown in Figure 2. LightGCN uses an interaction graph between users and items to strengthen interactions between the nodes. As a result, the recommendation performance is improved by generating the node embedding value using only the connection information between the user and the item. LightGCN generates node embeddings using the interaction graph between the nodes instead of the weight learning of the GCN algorithm; therefore, the learning and inference time can be significantly reduced compared to NGCF. Additionally, its architecture is simpler than that of NGCF, so it can be usefully used in various recommendation scenarios and large-scale recommendation systems. Due to these various advantages, the LightGCN model is the most widely used backbone network in follow-up research.
e u k + 1 = i N u 1 N u N i e i k ;     e i k + 1 = i N i 1 N i N u e u k
e u = k = 0 n α k e u k ;     e i = k = 0 n α k e i k
y ^ u i = e u T e i
However, NGCF and LightGCN repeatedly perform a symmetric normalization term to obtain the embedding of the next layer and transmit the value to the next layer. They experience an underfitting problem, in which the embedding value decreases each time that it is transmitted to the next layer. These problems impede LightGCN’s learning process and degrade its recommendation performance, preventing the original purpose of the recommendation system from being achieved. Additionally, the embedding value used for the recommendation of the LightGCN model is used after averaging the averages, as shown in Equation (4). Layer 1 learns only the 1-hop relationship and repeats it by increasing the length of the hop. The learning results of each hop are passed on to the next layer. The relationship data with a short previous hop length are the items that are most relevant to the user. The most related relationships are averaged and overlapped with the less related ones. The stronger the embedding value for this most relevant relationship, the lower the recommendation accuracy and recommendation diversity in the recommendation system. To overcome these problems, we improve the learning speed of the GCN algorithm and propose techniques and improvement structures to increase its recommendation performance.

3. Proposal Method (Embedding Enhancement Method)

LightGCN, whose embedding values continued to decrease during the learning process, robustly learns only the item features that the user has continuously viewed, allowing the users to discover new and relevant content that they would not have encountered otherwise, which is the essential reason for using a recommender system. With this, the probability decreases further. As a way to overcome this issue, in the process of passing the embedding of the previous layer to the next layer (as shown in Equation (6)), a scalar multiplication is performed on the embedding value that has performed a symmetric normalization term, resulting in enhancement embedding e u k + 1 , e i k + 1 being passed to the next layer. Through this process, you will obtain enhancement embedding. Through the obtained embedding value, the user not only learns the characteristics of the item that he or she has continuously viewed, but also learns the relationship between other users and items, so that new items that have a relationship between the user and the item, but were not previously recommended, can be recommended. From this, the probability increases. Additionally, in the existing LightGCN structure, the self-connected effect of graph convolution is maintained by weight summing the embedding value derived from the previous hop layer, leading to the extraction of comprehensive representation. The embedding calculated by taking the weight sum overlaps the embedding value in the order of short hop length, due to the summation process, which has the effect of being further emphasized. However, the embedding value obtained through the proposed embedding enhancement method robustly contains the information of all hops, and, instead of using the enhancement embedding to perform a weight sum process (as shown in Equation (7)), it is recommended to use only the last layer embedding value, as shown in Equation (8). This makes it possible to provide more diverse recommendations than the existing LightGCN. The proposed non-combination structure specifically addresses LightGCN’s weight sum issue by avoiding the averaging of embedding values from each layer and instead utilizing only the embedding value from the final layer. The reason for adopting this method is that the interference from the initial layer embeddings can actually degrade the recommendation performance, and increasing the layer depth beyond a certain point can also lead to performance degradation. Therefore, it is crucial to identify the optimal layer for each dataset or requirement and use only the results from this optimal layer for inference, which enhances the performance of the recommendation system. This approach minimizes the interference from the initial layers, allowing the model to learn deeper and more meaningful representations. By leveraging the embedding from the final layer, the model captures the most relevant and significant features for making recommendations. Furthermore, using the final layer embedding value helps us to provide more diverse recommendations, as it robustly includes the information from all hops without the overlapping effect seen in the weight-summed embeddings. This method increases the likelihood of recommending new items that have a relationship with the user but were not previously recommended, thus improving the overall recommendation quality. By implementing this non-combination structure, the recommender system not only learns the characteristics of the items that the user has continuously viewed, but also understands the relationships between the other users and items, facilitating the recommendation of previously unseen but relevant content. This ensures that the proposed method outperforms the existing LightGCN in providing diverse and accurate recommendations. A schematic diagram of the proposed method is shown in Figure 3.
e u k + 1 = w k i N u 1 N u N i e i k ;     e i k + 1 = w k i N i 1 N i N u e u k
e u = e u k ;     e i = e i k
y ^ u i = e u T e i

4. Experiment

In this study, a non-combination structure was proposed through the embedding enhancement method as a method to improve the problems that hinder the recommendation performance of LightGCN, as well as learning efficiency. The weight of the embedding enhancement of the proposed method is 2 of 0. To verify the proposed technique and structure, an experiment was conducted by applying the proposed method to LightGCN by increasing the weight from the power of 2 to the 11th power of 2. Since LightGCN is widely used as a backbone network in follow-up research, we believe that, if the performance and learning efficiency of LightGCN increases, the network proposed in follow-up research will also be able to produce meaningful results. SELFRec [25] was used in this experiment, which is an open-source library for GCN-based recommendation system algorithms. This library implements various recommendation algorithms based on PyTorch and facilitates recommendation system research and development. Additionally, several subsequent studies have used the same library as this experiment. Section 4.1 shows the experimental environment and data, and Section 4.2 shows the analysis results of the overall experiment. The link to the code written referring to SELFRec is as follows: Github, https://github.com/d9249/OptGCN (accessed on 22 February 2024). The learning records are as follows and have been posted on the Internet: Train log: https://wandb.ai/d9249/OptGCN (accessed on 22 February 2024).

4.1. Experimental Data and Environment

In experiments conducted to verify the proposed method, four types of datasets were used, as shown in Table 1. The MovieLens dataset is a dataset that is widely used in movie recommendation systems and contains interaction information about users’ ratings of movies. Here, the interaction data refers to the viewing history between the user and the item, and this interaction relationship is expressed as a bipartite graph; furthermore, the interaction is learned in order to predict the mapping between the user and the item. The size of the dataset is divided based on the number of user viewing records. Likewise, the FilmTrust dataset is a dataset created by crawling users’ movie evaluation information on the FilmTrust website, and, like the MovieLens dataset, it is often used in movie recommendation systems. In this experiment, only the rating information of the MovieLens dataset was used for learning between 4 and 5 points, but the FilmTrust dataset used rating information from 0.5 to 4 points for learning. The Douban-book dataset is a dataset created by crawling from Douban, a Chinese book review website, and includes book ratings and user information. The rating range for the dataset used in this experiment was four to five points. The Yelp2018 dataset comes from the Yelp Challenge 2018 and records ratings for local businesses, such as restaurants and bars, in 10 metropolitan areas across two countries, showing only those for which users have interacted with items.
The datasets used in the experiments of this study were the same as those used in the LightGCN follow-up study, and no preprocessing was performed on the data. These datasets are widely used as benchmark datasets for evaluating the performance of recommender system algorithms. To avoid overfitting problems, we implemented several strategies. First, we limited the training to 300 epochs and monitored the performance metrics during the training process. In particular, the epochs showing the highest performance metrics were selected to evaluate and compare the results. This approach prevents overfitting by preventing the model from continuing to learn beyond its optimal performance point. Second, we implemented early stopping based on the performance of the validation set. If the validation loss did not improve over a period of epochs, we stopped training to avoid overfitting the model to the training data. Finally, all accuracy analyses were performed using a separate validation dataset that the model did not see during training. This ensured that the reported performance metrics reflected the model’s ability to generalize to unknown data. These strategies ensured the robustness of the results and prevented overfitting problems. By using early stopping and monitoring the performance of a separate validation dataset, we can provide strong support that our model maintains its ability to generalize across different datasets.
Information about the experiment environment of this study is shown in Table 2, and the parameters used for learning LightGCN are also shown in the table.

4.2. Experimental Results and Analysis

Figure 4 shows the analysis results achieved through dimensionality reduction techniques (DRT) using embedding values and datasets. In this analysis, principle component analysis (PCA) and Gaussian kernel density estimation (GKDE) were performed for each dataset to analyze the distribution and density of the embeddings. Additionally, for visualization purposes, we performed a data sampling process to rank the users and items by popularity, and then randomly selected 500 hot users (blue) and hot items (green) from the top 5% of users and item groups. Here, 500 cold users (red) and cold items (orange) were randomly sampled from the bottom 80% of the user and item groups. The expression learned through the sampling process was mapped to a two-dimensional space using t-SNE, and PCA visualization was performed through the expression mapped to the two-dimensional space. Additionally, the GKDE visualization results are shown below, and each point (x,y) was visualized using GKDE of atan(x,y).
The visualizations shown in Figure 4 allow us to evaluate the distribution and density of the embeddings, providing insights into factors such as separability, density, uniformity, clustering, noise, outliers, and stability. The visualizations represent the best-performing proposed method for each dataset. For the existing LightGCN, multiple clusters appeared in the PCA visualization, but the boundaries between the clusters were relatively less distinct compared to those of the proposed method. This suggests that the embeddings are not clearly separated and are intermingled. A similar pattern was observed in the GKDE visualization, with peaks indicating that specific feature values are concentrated in certain areas. In contrast, the embeddings resulting from the proposed algorithm showed relatively distinct clustering and wider dispersion in the PCA visualization. This indicates that the proposed method captures the diversity and features of the data more effectively. The GKDE results also showed relatively smoother curves, with the peaks being more subdued, indicating that the embeddings are more uniformly distributed. The use of t-SNE and PCA for dimensionality reduction and visualization is appropriate for several reasons. The t-SNE is particularly effective at preserving local structures in high-dimensional data, making it well-suited for visualizing the clusters and relationships within the data. By mapping the high-dimensional embeddings to a 2D space, t-SNE helps us to reveal natural groupings and separations that might not be apparent in the original high-dimensional space. PCA, on the other hand, is a linear dimensionality reduction technique that captures the global variance structure of the data. By projecting the embeddings onto the principal components, PCA helps us to understand the overall distribution and spread of the data. It provides insights into the major directions of variance and helps us to identify the primary factors contributing to the data structure. Combining t-SNE for initial dimensionality reduction with PCA for subsequent visualization leverages the strengths of both methods. t-SNE ensures that the local relationships are maintained, making the clusters more apparent, while PCA provides a broader view of the data’s variance and distribution. This combination allows for a more comprehensive analysis of the embeddings, making it easier to evaluate their quality in terms of clustering, dispersion, and uniformity. These visualization results suggest that the proposed method learns the distribution of the embeddings more effectively than the existing method, better reflecting the diverse features of the data. Consequently, the proposed method enhances the quality of the embeddings by achieving clearer clustering, wider dispersion, and more uniform distribution, thus better representing the complexity of the data. The embedding values changed through the proposed method are shown in Table 3.
Four evaluation functions were used to verify the entire experiment, as follows: Precision, Recall, NDCG, and Hit Ratio, which were used to evaluate the quality of the model from various aspects. The Precision@K represents the ratio of items of interest to actual users among the recommendations proposed by the recommendation system. In other words, it measures how accurately the system actually recommends the preferred items to the user, and high precision indicates that the recommendation system accurately recommends the items that are highly relevant to the user. The Recall@K represents the ratio of items suggested by the recommendation system among the items that the actual user is interested in. In other words, it measures how many related items the recommender system recommends, and a high recall rate means that it suggests many related items to the user without missing them. Normalized discounted cumulative gain (NDCG)@K is a function that evaluates the relevance and ranking of the recommended items. The higher the user’s preferred items are, the higher the value is. Therefore, a high NDCG indicates that the recommendation system accurately suggests items of interest to the user with a high ranking. The Hit Ratio@K represents the rate at which the recommendation system successfully includes the user’s preferred items in the recommendation list. This measures whether the recommendation system recommends at least one item of interest to the user, and a high hit rate means that the recommendation system includes the user’s preferred items well, without missing them.
Through these various evaluation indicators, the performance of the recommendation system was evaluated and compared from various aspects. Precision and Recall evaluate the accuracy and diversity aspects of the recommendations, NDCG evaluates whether the items of interest to the user are accurately recommended at the top, and Hit Ratio measures how successful the recommendation system is in suggesting at least one related item to the user. The entire experiment was conducted with K = 20, and the quantitative results of the entire experiment are shown in Table 4.
The analysis of the experimental results for each dataset was performed using the results shown in Table 4. On the Yelp2018 dataset, the proposed method achieved the best performance, showing remarkable improvement in all evaluation indicators. This dataset achieved an increase of 15.07% in Precision, 15.53% in Recall, 15.74% in NDCG, and 15.06% in Hit ratio. These numbers indicate a significant improvement in the model’s ability to accurately recommend relevant items and encompass a wider range of user interests, and consistent improvements in all of these metrics suggest that LightGCN_N was particularly effective for the Yelp2018 dataset.
In the case of the MovieLens-1M dataset, the performance improvement was relatively small compared to that of Yelp2018. An improvement of approximately 2% was achieved in both the Recall and NDCG metrics, and this increase indicates a slight improvement in the model’s ability to capture relevant items for user recommendations. Precision and Hit Ratio also showed a slight improvement, with an increase of about 1%. This confirmed that LightGCN_N improved in recommendation quality, but the impact was not as clear as that observed for the Yelp2018 dataset. These results occurred due to inherent differences between the datasets.
The FilmTrust dataset showed an enhanced performance across all metrics, each showing an increase of approximately 1%. This slight improvement observed in the results after applying the proposed method indicates limited improvement compared to the basic LightGCN model. The relatively low increase compared to the other datasets suggests that the proposed technique may have a limited impact on datasets with similar characteristics to those of FilmTrust.
The Douban-book dataset showed the most notable improvement among all datasets, especially for NDCG, with an increase of 16.98%. Additionally, both Precision and Hit Ratio increased by about 14.4%, and Recall improved by 12.97%. This suggests that LightGCN_N is very effective on the Douban-book dataset. This increase in all metrics indicates that the proposed method not only improved the overall recommendation accuracy, but also better identified a wider range of relevant items for users. It can be seen that the proposed method is suitable for the characteristics of the Douban-book dataset, including item types and user interaction patterns.

5. Conclusions

In the digital age, users are overwhelmed by the vast amount of content and information available online, and, as a solution to this, recommendation systems play an important role in helping users to find relevant content. Traditional recommendation systems utilizing matrix factorization and collaborative filtering approaches have scalability issues, due to increasing data volumes. This causes scalability problems and a decrease in the learning and inference speed of the existing methods. Therefore, this study aimed to optimize and improve the accuracy of the LightGCN model, which is a widely used backbone network in GCN-based recommendation systems. For this purpose, the following method was proposed: As a way to improve the robustness of the embeddings, the robustness between the layers was maintained by increasing the embeddings through weights after symmetric normalization. Using the enhancement embedding obtained through this process, learning was completed without going through the existing weighted sum, and only the last layer’s embedding was used for recommendation; however, it surpassed the accuracy of the existing recommendation performance. What can be seen from these experiments is that the proposed method does not recommend only the items with a small number of hops—although there was a user–item relationship in the past—but is capable of recommending new and diverse items that had longer hops and were not otherwise recommended. The entire experiment was conducted on the MovieLens, FilmTrust, Douban-book, and Yelp2018 datasets using the SELFRec library, and Precision, Recall, NDCG, and Hit Ratio were evaluated as performance indicators for model verification. The analysis of the entire experiment showed notable improvements in all datasets, especially the Yelp2018 and Douban-book datasets. There were notable increases in these four indicators.
The detailed experimental results for each dataset are as follows: In most datasets, the proposed LightGCN_N showed improvements in all evaluation indicators. In the Yelp2018 dataset, Precision increased by 15.07%, Recall increased by 15.53%, NDCG increased by 15.74%, and Hit Ratio increased by 15.06%. This means that the proposed LightGCN_N can cover the user’s interests more broadly and recommend related items more accurately than the existing LightGCN model. In the MovieLens-1M dataset, Precision increased by 1.14%, Recall increased by 2.03%, NDCG increased by 1.84%, and Hit Ratio increased by 1.13%. This suggests that, although the LightGCN_N model improves in recommendation quality, the effect is relatively less pronounced compared to that of the Yelp2018 dataset. In the FilmTrust dataset, Precision increased by 0.72%, Recall increased by 0.58%, NDCG increased by 1.44%, and Hit Ratio showed a performance improvement of 0.71%. The impact of the proposed technique was found on the datasets with similar characteristics to those of FilmTrust. This means that it may be limited. The Douban-book dataset showed the most notable performance improvement, especially for NDCG, which increased by 16.98%. Additionally, Precision and Hit Ratio improved by 14.4%, and Recall increased by 12.97%. This indicates that the proposed method improved the recommendation accuracy overall and provided the users with a wider range of related items.
The comprehensive results show that the proposed methods effectively improved the recommendation performance and learning efficiency of LightGCN, especially showing notable improvements in the Yelp2018 and Douban-book datasets. This can be seen as an important contribution that significantly improves the performance of the LightGCN model in the field of recommender systems. In the future, we plan to verify the proposed methods by applying LightGCN (which was proposed in a follow-up study) to algorithms [29,30,31,32,33] that use the backbone network, and to learn to overcome the problem of having to manually search for weights to obtain enhancement embedding. In the process, we plan to perform a learning process in which the optimal weight is calculated. These methods will be applied to the latest recommendation model.

Author Contributions

Conceptualization, S.L., J.A. and N.K.; methodology, N.K.; software, S.L.; validation, S.L. and N.K.; formal analysis, N.K.; investigation, S.L.; resources, N.K.; data curation, S.L.; writing—original draft preparation, S.L.; writing—review and editing, S.L.; visualization, S.L.; supervision, N.K.; project administration, N.K.; funding acquisition, N.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was conducted with the support of the Korea Research Foundation and was funded by the Ministry of Education of Korea in 2020 (No. 2020R1A6A1A03040583) and the MSIT, Korea, under the National Program for Excellence in SW, supervised by the IITP (2021-0-01393).

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Goldberg, D.; Nichols, D.; Oki, B.M.; Terry, D. Using Collaborative Filtering to Weave an Information Tapestry. Commun. ACM 1992, 35, 61–70. [Google Scholar] [CrossRef]
  2. Resnick, P.; Iacovou, N.; Suchak, M.; Bergstrom, P.; Riedl, J. GroupLens: An open architecture for collaborative filtering of netnews. In Proceedings of the 1994 ACM Conference on Computer Supported Cooperative Work, Chapel Hill, NC, USA, 22–26 October 1994; pp. 175–186. [Google Scholar]
  3. Herlocker, J.L.; Konstan, J.A.; Borchers, A.; Riedl, J. An Algorithmic Framework for Performing Collaborative Filtering. In Proceedings of the 22nd Annual iIternational ACM SIGIR Conference on Research and Development in Information Retrieval, Berkley, CA, USA, 15–19 August 1999; pp. 230–237. [Google Scholar]
  4. Goyani, M.; Chaurasiya, N. A review of movie recommendation system: Limitations, Survey and Challenges. ELCVIA Electron. Lett. Comput. Vis. Image Anal. 2020, 19, 18–37. [Google Scholar]
  5. Wang, H.; Le, Z.; Gong, X. Recommendation System Based on Heterogeneous Feature: A Survey. IEEE Access 2020, 8, 170779–170793. [Google Scholar] [CrossRef]
  6. Ying, R.; He, R.; Chen, K.; Eksombatchai, P.; Hamilton, W.L.; Leskovec, J. Graph Convolutional Neural Networks for Web-Scale Recommender Systems. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, London, UK, 19–23 August 2018; pp. 974–983. [Google Scholar]
  7. Sarwar, B.; Karypis, G.; Konstan, J.; Riedl, J. Item-Based Collaborative Filtering Recommendation Algorithms. In Proceedings of the WWW, Hong Kong, China, 1–5 May 2001; pp. 285–295. [Google Scholar]
  8. Rajchakit, G.; Agarwal, P.; Ramalingam, S. Stability Analysis of Neural Networks; Springer: Singapore, 2021. [Google Scholar]
  9. Roy, D.; Dutta, M. A systematic review and research perspective on recommender systems. Big Data 2022, 9, 59. [Google Scholar] [CrossRef]
  10. Blanco-Fernandez, Y.; Pazos-Arias, J.J.; Gil-Solla, A.; Ramos-Cabrer, M.; Lopez-Nores, M. Providing entertainment by content-based filtering and semantic reasoning in intelligent recommender systems. IEEE Trans. Consum. Electron. 2008, 54, 727–735. [Google Scholar] [CrossRef]
  11. Isinkaye, F.O.; Folajimi, Y.O.; Ojokoh, B.A. Recommendation systems: Principles, methods and evaluation. Egypt. Inf. J. 2015, 16, 261–273. [Google Scholar] [CrossRef]
  12. Li, C.; Wang, Z.; Cao, S.; He, L. WLRRS: A new recommendation system based on weighted linear regression models. Comput. Electr. Eng. 2018, 66, 40–47. [Google Scholar] [CrossRef]
  13. He, M.; Wang, B.; Du, X. HI2Rec: Exploring knowledge in heterogeneous information for movie recommendation. IEEE Access 2019, 7, 30276–30284. [Google Scholar] [CrossRef]
  14. Afolabi, A.O.; Toivanen, P. Integration of recommendation systems into connected health for effective management of chronic diseases. IEEE Access 2019, 7, 49201–49211. [Google Scholar] [CrossRef]
  15. Kang, S.; Jeong, C.; Chung, K. Tree-based real-time advertisement recommendation system in online broadcasting. IEEE Access 2020, 8, 192693–192702. [Google Scholar] [CrossRef]
  16. Wei, J.; He, J.; Chen, K.; Zhou, Y.; Tang, Z. Collaborative filtering and deep learning based recommendation system for cold start items. Expert Syst. Appl. 2017, 69, 29–39. [Google Scholar] [CrossRef]
  17. Li, H.; Cui, J.; Shen, B.; Ma, J. An intelligent movie recommendation system through group-level sentiment analysis in microblogs. Neurocomputing 2016, 210, 164–173. [Google Scholar] [CrossRef]
  18. Chen, M.H.; Teng, C.H.; Chang, P.C. Applying artificial immune systems to collaborative filtering for movie recommendation. Adv. Eng. Inform. 2015, 29, 830–839. [Google Scholar] [CrossRef]
  19. Koren, Y.; Bell, R.; Volinsky, C. Matrix Factorization Techniques for Recommender Systems. Computer 2009, 42, 30–37. [Google Scholar] [CrossRef]
  20. Zhang, S.; Tong, H.; Xu, J.; Maciejewski, R. Graph convolutional networks: A comprehensive review. Comput. Soc. Netw. 2019, 6, 11. [Google Scholar] [CrossRef] [PubMed]
  21. You, J.; Leskovec, J.; He, K.; Xie, S. Graph Structure of Neural Networks. In Proceedings of the 37th International Conference on Machine Learning (ICML), Vienna, Austria, 12–18 July 2020; pp. 10881–10891. [Google Scholar]
  22. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Yu, P.S. A Comprehensive Survey on Graph Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 4–24. [Google Scholar] [CrossRef] [PubMed]
  23. Wang, X.; He, X.; Wang, M.; Feng, F.; Chua, T.S. Neural Graph Collaborative Filtering. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France, 21–25 July 2019; pp. 165–174. [Google Scholar]
  24. He, X.; Deng, K.; Wang, X.; Li, Y.; Zhang, Y.; Wang, M. LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Xi’an, China, 25–30 July 2020; pp. 639–648. [Google Scholar]
  25. Yu, J.; Yin, H.; Xia, X.; Chen, T.; Li, J.; Huang, Z. Self-Supervised Learning for Recommender Systems: A Survey. IEEE Trans. Knowl. Data Eng. 2022, 36, 335–355. Available online: https://github.com/Coder-Yu/SELFRec (accessed on 7 February 2024). [CrossRef]
  26. Harper, F.M.; Konstan, J.A. The MovieLens Datasets: History and Context. ACM Trans. Interact. Intell. Syst. 2015, 5, 19. [Google Scholar] [CrossRef]
  27. Guo, G.; Zhang, J.; Thalmann, D.; Yorke-Smith, N. ETAF: An Extended Trust Antecedents Framework for Trust Prediction. In Proceedings of the 2014 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Beijing, China, 17–20 August 2014; pp. 540–547. [Google Scholar]
  28. Ma, H.; Zhou, D.; Liu, C.; Lyu, M.R.; King, I. Recommender systems with social regularization. In Proceedings of the Fourth ACM International Conference on Web Search and Data Mining, Hong Kong, China, 9–12 February 2011; pp. 287–296. [Google Scholar]
  29. Huang, T.; Dong, Y.; Ding, M.; Yang, Z.; Feng, W.; Wang, X.; Tang, J. MixGCF: An Improved Training Method for Graph Neural Network-based Recommender Systems. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Singapore, 14–18 August 2021. [Google Scholar]
  30. Wu, J.; Wang, X.; Feng, F.; He, X.; Chen, L.; Lian, J.; Xie, X. Self-supervised Graph Learning for Recommendation. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Montréal, QC, Canada, 11–15 July 2021; pp. 726–735. [Google Scholar]
  31. Yu, J.; Yin, H.; Xia, X.; Chen, T.; Cui, L.; Nguyen, Q.V.H. Are Graph Augmentations Necessary? Simple Graph Contrastive Learning for Recommendation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, 11–15 July 2022; pp. 1294–1303. [Google Scholar]
  32. Yu, J.; Xia, X.; Chen, T.; Cui, L.; Hung, N.Q.V.; Yin, H. XSimGCL: Towards Extremely Simple Graph Contrastive Learning for Recommendation. IEEE Trans. Knowl. Data Eng. 2023, 36, 913–926. [Google Scholar] [CrossRef]
  33. Lee, D.; Kang, S.; Ju, H.; Park, C.; Yu, H. Bootstrapping user and item Representations for One-Class Collaborative Filtering. In Proceedings of the SIGIR, Montréal, QC, Canada, 11–15 July 2021. [Google Scholar]
Figure 1. (a) Bipartite graph representing user–item interactions. (b) Higher-order connectivity based on u1.
Figure 1. (a) Bipartite graph representing user–item interactions. (b) Higher-order connectivity based on u1.
Electronics 13 02282 g001
Figure 2. LightGCN’s basic learning structure.
Figure 2. LightGCN’s basic learning structure.
Electronics 13 02282 g002
Figure 3. Non-combination structure of LightGCN that does not perform weight sum using the embedding enhancement method.
Figure 3. Non-combination structure of LightGCN that does not perform weight sum using the embedding enhancement method.
Electronics 13 02282 g003
Figure 4. Visualization of each dataset and the embedding results of the base model and proposed method: (a) Yelp2018 dataset; (b) MovieLens-1M dataset; (c) FilmTrust dataset; (d) Douban-book dataset; (e) LightGCN_N at Yelp2018 dataset; (f) LightGCN_N at MovieLesn-1M dataset; (g) LightGCN_N at FilmTrust dataset; (h) LightGCN_N at Douban-book dataset. (eh) show visualizations based on the best performance of the applied methods for each dataset.
Figure 4. Visualization of each dataset and the embedding results of the base model and proposed method: (a) Yelp2018 dataset; (b) MovieLens-1M dataset; (c) FilmTrust dataset; (d) Douban-book dataset; (e) LightGCN_N at Yelp2018 dataset; (f) LightGCN_N at MovieLesn-1M dataset; (g) LightGCN_N at FilmTrust dataset; (h) LightGCN_N at Douban-book dataset. (eh) show visualizations based on the best performance of the applied methods for each dataset.
Electronics 13 02282 g004aElectronics 13 02282 g004b
Table 1. Datasets used in the experiment.
Table 1. Datasets used in the experiment.
Yelp2018MovieLens-1M [26]
TrainTestTrainTest
User149981212,63810,882
Item203334022,22219,075
Interaction33,7501747478,730119,690
Rating Range0.5~40.5~44~54~5
Sparsity98.89%99.37%99.83%99.94%
FilmTrust [27]Douban-book [28]
TrainTestTrainTest
User6038598931,66831,668
Item3492319038,04836,073
Interaction460,359114,9221,237,259324,147
Rating Range4∼54∼5NoneNone
Sparsity97.82%99.40%99.90%99.97%
Table 2. Experimental environment and parameters.
Table 2. Experimental environment and parameters.
ResourceProperty
OSUbuntu 18.04.06 LTS
Nvidia Driver525.78.01
CPUIntel i9-12900K
GPUGeForce RTX 3090
RAMSamsung DDR4 16 GB (4EA)
CUDA12.0
CuDNN11.6
Python3.7.11
Pytorch1.10.0
OptimizerAdam
Learning Epoch300
Batch size2048
Embedding size64
Learning Rate0.001
Reg lambda0.0001
Layer3
Table 3. Statistical analysis values of the existing LightGCN and the proposed method.
Table 3. Statistical analysis values of the existing LightGCN and the proposed method.
FilmTrustYelp2018
Existing
Average Value
Proposed Method Average ValueExisting
Average Value
Proposed Method Average Value
User
Embedding
Mean−5.57 × 10−3−3.11 × 100−1.67 × 10−2−5.19 × 10−2
Standard Deviation4.94 × 10−12.60 × 1016.78 × 10−14.78 × 10−1
Standard error3.47 × 10−44.86 × 10−27.54 × 10−45.32 × 10−4
Variance2.44 × 10−11.27 × 1024.59 × 10−12.29 × 10−1
Midian−8.12 × 10−3−3.86 × 100−1.90 × 10−2−2.62 × 10−2
Item
Embedding
Mean−2.73 × 10−4−9.37 × 10−1−1.35 × 10−2−3.83 × 10−2
Standard Deviation4.17 × 10−11.20 × 1015.50 × 10−13.82 × 10−1
Standard error2.67 × 10−43.31 × 10−24.62 × 10−43.20 × 10−4
Variance1.74 × 10−11.33 × 1023.03 × 10−11.46 × 10−1
Midian9.80 × 10−4−5.17 × 10−1−1.29 × 10−2−2.52 × 10−2
MovieLens-1MDouban-book
Existing
average value
Proposed method
average value
Existing
average value
Proposed method
average value
User
Embedding
Mean6.55 × 10−3−6.70 × 10−34.93 × 10−3−6.27 × 10−3
Standard Deviation3.23 × 10−13.49 × 10−13.72 × 10−13.06 × 10−1
Standard error3.59 × 10−43.88 × 10−45.98 × 10−44.93 × 10−4
Variance1.04 × 10−11.22 × 10−11.38 × 10−19.39 × 10−2
Midian6.83 × 10−3−6.99 × 10−36.57 × 10−3−1.03 × 10−2
Item
Embedding
Mean−7.72 × 10−1.96 × 10−33.82 × 10−3−2.51 × 10−2
Standard Deviation2.40 × 10−12.64 × 10−14.78 × 10−14.60 × 10−1
Standard error2.02 × 10−42.21 × 10−41.01 × 10−39.74 × 10−4
Variance5.77 × 10−26.96 × 10−22.28 × 10−12.12 × 10−1
Midian−2.03 × 10−3−1.96 × 10−34.15 × 10−3−1.95 × 10−2
Table 4. Quantitative results of the entire experiment with both the traditional model and the proposed method per dataset.
Table 4. Quantitative results of the entire experiment with both the traditional model and the proposed method per dataset.
DatasetWeightBest EpochModelPrecisionRecallNDCGHit Ratio
Yelp2018-101MF0.022350.049980.040550.04367
-231LightGCN0.026880.059490.048870.05253
8, 8, 856LightGCN_W0.028090.061910.050860.05488
8, 8, 8271LightGCN_N
(Propose method)
0.030930.068730.056560.06044
MovieLens-1M-46MF0.1890.2490.28310.197
-261LightGCN0.199870.271960.302740.20838
64, 64, 6411LightGCN_W0.199370.277160.308520.20786
128,128,128271LightGCN_N0.202140.277480.30830.21074
FilmTrust-296MF0.08680.8480.61690.8178
-276LightGCN0.087610.853760.615660.82547
8, 8, 876LightGCN_W0.088170.861310.61970.83481
32, 32, 32226LightGCN_N0.088240.858740.624530.83137
Douban-book-66MF0.054070.12470.10390.09662
-266LightGCN0.062090.147830.125080.11095
16, 16, 1626LightGCN_W0.063150.151270.128670.11284
16, 16, 16181LightGCN_N0.071030.167010.146320.12692
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, S.; Ahn, J.; Kim, N. Embedding Enhancement Method for LightGCN in Recommendation Information Systems. Electronics 2024, 13, 2282. https://doi.org/10.3390/electronics13122282

AMA Style

Lee S, Ahn J, Kim N. Embedding Enhancement Method for LightGCN in Recommendation Information Systems. Electronics. 2024; 13(12):2282. https://doi.org/10.3390/electronics13122282

Chicago/Turabian Style

Lee, Sangmin, Junho Ahn, and Namgi Kim. 2024. "Embedding Enhancement Method for LightGCN in Recommendation Information Systems" Electronics 13, no. 12: 2282. https://doi.org/10.3390/electronics13122282

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop