Disentangled Self-Attention with Auto-Regressive Contrastive Learning for Neural Group Recommendation
Abstract
:1. Introduction
- We combine disentangled neural networks with a multi-head self-attention mechanism and propose a disentangled multi-head self-attention module to generate precise group and user representations, dissecting user intents for preference encoding from multiple dimensions;
- We introduce optimization objectives based on auto-regressive user preference information and group-related mutual information maximization to counteract information scarcity in group data, thus facilitating network training and enhancing the fidelity of representations;
- Comprehensive experiments conducted across three public benchmark datasets for group recommendations corroborate the efficacy of our framework, showcasing its superiority over state-of-the-art baselines.
2. Scientific Method
2.1. Group Recommendation
2.2. Disentangled Neural Networks
- Encoder: IGRec considers the embedding of the disentanglement factors for the same user when producing the representations of user disentangled interests. However, our framework is realized by two MLP encoders and a multi-head disentangled self-attention block which additionally considers the relationships between different user representations within the same group when producing the representations of user disentangled interests;
- Aggregator: IGRec depends on Gumbel-SoftMax among the disentangled factors to aggregate the user interests. In our framework, the multi-factor representations are jointly projected for group preference prediction by an MLP, which can automatically extract and utilize the relevance between each factor and the group representation;
- Optimization Objectives: IGRec leverages user–item and group–item historical interactions to construct supervised signals for dual training. In our work, to alleviate the data sparsity issue, we leverage auto-regressive user preferences to train the disentangled multi-head self-attention encoder and employ the mutual information between user representations and group representations.
2.3. Mutual Information Optimization
3. Problem Statement
4. The Proposed Framework
4.1. Overview
4.2. Supervised Learning
4.3. Contrastive Leaning
4.4. Framework Details
4.4.1. Training Procedure
Algorithm 1 Training procedure of DAGA. |
|
4.4.2. Model Architecture
5. Experiments
5.1. Experimental Settings
- Weeplaces captures point-of-interest (POI) interactions within major US cities, offering rich categorical data ranging from entertainment data to dining data. Groups are constructed based on simultaneous check-ins, aligning with previous methodologies [12];
- CAMRa2011, a movie rating dataset, is adapted to the group recommendation context by focusing on households as implicit groups, with ratings transformed into binary interaction data for compatibility with our framework [11];
- Douban, sourced from a versatile social platform, facilitates the exploration of group dynamics within cultural contexts. Group interactions are inferred from shared participation in activities, following a strategy similar to that in [20].
- Popular: A rudimentary yet widely applied method that recommends items based solely on their overall popularity, disregarding personalized preferences or group dynamics;
- NeuMF [57]: NeuMF represents the integration of neural network architectures into collaborative filtering, enhancing recommendation performance by directly leveraging group–item interactions as the supervisory signal. NeuMF exemplifies the advancements in neural collaborative filtering by predicting group preferences with heightened accuracy;
- AGREE [11]: AGREE innovates by incorporating an attention mechanism to deduce group preferences from individual member preferences. It computes weights for each group member and aggregates them to obtain a unified group preference vector, employing a neural collaborative filtering framework to model group–item interactions;
- MoSAN [18]: An advanced group recommendation model that utilizes a neural network structure with multiple sub-attention modules. These modules are tasked with capturing nuanced user preferences and effectively modeling the intricate dynamics of user–item and group–item interactions;
- SIGR [19]: SIGR introduces an attention-based model augmented with latent variables to capture both global and local social influences among users. It conceptualizes interactions within a bipartite graph structure, applying graph embedding techniques to mitigate issues related to data sparsity;
- GroupIM [12]: A cutting-edge group recommendation model that aggregates individual member preferences through attention networks to forecast group preferences. It distinguishes itself by optimizing the mutual information between group and member representations, addressing the challenge of sparse data;
- HHGR [20]: HHGR utilizes a hypergraph neural network to encapsulate the complex high-order relationships among users, groups and items. It is particularly noted for its innovative double-scale self-supervised learning objective, designed to complement traditional supervised learning signals and enhance the training process;
- HHGR [20]: An extension of the HHGR model, this self-supervised variant represents the state of the art in group recommendation. It leverages self-supervised learning techniques to further refine the model’s ability to discern and predict group preferences;
- SGGCF [22]: SGGCF captures the high-order interactions between users, items and groups by modeling with a user-centered heterogeneous graph and a self-supervised learning framework. SGGCF realizes a state-of-the-art performance for ephemeral group recommendation.
- NDCG@K is a metric that measures the quality of the ranked recommendation list up to the K-th position, taking into account the position of relevant items. It is particularly well suited for evaluating recommendation systems where the order of recommendations is of importance. The computation of NDCG@K involves the gain of each item (with more weight given to items at higher ranks) and a normalization factor to ensure that the perfect ranking has a score of 1. This makes NDCG@K an effective measure for understanding how well a model can rank truly relevant items higher in the recommendation list;
- Recall@K, on the other hand, assesses the model’s ability to retrieve relevant items within the top K recommendations, regardless of their order. It is calculated as the fraction of relevant items that are successfully included in the top K recommended items out of all relevant items. Recall@K is particularly useful in scenarios where the goal is to capture as many relevant items as possible in the top K recommendations without an emphasis on the ranking order among these items.
- Learning Rate: Set at , the learning rate determines the step size during the gradient descent optimization process. This value is chosen to balance the trade-off between convergence speed and stability;
- Weight Decay: A regularization parameter employed to prevent overfitting by penalizing large weights. For our experiments, weight decay is set to 0, emphasizing model flexibility in learning from the data;
- Dropout Ratio: Applied within the MLPs to mitigate overfitting through the random omission of neurons during training. A dropout ratio of 0.4 is found to provide an optimal balance between regularization and model complexity;
- Batch Size: The number of samples processed before the internal parameters of the model are updated. A batch size of 256 is selected to leverage computational efficiency while maintaining sufficient gradient approximation;
- Epochs: Both the pre-training and fine-tuning phases are conducted over 100 epochs, ensuring ample opportunity for the model to learn and adapt to the training data;
- Factor Number: Our disentangled network architecture incorporates eight factors, allowing the model to capture a diverse array of latent user and group preferences;
- Embedded Size: The dimensionality of the user and group representations in the latent space is set to 192, with each factor contributing a 24-dimensional sub-representation. This configuration facilitates a comprehensive yet computationally efficient representation scheme;
- Number of Negative Samples (): In the mutual information maximization process (Section 4.3), five negative samples are used for each positive instance, balancing the learning signal derived from both positive and negative interactions;
- SSL Loss Weight (λ): The weight assigned to the SSL loss function during the fine-tuning phase is set to 1, underscoring its equal importance alongside the supervised loss components in model optimization.
5.2. Experimental Results
Comparative Analysis of Performance
5.3. Analysis of Parameter Complexity
5.4. Ablation Studies and Parameter Studies
5.4.1. Ablation Studies
5.4.2. Parameter Studies
6. Conclusions
- Theoretical Contributions: The innovative disentanglement approach allows for the independent transfer of user and group representations across distinct factors in the self-attention mechanism and MLP encoders, effectively mitigating the amalgamation of unrelated factor information. To facilitate the training of this framework, we formulate both supervised and self-supervised loss functions. These functions capitalize on the preference probability distributions at the input and output stages of DAGA, as well as on the contextual insights gleaned through the maximization of mutual information pertaining to group representations;
- Applied Contributions: Comprehensive empirical analyses conducted on three publicly available benchmark datasets substantiate the superior performance of DAGA in comparison to the state-of-the-art methodologies. Given the demonstrated efficacy of collaborative filtering (CF) in recommendation systems, our future endeavors will focus on the exploration of group recommendation models predicated on CF techniques and the development of more nuanced objective functions to further enhance model performance;
- Main Limitations: Additionally, there are also some limitations to our proposed framework. For example, DAGA performs best when the training flow follows the training procedure as shown in Algorithm 1. However, when it comes to an end-to-end training procedure, there will be some performance degradation. In our future work, we will explore the more efficient end-to-end training framework for group recommendation and also want to discover more applicable scenarios for the multi-head disentangled self-attention encoder, including a variety of recommended and non-recommended tasks.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Zheng, K.; Zhang, J.; Liu, X.; Fu, L.; Wang, X.; Jiang, X.; Zhang, W. Secrecy Capacity Scaling of Large-Scale Networks With Social Relationships. IEEE Trans. Veh. Technol. 2017, 66, 2688–2702. [Google Scholar] [CrossRef]
- Zheng, K.; Jia, X.; Chi, K.; Liu, X. DDPG-Based Joint Time and Energy Management in Ambient Backscatter-Assisted Hybrid Underlay CRNs. IEEE Trans. Commun. 2023, 71, 441–456. [Google Scholar] [CrossRef]
- Kabbur, S.; Ning, X.; Karypis, G. FISM: Factored item similarity models for top-N recommender systems. In Proceedings of the KDD ’13: Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Chicago, IL, USA, 11–14 August 2013; pp. 659–667. [Google Scholar]
- Wang, X.; He, X.; Wang, M.; Feng, F.; Chua, T.-S. Neural Graph Collaborative Filtering. In Proceedings of the SIGIR’19: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France, 21–25 July 2019. [Google Scholar]
- Rendle, S.; Freudenthaler, C.; Gantner, Z.; Schmidt-Thieme, L. BPR: Bayesian Personalized Ranking from Implicit Feedback. arXiv 2009, arXiv:1205.2618. [Google Scholar]
- He, X.; Deng, K.; Wang, X.; Li, Y.; Zhang, Y.; Wang, M. LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation. In Proceedings of the SIGIR ’20: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 25–30 July 2020. [Google Scholar]
- He, X.; He, Z.; Song, J.; Liu, Z.; Jiang, Y.; Chua, T.-S. NAIS: Neural Attentive Item Similarity Model for Recommendation. IEEE Trans. Knowl. Data Eng. 2018, 30, 2354–2366. [Google Scholar] [CrossRef]
- Koren, Y. Factorization meets the neighborhood: A multifaceted collaborative fltering model. In Proceedings of the KDD ’08: Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, NV, USA, 24–27 August 2008; pp. 426–434. [Google Scholar]
- Liang, D.; Krishnan, R.G.; Hoffman, M.D.; Jebara, T. Variational Autoencoders for Collaborative Filtering. In Proceedings of the WWW ’18: Proceedings of the 2018 World Wide Web Conference, Lyon, France, 23–27 April 2018; pp. 689–698. [Google Scholar]
- Hu, L.; Cao, J.; Xu, G.; Cao, L.; Gu, Z.; Cao, W. Deep modeling of group preferences for group-based recommendation. Proc. AAAI Conf. Artif. Intell. 2014, 28, 1861–1867. [Google Scholar] [CrossRef]
- Cao, D.; He, X.; Miao, L.; An, Y.; Yang, C.; Hong, R. Attentive group recommendation. In Proceedings of the SIGIR ’18: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 645–654. [Google Scholar]
- Sankar, A.; Wu, Y.; Wu, Y.; Zhang, W.; Yang, H.; Sundaram, H. GroupIM: A mutual information maximization framework for neural group recommendation. In Proceedings of the SIGIR ’20: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 25–30 July 2020; pp. 1279–1288. [Google Scholar]
- Baltrunas, L.; Makcinskas, T.; Ricci, F. Group recommendations with rank aggregation and collaborative fltering. In Proceedings of the RecSys ’10: Proceedings of the Fourth ACM Conference on Recommender Systems, Barcelona, Spain, 26–30 September 2010; pp. 119–126. [Google Scholar]
- Berkovsky, S.; Freyne, J. Group-based recipe recommendations: Analysis of data aggregation strategies. In Proceedings of the RecSys ’10: Proceedings of the fourth ACM conference on Recommender systems, Barcelona, Spain, 26–30 September 2010; pp. 111–118. [Google Scholar]
- Boratto, L.; Carta, S. State-of-the-Art in Group Recommendation and New Approaches for Automatic Identifcation of Groups. In Information Retrieval and Mining in Distributed Environments; Springer: Berlin/Heidelberg, Germany, 2011; Volume 324, pp. 1–20. [Google Scholar]
- Amer-Yahia, S.; Roy, S.B.; Chawla, A.; Das, G.; Yu, C. Group Recommendation: Semantics and Efciency. Proc. VLDB Endow. 2009, 2, 754–765. [Google Scholar] [CrossRef]
- He, Z.; Chow, C.-Y.; Zhang, J.-D. GAME: Learning Graphical and Attentive Multi-view Embeddings for Occasional Group Recommendation. In Proceedings of the SIGIR ’20: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 25–30 July 2020; pp. 649–658. [Google Scholar]
- Tran, L.V.; Pham, T.-A.N.; Tay, Y.; Liu, Y.; Cong, G.; Li, X. Interact and Decide: Medley of Sub-Attention Networks for Effective Group Recommendation. In Proceedings of the SIGIR’19: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France, 21–25 July 2019; pp. 255–264. [Google Scholar]
- Yin, H.; Wang, Q.; Zheng, K.; Li, Z.; Yang, J.; Zhou, X. Social Influence-Based Group Representation Learning for Group Recommendation. In Proceedings of the 2019 IEEE 35th International Conference on Data Engineering (ICDE), Macao, China, 8–11 April 2019; pp. 566–577. [Google Scholar]
- Zhang, J.; Gao, M.; Yu, J.; Guo, L.; Li, J.; Yin, H. Double-Scale Self-Supervised Hypergraph Learning for Group Recommendation. In Proceedings of the CIKM ’21: Proceedings of the 30th ACM International Conference on Information & Knowledge Management, Virtual Event, 1–5 November 2021; pp. 2557–2567. [Google Scholar]
- Chen, T.; Yin, H.; Long, J.; Nguyen, Q.V.H.; Wang, Y.; Wang, M. Thinking inside The Box: Learning Hypercube Representations for Group Recommendation. In Proceedings of the SIGIR ’22: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, 11–15 July 2022; pp. 1664–1673. [Google Scholar]
- Li, K.; Wang, C.-D.; Lai, J.; Yuan, H. Self-Supervised Group Graph Collaborative Filtering for Group Recommendation. In Proceedings of the WSDM ’23: Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining, Singapore, 27 February–3 March 2023; pp. 69–77. [Google Scholar]
- Quintarelli, E.; Rabosio, E.; Tanca, L. Recommending new items to ephemeral groups using contextual user influence. In Proceedings of the RecSys ’16: Proceedings of the 10th ACM Conference on Recommender Systems, Boston, MA, USA, 15–19 September 2016; pp. 285–292. [Google Scholar]
- Seko, S.; Yagi, T.; Motegi, M.; Muto, S. Group recommendation using feature space representing behavioral tendency and power balance among members. In Proceedings of the RecSys ’11: Proceedings of the Fifth ACM Conference on Recommender Systems, Chicago, IL, USA, 23–27 October 2011; pp. 101–108. [Google Scholar]
- Yin, H.; Wang, Q.; Zheng, K.; Li, Z.; Zhou, X. Overcoming data sparsity in group recommendation. IEEE Trans. Knowl. Data Eng. 2020, 34, 3447–3460. [Google Scholar] [CrossRef]
- Guo, L.; Yin, H.; Wang, Q.; Cui, B.; Huang, Z.; Cui, L. Group recommendation with latent voting mechanism. In Proceedings of the 2020 IEEE 36th International Conference on Data Engineering (ICDE), Dallas, TX, USA, 20–24 April 2020; pp. 121–132. [Google Scholar]
- Jia, R.; Zhou, X.; Dong, L.; Pan, S. Hypergraph Convolutional Network for Group Recommendation. In Proceedings of the 2021 IEEE International Conference on Data Mining (ICDM), Auckland, New Zealand, 7–10 December 2021; pp. 260–269. [Google Scholar]
- Gorla, J.; Lathia, N.; Robertson, S.; Wang, J. Probabilistic group recommendation via information matching. In Proceedings of the WWW ’13: Proceedings of the 22nd international conference on World Wide Web, Rio de Janeiro, Brazil, 13–17 May 2013; pp. 495–504. [Google Scholar]
- Liu, X.; Tian, Y.; Ye, M.; Lee, W.-C. Exploring personal impact for group recommendation. In Proceedings of the CIKM ’12: Proceedings of the 21st ACM International Conference on Information and Knowledge Management, Maui, HI, USA, 29 October–2 November 2012; pp. 674–683. [Google Scholar]
- Yuan, Q.; Cong, G.; Lin, C.-Y. COM: A generative model for group recommendation. In Proceedings of the KDD ’14: Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 163–172. [Google Scholar]
- Huang, Z.; Xu, X.; Zhu, H.; Zhou, M. An efcient group recommendation model with multiattention-based neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 4461–4474. [Google Scholar] [CrossRef]
- Ghaemmaghami, S.S.; Salehi-Abari, A. DeepGroup: Group Recommendation with Implicit Feedback. In Proceedings of the CIKM ’21: Proceedings of the 30th ACM International Conference on Information & Knowledge Management, Virtual Event, 1–5 November 2021; pp. 3408–3412. [Google Scholar]
- Cao, D.; He, X.; Miao, L.; Xiao, G.; Chen, H.; Xu, J. Social-Enhanced Attentive Group Recommendation. IEEE Trans. Knowl. Data Eng. 2021, 33, 1195–1209. [Google Scholar] [CrossRef]
- Deng, Z.; Li, C.; Liu, S.; Ali, W.; Shao, J. KnowledgeAware Group Representation Learning for Group Recommendation. In Proceedings of the 2021 IEEE 37th International Conference on Data Engineering (ICDE), Chania, Greece, 19–22 April 2021; pp. 1571–1582. [Google Scholar]
- Yang, M.; Liu, Z.; Yang, L.; Liu, X.; Wang, C.; Peng, H.; Yu, P.S. Ranking-based Group Identification via Factorized Attention on Social Tripartite Graph. In Proceedings of the WSDM ’23: Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining, Singapore, 27 February–3 March 2023; pp. 769–777. [Google Scholar]
- Yang, M.; Liu, Z.; Yang, L.; Liu, X.; Wang, C.; Peng, H.; Yu, P.S. Group Identification via Transitional Hypergraph Convolution with Cross-view Self-supervised Learning. In Proceedings of the CIKM ’23: Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, Birmingham, UK, 21–25 October 2023; pp. 2969–2979. [Google Scholar]
- Vaswani; Ashish; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
- Wu, J.; Wang, X.; Feng, F.; He, X.; Chen, L.; Lian, J.; Xie, X. Self-supervised graph learning for recommendation. In Proceedings of the SIGIR ’21: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 11–15 July 2021; pp. 726–735. [Google Scholar]
- Yu, J.; Yin, H.; Xia, X.; Chen, T.; Li, J.; Huang, Z. Self-Supervised Learning for Recommender Systems: A Survey. arXiv 2022, arXiv:2203.15876. [Google Scholar] [CrossRef]
- Zhou, Y.; Dou, Z.; Zhu, Y.; Wen, J.-R. PSSL: Selfsupervised Learning for Personalized Search with Contrastive Sampling. In Proceedings of the CIKM ’21: Proceedings of the 30th ACM International Conference on Information & Knowledge Management, Virtual Event, 1–5 November 2021; pp. 2749–2758. [Google Scholar]
- Bengio, Y.; Courville, A.; Vincent, P. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef] [PubMed]
- Hsieh, J.T.; Liu, B.; Huang, D.A.; Fei-Fei, L.F.; Niebles, J.C. Learning to decompose and disentangle representations for video prediction. Adv. Neural Inf. Process. Syst. 2018, 31, 517–526. [Google Scholar]
- Chen, X.; Duan, Y.; Houthooft, R.; Schulman, J.; Sutskever, I.; Abbeel, P. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. Adv. Neural Inf. Process. Syst. 2016, 29, 2180–2188. [Google Scholar]
- Ma, L.; Sun, Q.; Georgoulis, S.; Gool, L.V.; Schiele, B.; Fritz, M. Disentangled person image generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 99–108. [Google Scholar]
- Ma, J.; Cui, P.; Kuang, K.; Wang, X.; Zhu, W. Disentangled graph convolutional networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 4212–4221. [Google Scholar]
- Li, H.; Wang, X.; Zhang, Z.; Yuan, Z.; Li, H.; Zhu, W. Disentangled Contrastive Learning on Graphs. Adv. Neural Inf. Process. Syst. 2021, 34, 21872–21884. [Google Scholar]
- Higgins, I.; Matthey, L.; Pal, A.; Burgess, C.; Glorot, X.; Botvinick, M.; Mohamed, S.; Lerchner, A. beta-vae: Learning basic visual concepts with a constrained variational framework. In Proceedings of the 5th International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
- Burgess, C.P.; Higgins, I.; Pal, A.; Matthey, L.; Watters, N.; Desjardins, G.; Lerchner, A. Understanding disentangling in β-vae. In Proceedings of the NeurIPS Workshop on Learning Disentangled Representations, Long Beach, CA, USA, 9 December 2017. [Google Scholar]
- Liu, X.; Yang, L.; Liu, Z.; Li, X.; Yang, M.; Wang, C.; Yu, P.S. Group-Aware Interest Disentangled Dual-Training for Personalized Recommendation. In Proceedings of the 2023 IEEE International Conference on Big Data (BigData), Sorrento, Italy, 15–18 December 2023; pp. 393–402. [Google Scholar]
- Oord, A.v.; Li, Y.; Vinyals, O. Representation learning with contrastive predictive coding. arXiv 2018, arXiv:1807.03748. [Google Scholar]
- Hjelm, R.D.; Fedorov, A.; Lavoie-Marchildon, S.; Grewal, K.; Bachman, P.; Trischler, A.; Bengio, Y. Learning deep representations by mutual information estimation and maximization. arXiv 2018, arXiv:1808.06670. [Google Scholar]
- Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the 37th International Conference on Machine Learning, Virtual Event, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
- Veličković, P.; Fedus, W.; Hamilton, W.L.; Liò, P.; Bengio, Y.; Hjelm, R.D. Deep graph infomax. arXiv 2018, arXiv:1809.10341. [Google Scholar]
- Hassani, K.; Khasahmadi, A.H. Contrastive multi-view representation learning on graphs. In Proceedings of the 37th International Conference on Machine Learning, Virtual Event, 13–18 July 2020; pp. 4116–4126. [Google Scholar]
- Peng, Z.; Huang, W.; Luo, M.; Zheng, Q.; Rong, Y.; Xu, T.; Huang, J. Graph representation learning via graphical mutual information maximization. In Proceedings of the WWW ’20: Proceedings of The Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; pp. 259–270. [Google Scholar]
- You, Y.; Chen, T.; Sui, Y.; Chen, T.; Wang, Z.; Shen, Y. Graph contrastive learning with augmentations. Adv. Neural Inf. Process. Syst. 2020, 33, 5812–5823. [Google Scholar]
- He, X.; Liao, L.; Zhang, H.; Nie, L.; Hu, X.; Chua, T.-S. Neural Collaborative Filtering. In Proceedings of the WWW ’17: Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 April 2017; pp. 173–182. [Google Scholar]
Notation | Meaning | Notation | Meaning |
---|---|---|---|
Set of users | User i | ||
Set of items | Item j | ||
Set of groups | Group i | ||
R | User–item interaction matrix | Q | Group–item interaction matrix |
Z | User embedding matrix | Embedding matrix of factor k for users | |
P | Group embedding matrix | Embedding matrix of factor k for groups | |
K | Number of factors | Y | Predicted preference distribution matrix |
Dataset | #Users (#U) | #Items (#I) | #Groups (#G) | #U–I Interact | #G–I Interact |
---|---|---|---|---|---|
Weeplaces | 8643 | 25,081 | 22,733 | 1,358,458 | 180,229 |
CAMRa2011 | 602 | 7710 | 290 | 116,344 | 145,068 |
Douban | 2964 | 39,694 | 2630 | 823,851 | 463,040 |
Dataset | Weeplaces | CAMRa2011 | Douban | |||
---|---|---|---|---|---|---|
Metric | N@20 | N@50 | N@20 | N@50 | N@20 | N@50 |
Baseline Recommender | ||||||
Popular | 0.063 | 0.074 | 0.099 | 0.122 | 0.003 | 0.005 |
NeuMF | 0.193 | 0.244 | 0.305 | 0.367 | 0.124 | 0.167 |
AGREE | 0.224 | 0.267 | 0.307 | 0.418 | 0.201 | 0.224 |
MoSAN | 0.287 | 0.334 | 0.423 | 0.466 | 0.163 | 0.209 |
SIGR | 0.357 | 0.391 | 0.499 | 0.524 | 0.217 | 0.235 |
GroupIM | 0.431 | 0.456 | 0.637 | 0.659 | 0.257 | 0.284 |
HHGR | 0.379 | 0.398 | 0.517 | 0.532 | 0.254 | 0.267 |
HHGR | 0.456 | 0.478 | 0.645 | 0.671 | 0.279 | 0.294 |
SGGCF | 0.475 | 0.500 | 0.653 | 0.689 | 0.280 | 0.309 |
Our Recommender | ||||||
DAGA | 0.512 | 0.576 | 0.672 | 0.707 | 0.290 | 0.313 |
Dataset | Weeplaces | CAMRa2011 | Douban | |||
---|---|---|---|---|---|---|
Metric | R@20 | R@50 | R@20 | R@50 | R@20 | R@50 |
Baseline Recommender | ||||||
Popular | 0.126 | 0.176 | 0.172 | 0.226 | 0.009 | 0.018 |
NeuMF | 0.271 | 0.295 | 0.393 | 0.450 | 0.248 | 0.316 |
AGREE | 0.354 | 0.671 | 0.529 | 0.688 | 0.297 | 0.488 |
MoSAN | 0.548 | 0.738 | 0.572 | 0.801 | 0.384 | 0.459 |
SIGR | 0.524 | 0.756 | 0.585 | 0.825 | 0.436 | 0.560 |
GroupIM | 0.575 | 0.773 | 0.753 | 0.874 | 0.523 | 0.696 |
HHGR | 0.554 | 0.764 | 0.703 | 0.830 | 0.507 | 0.677 |
HHGR | 0.592 | 0.797 | 0.779 | 0.883 | 0.561 | 0.741 |
SGGCF | 0.689 | 0.794 | 0.789 | 0.891 | 0.573 | 0.744 |
Our Recommender | ||||||
DAGA | 0.723 | 0.814 | 0.801 | 0.905 | 0.608 | 0.771 |
Settings | Model | N@20 | N@50 | R@20 | R@50 |
---|---|---|---|---|---|
Aggregator | |||||
2-Layer with 4-Head | DAGA-Max | 0.486 | 0.514 | 0.699 | 0.791 |
DAGA-Mean | 0.498 | 0.558 | 0.715 | 0.808 | |
DAGA-Attention | 0.512 | 0.576 | 0.723 | 0.814 | |
Multi-Head Disentangled Self-Attention Encoder | |||||
4-Head with Att. Aggre. | 1-Layer | 0.493 | 0.539 | 0.713 | 0.802 |
2-Layer | 0.512 | 0.576 | 0.723 | 0.814 | |
3-Layer | 0.510 | 0.565 | 0.720 | 0.809 | |
2-Layer with Att. Aggre. | 2-Head | 0.505 | 0.548 | 0.715 | 0.805 |
4-Head | 0.512 | 0.576 | 0.723 | 0.814 | |
8-Head | 0.508 | 0.570 | 0.718 | 0.810 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Gao, L.; Zhang, H.; Fu, L. Disentangled Self-Attention with Auto-Regressive Contrastive Learning for Neural Group Recommendation. Appl. Sci. 2024, 14, 4155. https://doi.org/10.3390/app14104155
Gao L, Zhang H, Fu L. Disentangled Self-Attention with Auto-Regressive Contrastive Learning for Neural Group Recommendation. Applied Sciences. 2024; 14(10):4155. https://doi.org/10.3390/app14104155
Chicago/Turabian StyleGao, Linyao, Haonan Zhang, and Luoyi Fu. 2024. "Disentangled Self-Attention with Auto-Regressive Contrastive Learning for Neural Group Recommendation" Applied Sciences 14, no. 10: 4155. https://doi.org/10.3390/app14104155
APA StyleGao, L., Zhang, H., & Fu, L. (2024). Disentangled Self-Attention with Auto-Regressive Contrastive Learning for Neural Group Recommendation. Applied Sciences, 14(10), 4155. https://doi.org/10.3390/app14104155