Next Article in Journal
Literacy in Artificial Intelligence as a Challenge for Teaching in Higher Education: A Case Study at Portalegre Polytechnic University
Previous Article in Journal
Explainable Machine Learning Method for Aesthetic Prediction of Doors and Home Designs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

IUAutoTimeSVD++: A Hybrid Temporal Recommender System Integrating Item and User Features Using a Contractive Autoencoder †

1
LAVETE Laboratory, FST, Hassan First University of Settat, Settat 26000, Morocco
2
LAVETE Laboratory, ENSA, Hassan First University of Settat, Berrechid 26100, Morocco
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in Azri, A.; Haddi, A.; Allali, H. autoTimeSVD++: A Temporal Hybrid Recommender System Based on Contractive Autoencoder and Matrix Factorization. In Proceedings of the SADASC 2022, Marrakech, Morocco, 22–24 September 2022.
Information 2024, 15(4), 204; https://doi.org/10.3390/info15040204
Submission received: 8 March 2024 / Revised: 31 March 2024 / Accepted: 2 April 2024 / Published: 5 April 2024
(This article belongs to the Section Information Systems)

Abstract

:
Collaborative filtering (CF), a fundamental technique in personalized Recommender Systems, operates by leveraging user–item preference interactions. Matrix factorization remains one of the most prevalent CF-based methods. However, recent advancements in deep learning have spurred the development of hybrid models, which extend matrix factorization, particularly with autoencoders, to capture nonlinear item relationships. Despite these advancements, many proposed models often neglect dynamic changes in the rating process and overlook user features. This paper introduces IUAutoTimeSVD++, a novel hybrid model that builds upon autoTimeSVD++. By incorporating item–user features into the timeSVD++ framework, the proposed model aims to address the static nature and sparsity issues inherent in existing models. Our model utilizes a contractive autoencoder (CAE) to enhance the capacity to capture a robust and stable representation of user-specific and item-specific features, accommodating temporal variations in user preferences and leveraging item characteristics. Experimental results on two public datasets demonstrate IUAutoTimeSVD++’s superiority over baseline models, affirming its effectiveness in capturing and utilizing user and item features for temporally adaptive recommendations.

1. Introduction

In the era of big data, Recommender Systems (RS) [1] have become crucial for personalized recommendations and an indispensable part of our daily life. These systems facilitate the user’s experience by helping them to make decisions based on their historical preferences, such as ratings, clicks, saved carts, wishlists, and order history. Collaborative filtering (CF) [2], which relies on user preferences, remains a highly successful recommendation approach. Among collaborative filtering methods, matrix factorization, a latent factor model, gained popularity after its success in the Netflix prize challenge [3]. It continues to yield competitive results when compared with state-of-the-art deep learning-based models, as proved recently in [4,5]. Matrix factorization approximates the rating matrix R using two lower-rank matrices, P and Q , where k denotes the latent factor dimension. Despite the success of collaborative filtering, it struggles with sparsity (a lot of items without ratings) and cold-start (recommendations for a new user, for example) issues, particularly when recommending items to new users or handling new items. As the field of recommender systems advances, to address this limitation researchers have begun incorporating modern techniques such as deep learning (DL) [6,7] to propose innovative models. Other approaches adopt a hybrid strategy such as in [8,9,10], which combine latent factor models such as matrix factorization and deep learning techniques such as autoencoders or CNNs to leverage the benefits of both such as dimensionality reduction, efficient representation learning, and modeling of complex features. In this paper, our primary contribution lies in extending our previous model, autoTimeSVD++ [11], by incorporating user features. The proposed model integrates both user and item features, along with a temporal aspect based on timeSVD++ [12]. Our novel model, named IUAutoTimeSVD++, represents a temporal hybrid model that seamlessly integrates two techniques: temporal dynamics matrix factorization (timeSVD++) and the contractive autoencoder (CAE) [13]. Leveraging the effectiveness of CAE in feature extraction, it synergizes with temporal dynamics matrix factorization to enhance the model’s performance and recommendation accuracy. In light of this hybrid model, the present paper addresses the following research questions:
  • How can integrate user–item features within the timeSVD++ model to enhance the accuracy of rating predictions?
  • How to compare the performance of the IUAutoTimeSVD++ model with the baseline models, assessing its superiority and effectiveness?
  • How can we quantify the impact and significance of each component in our model to understand their individual contributions to improving recommendation predictions?
  • What are the limitations of the proposed model to provide insights into its potential areas for improvement?
By answering these research questions, our objective is to conduct a thorough evaluation of the IUAutoTimeSVD++ model, providing a detailed examination of its strengths, limitations, and possible future directions. The paper’s structure is organized as follows: The next section presents and discusses related works in the field of recommender systems by considering two main collaborative filtering methods: MF and PMF and its extensions. Section 3 is dedicated to presenting the architecture, prediction formula, and optimization process of the IUAutoTimeSVD++ model. In Section 4, we delve into the experimental results and present a comprehensive discussion of the findings. Finally, we summarize the main findings, underscore the contributions of our work in developing the IUAutoTimeSVD++ model, and suggest potential avenues for future research.

2. Related Work

In this section, our focus is on various collaborative filtering models, which include matrix factorization and its extensions, as well as contrastive learning (CL)-based models. The study is divided into three parts: the extensions of the native matrix factorization, probabilistic matrix factorization (PMF), and CL-based models as follows:

2.1. MF Extensions

Matrix factorization (MF) [14] has been a fundamental technique in recommendation systems, but researchers have continually sought to enhance its capabilities by suggesting novel extensions. Several models have been introduced to address specific challenges and incorporate additional data modalities, leading to improved recommendation accuracy and performance. One of the earliest extensions of MF is SVD++ [14] which incorporates the implicit feedback of users. Another extension of MF is timeSVD+ [12] which extends SVD++ by leveraging the time factor; this extension allows the model to extract user preferences over time. Another extension of MF is Visual Bayesian Personalized Ranking (VBPR) [15], which integrates visual information, such as images. By considering item visual features extracted using a CNN model, VBPR aims to make more informed and accurate recommendations, especially in the fashion domain, particularly when visual content plays a crucial role in item preferences. EMF (Explainable Matrix Factorization) [16] represents another notable extension of traditional matrix factorization techniques, aimed at achieving a balance between recommendation accuracy and interpretability. This model adopts a neighborhood explainability style to provide clear and intuitive explanations for its recommendations, making it more transparent and understandable to users. Another notable advancement is NeuMF (Neural Collaborative Filtering) [17]. It extends MF through a combination of two models: GMF, which is a generalized matrix factorization using pairwise product (Hadamard product), and an MLP-based model. Although NeuMF succeeds in capturing complex patterns, it suffers from computing efficiency issues due to the high cost of training the MLP branch of this model [4]. AutoSVD++ [8] introduces a hybrid approach by incorporating item features through a contractive autoencoder (CAE) into the SVD++ model. By leveraging item features, AutoSVD++ addresses the sparsity problem inherent in recommendation data, leading to improved recommendation accuracy and coverage. Similarly, SSAERec [10] extends the SVD++ [14] by adopting a hybrid strategy that involves utilizing a stacked autoencoder to extract item features. These extracted features are then incorporated into the latent factor model, which considers implicit feedback to enhance the recommendation process.
This review of related work is summarized in Table 1:

2.2. PMF Extensions

In this context, probabilistic matrix factorization (PMF) [18] stands as a foundational model. PMF is a technique used in CF models for recommendation systems. The fundamental principle behind PMF is to represent the user–item interaction matrix as a probabilistic model. The model formalized both user and item latent factors as random variables following a Gaussian distribution (normal distribution) with a zero mean. ConvMF (convolutional matrix factorization) [9] serves as an extension, introducing a convolutional neural network (CNN) to seamlessly integrate textual item reviews with item embedding. Leveraging user reviews, ConvMF enhances traditional PMF to provide more personalized and context-aware recommendations. Another notable extension is DBPMF (deep probabilistic matrix factorization) [19], which applies a CNN to extract user–item features, and the results are then averaged with PMF.
A comprehensive summary of the studies on PMF extensions is presented in Table 2:

2.3. Contrastive Collaborative Filtering

Contrastive learning (CL) has emerged as a promising approach in recommendation systems, garnering significant attention in recent years. CL involves learning representations of users and items by contrasting positive interactions with negative interactions. This methodology aims to capture underlying patterns in user–item interactions, thereby enhancing recommendation quality. In CL-based recommendation, positive interactions, such as user–item interactions indicating preference or engagement, are treated as positive samples, while negative interactions, such as non-interactions or randomly sampled items, serve as negative samples. By training the recommendation model to maximize the similarity between positive user–item pairs and minimize the similarity between negative user–item pairs in the learned representation space, CL-based recommendation endeavors to learn more informative representations of users and items. This approach enables the model to capture nuanced preferences, thereby improving its ability to make accurate recommendations. Recent research efforts, such as [20], have integrated CL into collaborative filtering by extending graph neural networks (GNNs). These models leverage bipartite graph representations of user–item interactions and incorporate high-order interactions and message passing to learn augmented user and item representations. Additionally, other recent models, such as [21], further enhance GNN-based approaches by learning different intents. These advancements highlight the growing importance and potential of CL-based techniques in recommendation systems.
Despite the significant advancements in recommender system models—particularly, the hybridization of collaborative techniques with deep learning (DL) models such as convolutional neural networks (CNNs) and autoencoders for extracting item features and the advancements in integrating the contrastive learning (CL) in recommendation, which is a promising and ongoing research area—certain crucial aspects remain overlooked. Many existing models heavily rely on static approaches like matrix factorization (MF) or probabilistic matrix factorization (PMF), disregarding the dynamic nature of user preferences over time. User rating tastes evolve with time; for instance, preferences for movie genres may differ between weekdays and weekends. Additionally, user behavior may vary based on seasonal factors; for example, individuals might purchase warm clothing during the winter season and switch to beachwear during the summer. Furthermore, while some models focus solely on extracting item features, they often neglect the equally important user features. In light of these limitations, our motivation drives the proposition of the IUAutoTimeSVD++ model. By integrating both item and user features within a temporal dynamic framework, IUAutoTimeSVD++ aims to address sparsity issues, extract meaningful user–item features, and effectively capture changes in user preferences over time through the utilization of timeSVD++.

3. Our Model IUAutoTimeSVD++

3.1. Preliminary

Prior to delving into the exposition of the model architecture, we provide a comprehensive definition of the symbols employed in current article, as summarized in Table 3:

3.2. Model Architecture

The proposed IUAutoTimeSVD++ model extends the temporal latent factors of the timeSVD++ model by incorporating a comprehensive representation of user–item features, facilitated by the powerful contractive autoencoder (CAE). Leveraging the CAE as an efficient technique for extracting meaningful representations of both user and item features, our model aims to enhance the accuracy of recommendations.
The architecture of the IUAutoTimeSVD++ model comprises the following key components:

3.2.1. Temporal Latent Factors Model

The timeSVD++ [12] model, an extension of the conventional collaborative filtering model SVD++ (Singular Value Decomposition Plus Plus) [14], builds upon its foundations to incorporate temporal dynamics. This extension allows the model to adapt to changes in user preferences over time. TimeSVD++ introduces several modifications to the original SVD++ formulation, affecting components such as item bias, user bias, and the introduction of a new term: time deviation d e v u O H E ( t ) to account for temporal variability. Each component comprises two parts: an initial static part and a dynamic part to capture evolving user preferences. The timeSVD++ model contains the following components:

Item Bias

The item bias is split into two parts, the initial static b i and dynamic binning B i n , depending on the time as follows:
b i ( t ) = b i + b i , Bin ( t )

Time Deviation

A deviation time to capture the rating time variability in linear form is as follows:
dev u ( t ) = sign ( t t u ) · | t t u | β

User Bias

We update the user bias by introducing the time deviation d e v u ( t ) .
b u ( t ) = b u + α u · dev u ( t ) + b u , t

User Latent Factors

In the same way as user bias, we introduce the deviation time in the user latent factors formula as follows:
p u ( t ) = p u + α u · dev u ( t ) + p u , t
Table 4 lists the parameters used in timeSVD++:
By combining the expressions of b u ( t ) , b i ( t ) , and p u ( t ) , we obtain the final expression of timeSVD++ model. Equation (5) represents the final timeSVD++ rating expression:
r ^ u i ( t ) = μ + b u ( t ) + b i ( t ) + q i ( p u ( t ) + | S ( u ) | 1 2 j S ( u ) y j )

3.2.2. Contractive Autoencoder

An autoencoder [22] is a type of artificial neural network designed for unsupervised learning. It consists of an encoder that compresses input data into a reduced-dimensional representation and a decoder that reconstructs the input from this representation. The training process encourages the autoencoder to learn efficient data encoding, making it well suited for tasks like feature learning and dimensionality reduction. A contractive autoencoder (CAE) [13] is a type of autoencoder. In the context of a CAE, the autoencoder architecture includes a regularization term inspired by the Jacobian matrix. The primary characteristic of a CAE is its use of this Jacobian-based regularization to enforce robustness in the learned representations. This regularization encourages the model to generate latent representations that are less sensitive to small variations or perturbations in the input data, making the autoencoder more robust and capable of capturing meaningful features.
The CAE components are summarized as follows:

Encoder

The encoder projects the input x into the hidden latent space h using the sigmoid activation function σ as follows:
h = σ ( W x + b h )

Decoder

The decoder produces an output through the reconstruction of the input from the hidden layer x assuming that W = W .
y = σ ( W h + b y )

Objective Function

The objective function is used for learning CAE parameters. J h ( x ) is the regulation term based on the Jacobian matrix in Frobenius norm.
L C A E ( θ ) = x D n ( L ( x , g ( f ( x ) ) ) ) + λ J h ( x ) F 2

Regulation Term

J h ( x ) F 2 = i j h j ( x ) x i 2
Table 5 outlines the parameters used to define the CAE model as follows:
The architecture of IUAutoTimeSVD++ is visually represented in Figure 1, providing a detailed overview of the model’s structure and the integration of temporal dynamics and item–user feature representations as follows:

3.3. IUAutoTimeSVD++ Methodology

To integrate user–item features into the timeSVD++ model, we follow a methodology outlined in the following steps:

3.3.1. One-Hot Encoding Representation

First, we collect the available user–item features, then we employ an algorithm based on one-hot encoding (OHE) to convert categorical and structural features into a binary representation. This technique transforms each categorical feature into multiple binary features, where each binary feature uniquely represents a category from the original feature. The one-hot encoding (OHE) process can be expressed through the following formulation: Let U and I denote the sets of users and items, respectively. A and B represents the set of user and item attributes, respectively. The one-hot encoding matrices for user and item features, denoted as O H E ( U , A ) and O H E ( I , B ) , respectively, have dimensions m × Y and n × X . Here, m stands for the number of users, n for the number of items, Y for the number of user attributes, and X for the number of item attributes. These matrices are defined as:
O H E ( U , A ) = x 11 x 12 x 1 Y x 21 x 22 x 2 Y x m 1 x m 2 x m Y
and
O H E ( I , B ) = x 11 x 12 x 1 X x 21 x 22 x 2 X x n 1 x n 2 x n X
Here, x i j is defined as:
x i j = 1 if the i - th user ( item ) has the j - th attribute 0 otherwise
In summary, while OHE is a simple, structured, and interpretable method for encoding categorical variables, it has drawbacks such as high dimensionality and sparse representations.

3.3.2. Features Extraction Using CAE

The contractive autoencoder (CAE) facilitates the learning of intricate non-linear mappings between the input, represented as one-hot encoding (OHE) vectors, and the resulting learned representation. This capability enables the model to discern and capture complex patterns and relationships within categorical data. Additionally, CAE has the capacity to derive a lower-dimensional representation of the initially high-dimensional OHE vectors. Rather than directly employing the sparse and high-dimensional OHE vectors, CAE learns a condensed and densely packed representation that encapsulates essential features. Unlike the binary vectors derived from OHE, the output of a CAE typically manifests as a continuous-valued vector. This continuous representation is adept at capturing the nuanced relationships and similarities between categories. Moreover, CAE is trained with the objective of reconstructing the input data from the learned representation. This training paradigm encourages the model to acquire robust features that exhibit decreased sensitivity to noise within the input data, potentially yielding more resilient representations compared to OHE, particularly in scenarios involving noisy or incomplete datasets.
To process the OHE features and extract meaningful item and user feature vectors, we employ two variants of CAE: Equation (10) delineates the CAE transformation for deriving the item features vector, and Equation (11) outlines the generation of the user features vector. The formulations are as follows:
cae ( v i O H E ) = σ ( W · v i O H E + b )
cae ( v u O H E ) = σ ( W · v u O H E + b )
Equation (10) applies the CAE transformation to the item vector v i O H E , with W denoting the weight matrix and b representing the bias term added to the hidden layer. Equation (11) applies a similar transformation to the user vector v u O H E , utilizing a weight matrix W and a bias term b . For simplicity, we assume W = W and b = b .

3.3.3. Rating Prediction Equation

Equation (12) expresses the final recommendation prediction in our model, utilizing the item and user feature vectors obtained from the CAE as follows:
r ^ u i ( t ) = μ + b i ( t ) + b u ( t ) + ( q i + ϕ . c a e ( v i O H E ) ) ( p u ( t ) + ψ . c a e ( v u O H E ) + j S ( u ) y j | S ( u ) | )
Here, r ^ u i ( t ) denotes the predicted rating for user u on item i at time t. The components of the prediction include the global mean μ , item bias b i ( t ) , user bias b u ( t ) , and the interactions between item factors q i , user factors p u ( t ) , and their respective CAE-derived feature vectors ϕ . c a e ( v i O H E ) and ψ . c a e ( v u O H E ) . The parameters ϕ and ψ are used to normalize the CAE vectors, ensuring the consistent scaling of the extracted features. The term j S ( u ) y j / | S ( u ) | represents the implicit feedback of a user u and a set of items S ( u ) , where y j is a binary variable indicating the interaction between the user u and the item i:
y j = 1 if the j - th user u has interaction with the item i 0 otherwise

3.3.4. Model Training and Optimization

To learn the model parameters, we train the model by minimizing the regularized squared error loss defined through the following loss function L o s s :
m i n L o s s = min u , i K [ r u i ( t ) ( q i + ϕ . c a e ( v i O H E ) ) ( p u ( t ) + ψ . c a e ( v u O H E ) + j S ( u ) y j | S ( u ) | ) μ b u b i ] 2 + λ . R
where λ is the learning rate and the R is the regularization terms. The term R is defined as:
R = B + q i 2 + p u 2 + j S ( u ) y j 2
where B is the bias regularization term. B is defined as:
B = b u 2 + b i 2 + b u ( t ) 2 + b i ( t ) 2 + α u 2
We employ the stochastic gradient descent (SGD) algorithm, a widely utilized optimization technique for recommender systems, to learn the parameters of the IUAutoTimeSVD++ model denoted by Θ = { p , q } . The objective of SGD, as illustrated in Equation (16), is to minimize the error. This error, denoted as e u i , represents the difference between the actual rating r u i ( t ) and the predicted rating r ^ u i ( t ) at time t:
e u i = r u i ( t ) r ^ u i ( t )
In the SGD algorithm, the model’s parameters p and q undergo iterative updates, driven by the computed error, to refine predictions and enhance overall model performance. The update rules for the IUAutoTimeSVD++ model, outlined in Equations (20)–(24), govern the evolution of key components. These include the user bias ( b u ), item bias ( b i ), latent factors ( p u and q i ), and time-dependent parameters ( b u ( t ) , b i ( t ) , and α u ). Each of these elements plays a crucial role in adapting the model to observed data, contributing to improved recommendation accuracy.
q i = q i + γ 2 ( e u i ( p u + ψ . c a e ( v u O H E ) ) λ 2 q i )
p u = p u + γ 2 ( e u i ( q i + ϕ . c a e ( v i O H E ) ) λ 2 p u )
y j = y j + γ 2 ( e u i | S ( u ) | 1 2 ( ϕ . c a e ( v i O H E ) ) λ 2 y i )
b u = b u + γ 1 ( e u i λ 1 b u )
b i = b i + γ 1 ( e u i λ 1 b i )
b u ( t ) = b u + γ 3 ( e u i λ 3 b u )
b i ( t ) = b i + γ 3 ( e u i λ 3 b i )
α u = α u + γ α ( e u i λ α α u )
Here, γ 1 , γ 2 , γ 3 , and γ α represent the learning rates, and λ 1 , λ 2 , λ 3 , and λ α represent the regularization values.

4. Experimental Results

4.1. Datasets

To evaluate our model, we conducted many experiments using two public datasets: M L 100 K and M L 1 M . The datasets are sourced from MovieLens [23]. The statistics of both datasets, including the number of interactions (# Interactions), the number of unique users (# Users) and the number of unique items (# Items), are presented in Table 6:
Table 7 presents an example of user and item (movie) features extracted from the MovieLens datasets:

4.2. Evaluation Metrics

We use RMSE and MAE [24] to evaluate the performance of our model; they are widely utilized as metrics for evaluating the RS accuracy. Equations (25) and (26) define RMSE and MAE, respectively:
R M S E = 1 | T | ( u , i ) T ( r ^ u i r u i ) 2
M A E = 1 | T | ( u , i ) T r ^ u i r u i
where T denotes the test set, r ^ u i denotes the score predicted by the model, and r u i represents the actual score of the test set. By calculating the RMSE and MAE, we can assess the accuracy of our model’s predictions compared to the ground truth ratings in the test set. A reduced RMSE (or MAE) value signifies improved performance, indicating that the model’s predictions closely align with the actual ratings.

4.3. Baseline Models

We compared our IUAutoTimeSVD++ model with the following baselines:
  • MF: The basic variant of the matrix factorization model proposed in [14].
  • PMF: A probabilistic variant of the matrix factorization proposed in [18].
  • SVD++: An extension of the MF integrating the user implicit feedback proposed in [14].
  • convMF: An extension of the PMF leveraging the textual content of the item and user reviews using CNN. [9].
  • DBPMF: An extension of the PMF proposed in [19], leveraging item–user features using CNN.
  • autoSVD++: A hybrid model proposed in [8] extending SVD++ through incorporating item features using CAE.
  • SSAERec: A hybrid model proposed in [10] extending SVD++ and incorporating item features using stacked autoencoders.
  • timeSVD++: The baseline temporal dynamic latent factors model proposed in [12].

4.4. Hyperparameter Analysis of Baseline Models and the Proposed Model

Hyperparameters play a crucial role in shaping the performance and behavior of machine learning models. Unlike the parameters of a model, which are learned during training, hyperparameters are set prior to training and govern various aspects of the learning process. They influence the model’s capacity, regularization, optimization strategy, and ability to generalize to unseen data. Understanding the importance of hyperparameters and categorizing them appropriately is essential for effectively tuning models to achieve optimal performance.

4.4.1. Categorization of Hyperparameters

Hyperparameters can be categorized based on their role and influence on the model. Some parameters are fundamental and common across different types of models, while others are specific to certain model architectures or techniques. Common hyperparameters include those related to model complexity (e.g., number of latent factors, number of layers), regularization (e.g., learning rate, regularization term), and optimization (e.g., optimization algorithm, batch size). On the other hand, hyperparameters associated with hybrid aspects of models, such as convolutional neural networks (CNNs) or autoencoders, may include parameters specific to the architecture (e.g., number of filters, filter sizes) or feature extraction techniques (e.g., sparsity constraints, dropout rates).

4.4.2. Fine-Tuning Model Performance across Baseline and Proposed Model

Matrix Factorization (MF)

MF [12] model parameters include:
  • Number of latent factors (k): This parameter controls the dimensionality of the latent factor space. It typically ranges from 10 to 100. Larger values may capture more complex interactions but increase computational cost and risk overfitting.
  • Learning rate ( α ): Determines the step size in the optimization process. Common values range from 0.001 to 0.1. Higher values can speed up convergence but may lead to instability.
  • Regularization term ( λ ): Balances the model’s fit to the training data and its complexity. Values typically range from 0.01 to 0.1. Higher values penalize large parameter values, preventing overfitting.
  • Number of epochs: The number of iterations for optimization. This parameter affects convergence and computational cost. The appropriate number of iterations can vary depending on factors such as the complexity of the dataset, the size of the model, and the optimization algorithm used.

SVD++

Similar to MF, SVD++ includes parameters like the number of latent factors, learning rate, and regularization term. It additionally includes:
  • Implicit feedback weight ( α ) Balances the influence of implicit feedback signals. It is crucial to tune this parameter to weigh implicit feedback appropriately.

Probabilistic Matrix Factorization (PMF)

PMF [18] incorporates probabilistic modeling, introducing additional hyperparameters:
  • Variance of the Gaussian prior ( σ 2 ): Controls the spread of the Gaussian distribution over latent factors. A smaller value imposes stronger regularization.
  • Hyperparameters of the prior distribution: PMF often uses priors such as Gaussian or Laplace distributions. Tuning these hyperparameters influences the model’s robustness to overfitting.

ConvMF (Convolutional Matrix Factorization)

Similar to MF, ConvMF [9] uses learning rate and regularization for optimization and controlling model complexity. Furthermore, it uses the following parameters:
  • Dimension: Similar to latent factor. It represents the size of latent dimension for users and items.
  • λ u : User regularization parameter.
  • λ v : Item regularization parameter.
  • CNN hyperparameters: Includes hyper parameters used for training the CNN model to extract embedding representation of the word such as number of kernels per window size and the size of latent dimension for word vectors.

Deep Bias Probabilistic Matrix Factorization (DBPMF)

DBPMF [19] uses the following parameters:
  • Number of latent factors (K): Similar to MF and PMF, controls the dimensionality of latent factors.
  • Hyperparameters of the dynamic model: DBPMF typically incorporates additional hyperparameters related to modeling temporal dynamics, such as transition matrices, state transition probabilities, and variance of state dynamics.
  • Learning rate, regularization term, and variances: hyperparameters for optimizing the model and controlling the balance between fit to the data and model complexity.

AutoSVD++

AutoSVD++ [8] extends SVD++ with item features extracted using CAE [13] for improved performance. hyperparameters include, in addition to SVD++ parameters, the following parameters:
  • Hyperparameter β for feature extraction: This parameter is used to normalize the vector features extracted using CAE and avoid the model overfitting. The paper [8] does not provide any additional information about CAE model training.

SSAERec

SSAERec [10] extends the SVD++ model, incorporating additional parameters beyond those employed in SVD++. Among these, the following parameter is introduced:
  • Hyperparameter β : This hyperparameter serves to normalize the Stacked Sparse Auto-Encoder (SSAE) utilized for extracting feature representations.

TimeSVD++

For timeSVD++ [12], the hyperparameters include those from the SVD++ model as well as additional parameters related to temporal dynamics as follows:
  • α u : A regularization hyperparameter to control the time deviation term.
  • β : A hyperparameter used to control the time deviation. It depends on the dataset determined through cross validation.
  • Number of Time Bins: The timeSVD++ splits the time dimension into multiple time bins to capture temporal dynamics. The number of time bins is a hyperparameter that determines the granularity of the temporal modeling.

IUAutoTimeSVD++

Our model IUAutoTimeSVD++ uses the same hyperparameters corresponding to the timeSVD++ model. Additionally, it uses the following hyperparameters to control the user–item features contribution in the rating as follows:
  • Item CAE vector hyperparameter ϕ : Used to normalize the item vector features extracted using CAE and avoid the model overfitting. Based on experiments and validation, the values range from 0.01 to 0.1 .
  • User CAE vector hyperparameter ψ : Used to normalize the user vector features. Similar to ϕ , common values vary from 0.01 to 0.1 .
  • CAE Hyperparameters: The CAE Hyperparameters used to train CAE to generate the user and item features are illustrated in the next subsection.

4.5. Hyperparameter Settings

4.5.1. CAE Training Hyperparameters

We use the Keras library to train the CAE model and extract the pre-trained user–item features. During the training phase, we configure the model with the following parameters: h s i z e = 600 to set the size of the hidden layer, λ = 1 × 10 5 as the regularization parameter, b a t c h s i z e = 128 to determine the batch size used in training, and e p o c h s = 20 to specify the number of training epochs. The Adam optimizer is selected to optimize the model’s performance throughout the training process. With these settings, we aim to obtain effective and robust representations of the user–item features through the CAE model.

4.5.2. IUAutoTimeSVD++ Training Hyperparameters

To train the IUAutoTimeSVD++ model, we use the parameter settings illustrated in Table 8:
The learning rate and regularization values utilized in both SVD++ [14] and autoSVD++ [8] were kept consistent. However, when it comes to the parameters λ α , γ α , β , and α u , which are associated with the time aspects, a cross-validation approach was employed to identify the optimal values. Similarly, for the parameters ϕ and ψ , cross-validation was utilized to determine the best values based on a set of values for each hyperparameter. In contrast, the learning rate and regularization values were kept consistent between SVD++ and autoSVD++, providing a fair comparison between the two models with respect to their base configuration. This consistency allows us to focus on the impact of the additional components introduced in IUAutoTimeSVD++ model and their contributions in improving recommendation accuracy. The model hyperparameters are set as mentioned in the author’s paper.

4.6. Results

4.6.1. Numerical Results

Table 9 presents the experiment results of our model in comparison with the other baseline models on both ML-100K and ML-1M datasets as follows:
Upon analyzing the RMSE results presented in Table 9, which includes the proposed IUAutoTimeSVD++ model evaluated on two distinct datasets, ML-100K and ML-1M, it is evident that IUAutoTimeSVD++ surpasses timeSVD++. The achieved lower RMSE values on both datasets (0.892 on ML-100K and 0.845 on ML-1M) indicate that our proposed model, incorporating temporal dynamics and user–item features, leads to enhanced prediction accuracy.
Furthermore, IUAutoTimeSVD++ outperforms various baseline models (PMF, MF, SVD++, convMF, DBPMF, autoSVD++, and SSAERec) on both datasets, showcasing its superiority over the compared baseline models.

4.6.2. Improvement Analysis

To quantify the effectiveness of the IUAutoTimeSVD++ model over the baseline models, we compute the improvements of our model and other baseline models compared to the baseline model MF [14] as shown in Figure 2:
Indeed, based on the improvement analysis results in Figure 2, it is evident that the IUAutoTimeSVD++ model consistently outperforms the baseline models. The improvement plot clearly shows that the bars for IUAutoTimeSVD++ are above the horizontal dashed line (representing the baseline performance) for both datasets. This indicates that the IUAutoTimeSVD++ model achieves a lower RMSE value as compared to the baseline models, indicating superior predictive accuracy in making recommendations. The improvements showing that the integration of temporal dynamics and user–item features extracted through the contractive autoencoder leads to enhanced recommendation performance. The finding that IUAutoTimeSVD++ outperforms the baselines is valuable, as it demonstrates the effectiveness of the proposed model in addressing the limitations of traditional matrix factorization models through the use of CAE to extract user–item features. It also indicates that the inclusion of temporal dynamics and user–item features enhances the accuracy recommendation.

4.7. Effectiveness Study of Training Size Impact

To study the impact of the training/test dataset size, we conducted many experiments using several values of training size { 60 , 70 , 80 , 90 } . The results are presented in Table 10 and Table 11:
We complement the effectiveness study results presented in Table 10 and Table 11 with visual representations in the following figures:

Effectiveness Study Results Analysis

Based on the comprehensive analysis of performance results across different test ratio configurations, as illustrated in Table 10 and Table 11 (Figure 3 and Figure 4) it becomes evident that the IUAutoTimeSVD++ model consistently outperforms other models. This is reflected in its superior accuracy, with lower values of RMSE and MAE on both the ML-100K and ML-1M datasets. Notably, the IUAutoTimeSVD++ model exhibits remarkable performance regardless of the proportion of data used for testing (10%, 20%, 30%, or 40%). It consistently demonstrates robustness and effectiveness in making accurate predictions and recommendations across various testing scenarios. Additionally, it is noteworthy that all the compared models achieve a better accuracy when utilizing a higher training ratio (90% of the data), as evidenced by lower RMSE values for both the ML-100K and ML-1M datasets. This observation emphasizes the importance of allocating more resources for the training phase. A higher training ratio allows the model to learn from a more extensive and representative portion of the data.

4.8. Ablation Study

In this subsection, we conduct an ablation study to dissect and scrutinize the key components influencing our model’s performance. Specifically, we isolate and analyze the impact of three critical factors: the temporal aspect, user features, and item features. By systematically varying or removing each component individually, we aim to discern their respective contributions to the overall model efficacy. Table 12 enumerates the components employed in each variant examined in our ablation study:

4.8.1. Results

The results of the ablation study on both datasets are presented in Table 13 and Table 14:
Based on analysis of the ablation study results in Table 13 and Table 14, the IUAutoTimeSVD++ model, which encompasses all three components (time, user, and item features), demonstrated the lowest RMSE and MAE values. This performance advantage signifies that IUAutoTimeSVD++ outperformed the other variants concerning predictive accuracy. The ablation study provided valuable insights into the significance of each component within the model. The presence of time, user, and item features working together smoothly enhanced the model’s ability to make the predictions more accurate. Notably, the results indicate that user and item features hold more importance than the time factor in predicting ratings. The integration of user features had a more substantial impact on the model’s performance compared to incorporating item features. Moreover, it was observed that leveraging only the user features yielded better accuracy.

4.8.2. Improvement Ablation Study Analysis

To gain a comprehensive understanding of the performance improvement of each variant of the model, we assess the improvement of all variants in comparison to the baseline model, timeSVD++. In this analysis, we calculate the difference in performance metrics (e.g., RMSE and MAE) between each variant and the baseline. This allows us to quantify how much better or worse each variant performs relative to the baseline. The results are illustrated in Figure 5 and Figure 6:
In light of analysis of the improvement plots (Figure 5 and Figure 6) the IUAutoTimeSVD++ model consistently outperforms the other variants and the baseline model timeSVD++, demonstrating its effectiveness in improving the recommendation accuracy. The inclusion of both user and item features (IUAutoTimeSVD++) appears to be more beneficial than including either user or item features alone (I-AutoTimeSVD++ and U-AutoTimeSVD++). The IU-autoSVD++ variant, which does not include the time component, performs slightly worse compared to the IUAutoTimeSVD++ model, suggesting that consideration of the temporal dynamics is essential for enhancing the model’s prediction accuracy. Overall, the ablation study highlights the significance of each component in the proposed IUAutoTimeSVD++ model and underscores its superiority in making accurate predictions and recommendations on both ML-100K and ML-1M datasets. The results provide valuable insights into the model’s effectiveness and the importance of incorporating user, item, and temporal features to tackle the sparsity problem and enhance the matrix factorization model for recommender systems.

4.9. Experimental Analysis on Sparse Synthetic Dataset: Subsampling MovieLens Data

We delve into the experimental analysis conducted on a synthetic dataset derived from the MovieLens dataset. The synthetic dataset is generated through subsampling (selecting the 30% of the user reviews), where a fraction of the original MovieLens data is retained to create a sparser version. This process allows us to explore the behavior of recommendation algorithms under varying degrees of data sparsity and evaluate the robustness of our proposed model, mimicking real-world scenarios with limited interaction data. Table 15 present the statistics of the synthetic datasets (ML-100K and ML-1M) after resampling and only selecting (30%) of reviews.
In our evaluation of compared recommendation models on synthetic datasets derived from the original MovieLens datasets (ML-100K and ML-1M) by retaining only 30% of ratings, we aimed to assess the models’ adaptability and performance under sparse data conditions. The results indicate that IUAutoTimeSVD++ consistently demonstrates robust performance on the synthetic datasets, despite the lower data density. With RMSE scores of 0.959 on Synthetic ML-100K and 0.906 on Synthetic ML-1M as shown in Table 16, IUAutoTimeSVD++ exhibits resilience and effectiveness in capturing underlying user–item interactions even with limited rating information. This highlights its ability to generalize well and maintain performance across varying data densities. However, it is important to note that while IUAutoTimeSVD++ excelled on the synthetic datasets, the evaluation was limited to this model due to the unavailability of code for other baseline models. As a result, the performance of these models under sparse conditions remains unexplored. Nonetheless, the strong performance of IUAutoTimeSVD++ on the synthetic datasets underscores its potential for real-world applications where data sparsity is a challenge due to limited user–item interactions. This analysis provides valuable insights into how these models fare when faced with reduced data density, which is common in real-world applications.

5. Conclusions

We proposed IUAutoTimeSVD++, a novel hybrid model extending the temporal dynamic latent factor model timeSVD++ by incorporating user side information and item features. Our primary goal was to tackle the sparsity issue in matrix factorization and ensure a dynamic rating prediction mechanism. The proposed model leverages a contractive autoencoder to combine pre-trained item and user features, enabling a robust representation of the input features. Our experimental results have shown that IUAutoTimeSVD++ outperforms several baseline models, including those based on time and latent factors. The model’s superior performance validates the effectiveness of incorporating user and item features and the temporal dynamics into the rating prediction process. As a future direction, we aim to enhance the model’s capability in capturing short-term interactions. Currently, IUAutoTimeSVD++ focuses on modeling long-term user tastes, but incorporating short-term preferences, such as session-based interactions or recent user behaviors, is essential for providing more timely and accurate recommendations. To achieve this, we plan to adopt a recurrent neural network (RNN) and techniques like gated recurrent units (GRUs) [25] and long short-term memory (LSM) [26], which are adept at capturing sequential patterns in user–item interactions. Another promising research direction combining contrastive learning (CL) framework for matrix factorization models—where positive user–item interactions are contrasted with negative interactions to learn more informative latent representations inspired by some recent research such as [20] and the contractive autoencoder (CAE) providing a compact and dense representations of user and item features—could focus on developing a novel recommendation model that leverages the strengths of each technique. Additionally, we intend to explore additional rich features, such as visual and textual item attributes, to create a more comprehensive and personalized recommendation system. Moreover, we plan to explore parallel and distributed computing models like DSGD (distributed stochastic gradient descent) [27] to enhance the model’s efficiency and scalability for handling larger datasets.

Author Contributions

Conceptualization, A.A. and A.H.; methodology, A.A.; software, A.A.; validation, A.A., A.H. and H.A.; formal analysis, A.A.; investigation, A.A.; resources, A.A.; data curation, A.A.; writing—original draft preparation, A.A.; writing—review and editing, A.A. and A.H.; visualization, A.A.; supervision, A.H. and H.A.; project administration, H.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ricci, F.; Rokach, L.; Shapira, B. (Eds.) Recommender Systems Handbook, 2nd ed.; Springer: New York, NY, USA, 2015. [Google Scholar] [CrossRef]
  2. Ekstrand, M.D.; Riedl, J.T.; Konstan, J.A. Collaborative Filtering Recommender Systems. Found. Trends Hum.-Comput. Interact. 2011, 4, 81–173. [Google Scholar] [CrossRef]
  3. Bell, R.M.; Koren, Y. Lessons from the Netflix Prize Challenge. SIGKDD Explor. Newsl. 2007, 9, 75–79. [Google Scholar] [CrossRef]
  4. Rendle, S.; Krichene, W.; Zhang, L.; Anderson, J. Neural Collaborative Filtering vs. Matrix Factorization Revisited. arXiv 2020, arXiv:2005.09683. [Google Scholar]
  5. Anelli, V.W.; Bellogín, A.; Di Noia, T.; Pomo, C. Reenvisioning Collaborative Filtering vs. Matrix Factorization. In Proceedings of the Fifteenth ACM Conference on Recommender Systems, Amsterdam, The Netherlands, 27 September–1 October 2021; pp. 521–529. [Google Scholar]
  6. Zhang, S.; Yao, L.; Sun, A.; Tay, Y. Deep Learning Based Recommender System: A Survey and New Perspectives. ACM Comput. Surv. 2019, 52, 1–38. [Google Scholar] [CrossRef]
  7. Mu, R. A Survey of Recommender Systems Based on Deep Learning. IEEE Access 2018, 6, 69009–69022. [Google Scholar] [CrossRef]
  8. Zhang, S.; Yao, L.; Xu, X. AutoSVD++: An Efficient Hybrid Collaborative Filtering Model via Contractive Auto-encoders. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, Tokyo, Japan, 7–11 August 2017; pp. 957–960. [Google Scholar] [CrossRef]
  9. Kim, D.; Park, C.; Oh, J.; Lee, S.; Yu, H. Convolutional Matrix Factorization for Document Context-Aware Recommendation. In Proceedings of the 10th ACM Conference on Recommender Systems RecSys ’16, New York, NY, USA, 15–19 September 2016; pp. 233–240. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Zhao, C.; Chen, M.; Yuan, M. Integrating Stacked Sparse Auto-Encoder Into Matrix Factorization for Rating Prediction. IEEE Access 2021, 9, 17641–17648. [Google Scholar] [CrossRef]
  11. Azri, A.; Haddi, A.; Allali, H. autoTimeSVD++: A Temporal Hybrid Recommender System Based on Contractive Autoencoder and Matrix Factorization. In Proceedings of the Smart Applications and Data Analysis—4th International Conference, SADASC 2022, Marrakesh, Morocco, 22–24 September 2022; Hamlich, M., Bellatreche, L., Siadat, A., Ventura, S., Eds.; Springer: Berlin/Heidelberg, Germany, 2022; Volume 1677, Communications in Computer and Information Science. pp. 93–103. [Google Scholar] [CrossRef]
  12. Koren, Y. Collaborative filtering with temporal dynamics. In Proceedings of the KDD09: The 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Paris, France, 28 June–1 July 2009; pp. 447–456. [Google Scholar]
  13. Rifai, S.; Vincent, P.; Muller, X.; Glorot, X.; Bengio, Y. Contractive auto-encoders: Explicit invariance during feature extraction. In Proceedings of the 28th International Conference on International Conference on Machine Learning ICML’11, Madison, WI, USA, 28 June–2 July 2011; pp. 833–840. [Google Scholar]
  14. Koren, Y.; Bell, R.; Volinsky, C. Matrix Factorization Techniques for Recommender Systems. Computer 2009, 42, 30–37. [Google Scholar] [CrossRef]
  15. He, R.; McAuley, J. VBPR: Visual Bayesian Personalized Ranking from Implicit Feedback. arXiv 2015, arXiv:1510.01784. [Google Scholar] [CrossRef]
  16. Abdollahi, B.; Nasraoui, O. Explainable Matrix Factorization for Collaborative Filtering. In Proceedings of the 25th International Conference Companion on World Wide Web WWW ’16 Companion, Geneva, Switzerland, 11–15 April 2016; pp. 5–6. [Google Scholar] [CrossRef]
  17. He, X.; Liao, L.; Zhang, H.; Nie, L.; Hu, X.; Chua, T.S. Neural Collaborative Filtering. In Proceedings of the 26th International Conference on World Wide Web WWW ’17, Perth, Australia, 3–7 April 2017; pp. 173–182. [Google Scholar] [CrossRef]
  18. Mnih, A.; Salakhutdinov, R.R. Probabilistic Matrix Factorization. In Advances in Neural Information Processing Systems 20; Platt, J.C., Koller, D., Singer, Y., Roweis, S.T., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2008; pp. 1257–1264. [Google Scholar]
  19. Li, K.; Zhou, X.; Lin, F.; Zeng, W.; Alterovitz, G. Deep Probabilistic Matrix Factorization Framework for Online Collaborative Filtering. IEEE Access 2019, 7, 56117–56128. [Google Scholar] [CrossRef]
  20. Ren, X.; Xia, L.; Zhao, J.; Yin, D.; Huang, C. Disentangled contrastive collaborative filtering. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, Taipei, Taiwan, 23–27 July 2023; pp. 1137–1146. [Google Scholar]
  21. Wang, Y.; Wang, X.; Huang, X.; Yu, Y.; Li, H.; Zhang, M.; Guo, Z.; Wu, W. Intent-aware Recommendation via Disentangled Graph Contrastive Learning. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI-2023, Macao, 19–25 August 2023. [Google Scholar] [CrossRef]
  22. Bank, D.; Koenigstein, N.; Giryes, R. Autoencoders. arXiv 2021, arXiv:2003.05991. [Google Scholar]
  23. Harper, F.M.; Konstan, J.A. The movielens datasets: History and context. Acm Trans. Interact. Intell. Syst. 2015, 5, 1–19. [Google Scholar] [CrossRef]
  24. Aggarwal, C.C. Recommender Systems: The Textbook, 1st ed.; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  25. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. On the properties of neural machine translation: Encoder-decoder approaches. arXiv 2014, arXiv:1409.1259. [Google Scholar]
  26. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  27. Gemulla, R.; Nijkamp, E.; Haas, P.J.; Sismanis, Y. Large-scale matrix factorization with distributed stochastic gradient descent. In Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Diego, CA, USA, 21–24 August 2011; pp. 69–77. [Google Scholar] [CrossRef]
Figure 1. IUAutoTimeSVD++ model architecture.
Figure 1. IUAutoTimeSVD++ model architecture.
Information 15 00204 g001
Figure 2. Improvement in RMSE for IUAutoTimeSVD++ model over baseline (MF) on ML-100K and ML-1M Datasets.
Figure 2. Improvement in RMSE for IUAutoTimeSVD++ model over baseline (MF) on ML-100K and ML-1M Datasets.
Information 15 00204 g002
Figure 3. RMSE results for different training percentages on ML-100K.
Figure 3. RMSE results for different training percentages on ML-100K.
Information 15 00204 g003
Figure 4. RMSE results for different training percentages on ML-1M.
Figure 4. RMSE results for different training percentages on ML-1M.
Information 15 00204 g004
Figure 5. Improvement in ablation study results for ML-100K dataset.
Figure 5. Improvement in ablation study results for ML-100K dataset.
Information 15 00204 g005
Figure 6. Improvement in ablation study results for ML-1M dataset.
Figure 6. Improvement in ablation study results for ML-1M dataset.
Information 15 00204 g006
Table 1. Matrix factorization extension models.
Table 1. Matrix factorization extension models.
Baseline ModelExtension
MF [14]ModelTypeTechniqueYearTime factorUser featuresItem featuresVisual features
SVD++ [14]Implicit feedbackDot product2008
timeSVD++ [12]TemporalDot product2009
VBPR [15]VisualCNN2016
EMF [16]ExplainabilitySimilarity2016
autoSVD++ [8]HybridAutoencoder2017
NeuMF [17]EmbeddingMLP2017
SSAERec [10]HybridAutoencoder2021
Table 2. Probabilistic matrix factorization extension models.
Table 2. Probabilistic matrix factorization extension models.
Baseline ModelExtension
PMF [18]ModelTypeTechniqueYearTime factorUser featuresItem featuresVisual features
ConvMF [9]TextualCNN2016
DBPMF [19]HybridCNN2019
Table 3. Notations and symbols used in the article.
Table 3. Notations and symbols used in the article.
NotationDescription
U The user’s set defined as { u 1 , u 2 , , u m }
I The item’s set defined as { i 1 , i 2 , , i n }
nThe number of items where n N
mThe number of users where m N
v i O H E OHE item features vector v i O H E = { x 1 , x 2 , , x X } of size X
v u O H E OHE user features vector v u O H E = { x 1 , x 2 , , x Y } of size Y
R Interaction of rating matrix
Q Low rank matrix representing items
P Low rank matrix representing users
KThe latent factors space
kThe latent factors dimension
r u i The rating value provided by a user u of an item i
r u i ( t ) The rating value at the time t
CAEContractive autoencoder
SVDSingle value decomposition
SGDStochastic gradient Descent
Table 4. TimeSVD++ Parameters.
Table 4. TimeSVD++ Parameters.
ParameterDescription
B i n ( t ) Time-based bin for mapping a day to its rank in a month
β Parameter for the drift concept
α u Coefficient for the drift in user bias
Table 5. Description of CAE Parameters.
Table 5. Description of CAE Parameters.
ParameterDefinition
b h Bias term added to hidden layers for shifting output
b y Additional bias term contributing to model flexibility
D n Dimensionality of the latent space
WWeight matrix for transforming input data in the encoding process
W Additional weight matrix, often associated with decoding in the autoencoder
λ Regularization parameter controlling Jacobian regularization strength
Table 6. Dataset Statistics.
Table 6. Dataset Statistics.
Dataset# Interactions# Users# ItemsSparsity (%)
MovieLens 100K100,000943168293.70
MovieLens 1M1,000,2096040395295.53
Table 7. Example of User–Item Features.
Table 7. Example of User–Item Features.
ElementAvailable Features
UserAge, gender, job
ItemGenre, Year of apparition
Table 8. Hyperparameter settings of the IUAutoTimeSVD model.
Table 8. Hyperparameter settings of the IUAutoTimeSVD model.
ParameterDescriptionValue
ϕ I-CAE learning rate 0.1 (ML-100K), 0.01 (ML-1M)
ψ U-CAE learning rate 0.1 (ML-100K), 0.01 (ML-1M)
kLatent factors dimension10
λ 1 Learning rate0.005
λ 2 Learning rate0.015
λ 3 Learning rate0.015
λ α Learning rate of user0.0004
γ 1 Regularization0.005
γ 2 Regularization0.007
γ 3 Regularization0.001
γ α Regularization of user0.00001
β Control time deviation0.015
epochsNumber of iteration25
Table 9. RMSE Results for ML-100K and ML-1M Datasets.
Table 9. RMSE Results for ML-100K and ML-1M Datasets.
ModelML-100K RMSEML-1M RMSE
MF0.9350.873
PMF0.9150.853
SVD++0.9250.855
timeSVD++0.9190.852
convMF0.9140.856
DBPMF0.9990.948
autoSVD++0.9090.851
SSAERec0.9020.852
IUAutoTimeSVD++0.8920.845
Table 10. RMSE and MAE experimental results for different training percentages on ML-100K.
Table 10. RMSE and MAE experimental results for different training percentages on ML-100K.
DatasetTrainingMetricsSVD++autoSVD++timeSVD++IUAutoTimeSVD++
ML-100K90%RMSE0.9250.9090.9190.892
MAE0.7310.7120.7240.699
80%RMSE0.9430.9250.9290.912
MAE0.7420.7260.7330.716
70%RMSE0.9350.9200.9390.915
MAE0.7390.7250.7420.720
60%RMSE0.9440.9380.9410.936
MAE0.7440.7380.7410.738
Table 11. RMSE and MAE experimental results for different training percentages on ML-1M.
Table 11. RMSE and MAE experimental results for different training percentages on ML-1M.
DatasetTrainingMetricsSVD++autoSVD++timeSVD++IUAutoTimeSVD++
ML-1M90%RMSE0.8550.8510.8520.845
MAE0.6730.6660.6690.661
80%RMSE0.8610.8540.8580.853
MAE0.6760.6690.6730.668
70%RMSE0.8680.8590.8670.860
MAE0.6830.6720.6810.672
60%RMSE0.8730.8710.8710.867
MAE0.6870.6810.6840.678
Table 12. Ablation Study on Model Components.
Table 12. Ablation Study on Model Components.
ModelTimeItem FeaturesUser Features
timeSVD++
I-AutoTimeSVD++
U-AutoTimeSVD++
IU-autoSVD++
IUAutoTimeSVD++
Table 13. Ablation study results for ML-100K.
Table 13. Ablation study results for ML-100K.
ModelRMSEMAE
timeSVD++0.9190.724
I-AutoTimeSVD++0.8960.704
U-AutoTimeSVD++0.8960.704
IU-autoSVD++0.9090.712
IUAutoTimeSVD++0.8920.699
Table 14. Ablation study results for ML-1M.
Table 14. Ablation study results for ML-1M.
ModelRMSEMAE
timeSVD++0.8520.669
I-AutoTimeSVD++0.8510.667
U-AutoTimeSVD++0.8450.662
IU-autoSVD++0.8510.666
IUAutoTimeSVD++0.8450.661
Table 15. Dataset Statistics.
Table 15. Dataset Statistics.
Dataset# Interactions# Users# ItemsSparsity (%)
Synthetic MovieLens 100K30,000943149998.09
Synthetic MovieLens 1M300,0636040353198.76
Table 16. RMSE results for synthetic ML-100K and ML-1M Datasets.
Table 16. RMSE results for synthetic ML-100K and ML-1M Datasets.
ModelSynthetic ML-100K RMSESynthetic ML-1M RMSE
MF0.9620.915
SVD++0.9610.913
timeSVD++0.9590.909
autoSVD++0.9611.010
IUAutoTimeSVD++0.9590.906
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Azri, A.; Haddi, A.; Allali, H. IUAutoTimeSVD++: A Hybrid Temporal Recommender System Integrating Item and User Features Using a Contractive Autoencoder. Information 2024, 15, 204. https://doi.org/10.3390/info15040204

AMA Style

Azri A, Haddi A, Allali H. IUAutoTimeSVD++: A Hybrid Temporal Recommender System Integrating Item and User Features Using a Contractive Autoencoder. Information. 2024; 15(4):204. https://doi.org/10.3390/info15040204

Chicago/Turabian Style

Azri, Abdelghani, Adil Haddi, and Hakim Allali. 2024. "IUAutoTimeSVD++: A Hybrid Temporal Recommender System Integrating Item and User Features Using a Contractive Autoencoder" Information 15, no. 4: 204. https://doi.org/10.3390/info15040204

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop