Next Article in Journal
Understanding Statistical Hypothesis Testing: The Logic of Statistical Inference
Previous Article in Journal
A Radiative Transfer Model-Based Multi-Layered Regression Learning to Estimate Shadow Map in Hyperspectral Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Matrix Factorization Algorithm for Efficient Recommendations in Social Rating Networks Using Constrained Optimization

by
Nicholas Ampazis
*,
Theodoros Emmanouilidis
and
Flora Sakketou
Department of Financial and Management Engineering, University of the Aegean, Kountouriotou 41, Chios 82100, Greece
*
Author to whom correspondence should be addressed.
Mach. Learn. Knowl. Extr. 2019, 1(3), 928-944; https://doi.org/10.3390/make1030053
Submission received: 4 June 2019 / Revised: 31 July 2019 / Accepted: 9 August 2019 / Published: 11 August 2019
(This article belongs to the Section Learning)

Abstract

:
In recent years the emergence of social media has become more prominent than ever. Social networking has become the de facto tool used by people all around the world for information discovery. Consequently, the importance of recommendations in a social network setting has urgently emerged, but unfortunately, many methods that have been proposed in order to provide recommendations in social networks cannot produce scalable solutions, and in many cases are complex and difficult to replicate unless the source code of their implementation has been made publicly available. However, as the user base of social networks continues to grow, the demand for developing more efficient social network-based recommendation approaches will continue to grow as well. In this paper, following proven optimization techniques from the domain of machine learning with constrained optimization, and modifying them accordingly in order to take into account the social network information, we propose a matrix factorization algorithm that improves on previously proposed related approaches in terms of convergence speed, recommendation accuracy and performance on cold start users. The proposed algorithm can be implemented easily, and thus used more frequently in social recommendation setups. Our claims are validated by experiments on two real life data sets, the public domain Epinions.com dataset and a much larger dataset crawled from Flixster.com.

1. Introduction

Matrix factorization in collaborative filtering recommender systems is usually performed by unconstrained gradient descent for learning the feature components of the user and item factor matrices [1]. This is essentially a “black box” approach, where apart from the minimization of an objective function (usually the Root Mean Squared Error (RMSE) over the known ratings), generally no other information or knowledge is taken into account during the factorization process. However this approach alone cannot solve efficiently a majority of recommendation problems. The most notable example is the Netflix Prize problem, which dealt with a very sparse, high-dimensional dataset with more than 100 million training patterns. The winning solutions of the “Bellkor’s Pragmatic Chaos” team [2,3,4], made apparent that a huge number of latent factor models, clustering, and K-nearest neighbors approaches was necessary to be combined (both linearly and non-linearly) in order to provide an accurate final solution.
In earlier work where our focus was on training feedforward neural networks [5] we showed that, during training, it is useful to incorporate additional “incremental conditions”, i.e., conditions involving quantities that must be optimized incrementally at each epoch of the learning process (by the term “epoch” we denote a training cycle of presenting the entire training set). We thus formulated a general problem, whereby we sought the minimization of the objective function representing the distance of the network’s outputs from preset target values, subject to other constraints that represented the additional knowledge. We then demonstrated how to formulate the general problem of incorporating the additional knowledge as a constrained optimization task, whose solution lead to a powerful generic learning algorithm accounting for both target and incremental conditions. Advancing that work, in a recent study [6], and similarly to other proposed methods [7], we cast the factorization of the user-by-item ratings matrix as a feedforward neural network training problem, and thus concentrated on the development of constrained optimization methods that could lead to efficient factorization algorithms. In that study, we introduced a general constrained matrix factorization framework that we refer to as FALCON (Factorization Algorithms for Learning with Constrained OptimizatioN), and presented two examples of algorithms which can be derived from that framework, that incorporate additional knowledge about learning the factor matrices. The first example (FALCON-M), incorporated an extra condition that seeks to facilitate factor updates in long narrow valleys of the objective function landscape, thus avoiding getting trapped in suboptimal solutions. The second example (FALCON-R), considered the importance of regularization on the success of the factorization models and adapted the regularization parameter automatically while training.
In a social rating network, recommendations for a user can be produced on the basis of the ratings of the users that have direct or indirect social relations with the given user. This approach is supported by sociological models [8], and their verification due to the increasing availability of online social network data [9]. The models propose that people tend to relate to other people with similar attributes, and due to the effects of social influence, related people in a social network, in turn, influence each other to become even more similar. Thorough surveys summarizing the various approaches proposed for social recommender systems in general, can be found in [10,11].
In this paper, we exploit this information within the FALCON framework and propose a matrix factorization algorithm for recommendation in social rating networks, called SocialFALCON. Here the additional information is incorporated as an extra condition that, during training, seeks to render the latent feature vector of each user as close as possible to the weighted average of the latent feature vectors of his direct neighbors, without compromising the need for decreasing the objective function. By achieving maximum possible alignment between the successive feature vectors updates of each user to that of his direct neighbors, the latent features of users indirectly connected in the social network also become dependent and hence social network influence gets propagated during training. An important benefit from this approach also arises for cold start users, that is, for users who have expressed only a few ratings. These users are more dependent on the propagation of social infuence compared to users with more ratings, and thus the SocialFALCON model, through the enforcement of the constraint that their latent features is close to those of their neighbors, can learn user feature vectors for this challenging category of users as well.
To evaluate the effectiveness of our approach in terms of prediction accuracy and computational efficiency, we present experimental results to compare its performance with the following baseline and state of the art algorithms for matrix factorization in recommender systems:
  • Regularized SVD (RegSVD): This method is the baseline matrix factorization approach which does not take into account the social network [12]. We compare our approach against this method in order to evaluate the improvement in performance induced by the social network information.
  • Probabilistic Matrix Factorization (PMF): This is an extension of the baseline matrix factorization which bounds the range of predictions by introducing non-linearities in the prediction rule [13]. Similarly to RegSVD we compare our approach against this method in order to evaluate the importance of utilizing social network information.
  • SVD++: This is also a matrix factorization approach which also does not take into account any social network information [14]. However SVD++ takes into account additional information in the factorization model in the form of implicit feedback for each user’s item preference history. We compare our approach against this method since it achieves state of the art performance in recommender system tasks against which other algotihms should be compared, and because it provides a paradigm for extending the baseline matrix factorization approach with additional information.
  • SocialMF: This is a very closely related algorithm to our proposed approach but utilizes the social network information in a different way in the factorization model [15].
Detailed derivations and descriptions of the above algorithms are presented in the next sections. To elucidate on the transparency of the results, we have implemented each algorithm from scratch in low-level C with no dependencies on additional libraries, and we’ve made the source codes publicly available with documentation and additional supporting materials from Github as described in Section 8.
Due to the lack of publicly available social rating network datasets, the performance of all algorithms was evaluated on the two standard social rating network datasets utilized and introduced respectively in [15], namely the existing Epinions dataset and a large scale dataset that was gained from Flixster.com. Again for transparency reasons in the reported results, we’ve made publicly available all the training/validation cross-validation splits that we utilized in our experiments as described in Section 7.
The rest of this paper is organized as follows: The problem definition of matrix factorization in recommender systems is presented in Section 2. SVD++, which we consider in our benchmarks as a state of the art extension of the basic matrix factorization model, and which utilizes implicit rating information, is discussed in Section 3. Section 4 discusses the recommendation problem in a social rating network setting and describes SocialMF which is the most closely related method to our approach. The general constrained optimization FALCON framework is outlined in Section 5. In Section 6 we derive the SocialFALCON algorithm from the FALCON framework. In the same section we also discuss some desirable properties of the algorithm, and its advantages over the SocialMF approach, especially with regards to computational complexity. The real life data sets used in our experiments are described in Section 7, and the experiments are reported in Section 8. Finally, in Section 9 conclusions are drawn and we outline some interesting directions for future work.

2. Matrix Factorization for Recommender Systems

In its basic form, matrix factorization for recommender systems describes both users and items by vectors of (latent) factors which are inferred from the user-by-item rating matrix. High correlation between user and item factors leads to a recommendation of an item to a particular user [16].
The idea behind matrix factorization is very simple. We define the corresponding matrix factorization R ^ K of the user-by-item ratings matrix R R N × M as
R ^ K = U T V
where U R K × N and V R K × M , in order to approximate all the unknown elements of R.
For the recommendation problem, R has many unknown elements, which cannot be treated as zero, and the application of formal Singular Value Decomposition (SVD), e.g., by Lanczos’ method [17] would provide erroneous results. Thus, for this case, the approximation task can be defined as follows:
Let r ^ u i denote how the u-th user would rate the i-th item, according to the factorization model, and e u i denote the error on the ( u , i ) -th known rating example (training pattern):
r ^ u i = k = 1 K u u k v k i
and
e u i = r u i r ^ u i .
In order to minimize this error (and consequently the error over all training patterns) we can apply the stochastic version of the gradient descent method on each ( 1 / 2 ) · e u i 2 to find a local minimum. Hence the elements of U and V can be updated as follows:
u u k = u u k + η · e u i · v k i , v k i = v k i + η · e u i · u u k
where η is the learning rate.
To better generalize on unseen examples, we can apply regularization with factors λ u and λ v for the user and item factors respectively, in order to prevent large weights:
u u k = u u k + η · ( e u i · v k i λ u · u u k ) , v k i = v k i + η · ( e u i · u u k λ v · v k i ) .
Thus, Equation (5) are the user and item factors updates when minimizing the following objective function by performing unconstrained stochastic gradient descent on U u and V i for all users u and all items i:
L = 1 2 u = 1 N i = 1 M I u i ( r u i U u T V i ) 2 + λ u 2 u = 1 N | | U u | | 2 + λ v 2 i = 1 M | | V i | | 2 .
In the above equation I u i is an indicator function that is equal to 1 if user u rated item i and equal to 0 otherwise, and “ | | . | | ” denotes the Euclidean norm. In this linear model the predictions are usually clipped to the [ m i n r , m a x r ] range appearing in the ratings matrix R.
It is also customary to adjust the predicted rating by accounting for user and item biases, that is, for systematic tendencies for some users to give higher ratings than others, and for some items to receive higher ratings than others. We can encapsulate these effects using baseline estimates [14] as follows:
Denote by μ the overall average rating. A baseline estimate for an unknown rating r ^ u i is denoted by b u i and accounts for the user and item effects:
b u i = μ + β u + γ i .
The vectors β R N and γ R M contain the biases for all users and items respectively, and the elements β u and γ i indicate the biases of user u and item i respectively. These can be also learned by gradient descent if we plugin into Equation (6) the baseline corrected expression for r ^ u i as:
r ^ u i = b u i + U u T V i = μ + β u + γ i + U u T V i .
In this case the updates for β u and γ i are given by:
β u = β u + η · ( e u i λ β · β u ) , γ i = γ i + η · ( e u i λ γ · γ i )
where λ β and λ γ are regularization parameters for the user and item biases respectively. This approach is usually refered to in the literature as Regularized SVD (RegSVD) [12].
Instead of using a simple linear model, which can make predictions outside of the range of valid rating values, the dot product between user and item specific feature vectors can be passed through the logistic function (sigmoid) s ( z ) = 1 / ( 1 + e x p ( z ) ) , which bounds the range of predictions within [ 0 , 1 ] . In this case the objective function is modified as follows:
L = 1 2 u = 1 N i = 1 M I u i ( r u i g ( θ u i ) ) 2 + λ u 2 u = 1 N | | U u | | 2 + λ v 2 i = 1 M | | V i | | 2 + λ β 2 | | β | | 2 + λ γ 2 | | γ | | 2
where
θ u i = β u + γ i + V i T U u
and g ( z ) = ( m a x r m i n r ) s ( z ) + m i n r . This model is usually refered to as Probabilistic Matrix Factorization (PMF) [13].
Usually the learning rate and the regularization values for each latent factor are determined by a search for optimal values that optimize the performance on a withheld validation set. There are various approaches on how this search is performed, but the most common approach is by grid search [18], where a predefined grid of candidates is set up and on each candidate a model is trained and evaluated on the validation set. It is obvious that such a scheme is very expensive in terms of runtime as many different models have to be trained and then evaluated, neverthess it is the dominating method for parameter selection in matrix factorization models.

3. SVD++

SVD++ [14] is a state of the art extension to the RegSVD model by taking into account implicit information. The most natural choice for implicit feedback is the item preference history, which tells us about the items for which the user has expressed a preference (either explicit or implicit). Thus, the idea is to add a second set of item factors, relating each item i to a factor vector Y i R K , and model the factor vector of each user as a function of all factor vectors Y i , for which the user has expressed a preference.
Following [12,14], SVD++ gives the predicted rating of user u on item i as follows:
r ^ u i = b u i + V i T ( U u + | R u | 1 2 j R u Y j )
where b u i is given by Equation (7) and the set R u contains the items rated by user u (for which we may or may not know the exact explicit rating). For example, in the Netflix challenge there was extra information in the qualifying set that users had rated certain movies but the actual value of each rating was withheld in order to be predicted by the contestants.
Therefore, in order to predict an unknown rating r ^ u i , as in (8), SVD++ maintains a vector of factors U u for each user but this vector is also complemented by the sum | R u | 1 2 j R u Y j that represents the combination of explicit and implicit feedback.
Plugging in Equation (12) into (6) and taking into account regularization parameters λ β and λ γ for the user and item biases as in Equation (9), we can learn U u , V i , and Y i by gradient descent updates as follows:
U u = U u + η · ( e u i · V i λ u U u ) ,
V i = V i + η · ( e u i · ( U u + | R u | 1 2 j R u Y j ) λ v V i ) ,
Y j = Y j + η · ( e u i · | R u | 1 2 V i λ y Y j ) , j R u ,
where λ y is the regularization parameter for the items for which the user has expressed a preference.

4. SocialMF

In a social rating network, each user u also has a set N u of other users (direct neighbors) with which she has a connection. We denote by T u v the value of social “trust” [15] that user u has on user v as a real number in [0, 1]. A value of 0 means that there is no trust and a value of 1 means full trust from u to v. In practice, in most social networks the trust values are binary with T u v = 0 meaning that user u does not “follow” user v, and T u v = 1 meaning that u “follows” v. The trust values can be stored in a matrix T = [ T u v ] N x N , where usually each row is normalized so that u = 1 N T u v = 1 . Note that the matrix T is asymmetric in general.
SocialMF [15] is a probabilistic model-based solution which incorporates the propagation of trust in the model, and which has been shown to improve both in terms of the quality of recommendations and speed over previously proposed approaches such as the Social Trust Ensemble (STE) proposed in [19]. The main idea behind SocialMF is that the latent feature vectors of a user should be made dependent on the feature vectors of his direct neighbors in the social network. Using this idea, latent features of users indirectly connected in the social network will be also dependent and hence the trust gets propagated. Thus, latent factors for all users and items are learned jointly from the ratings and the social graph. SocialMF therefore models the users’ feature vectors as:
U ^ u = v N u T u v U v
where U ^ u is the estimated latent feature vector of u given the feature vectors of his direct neighbors.
To this end, SocialMF extends the basic PMF model by considering the following objective function:
L ( R , T , U , V ) = 1 2 u = 1 N i = 1 M I u i ( r u i g ( U u T V i ) ) 2 + λ u 2 u = 1 N | | U u | | 2 + λ v 2 i = 1 M | | V i | | 2 + λ t 2 u = 1 N | | ( U u v N u T u v U v ) | | 2 ,
where g ( U u T V i ) is the predicted rating r ^ u i of user u on item i. As in PMF, λ u , λ v and λ t are regularization parameters and I u i is an indicator function that is equal to 1 if user u rated item i and equal to 0 otherwise.
We can find a local minimum of the objective function in Equation (17) by performing gradient descent on U u and V i for all users u and all items i:
L U u = i = 1 M I u i V i g ( U u T V i ) ( g ( U u T V i ) r u i ) + λ u U u + λ t ( U u v N u T u v U v ) v | u N v T v u ( U v w N v T v w U w ) ,
L V i = u = 1 N I u i U u g ( U u T V i ) ( g ( U u T V i ) r u i ) + λ v V i .
Note that the derivative of the objective function with respect to the item feature vectors V i is simply the same to that of the PMF model.
A variation of SocialMF has been proposed in [20] in which the authors train separate matrix factorization models for each category c that the items belong to, and for which a user u has issued a trust statement towards v, given that u and v simultaneously have ratings in a category c.

5. Overview of the Falcon Framework

Based on the discussion in the previous sections, it is clear that in all approaches matrix factorization in recommender systems involves minimization by unconstrained gradient descent of a suitably chosen objective function with respect to the user and item features and biases. In our approach we suppose that there are additional relations to be satisfied, that represent additional knowledge we wish to incorporate into the learning mechanism, and involve all the available features U u , V i , β u and γ i (u = 1 ⋯ N and i = 1 ⋯ M).
For the rest of our discussion it is convenient to group all the features into a single column vector w (which we will refer to as the “weight" vector) as follows:
w = ( β 1 β N , U 1 T U N T , γ 1 γ M , V 1 T V M T ) .
Before introducing the form of the additional relations, we note that in this work we’ll concentrate on the minimization of the Mean Squared Error (MSE) objective function given by:
L = 1 2 u = 1 N i = 1 M I u i ( r u i g ( θ u i ) ) 2 ,
with the quantities involved explained in Section 2. We will also adopt an epoch-by-epoch (i.e., batch) optimization framework with the following objectives:
  • At each epoch of the learning process, the vector w will be incremented by d w , so that the search for an optimum new point in the space of w is restricted to a hypersphere of known radius δ P centered at the point defined by the current w
    d w T d w = ( δ P ) 2 .
  • At each epoch, the objective function L must be decremented by a quantity δ Q , so that, at the end of learning, L is rendered as small as possible. To first order, we can substitute the change in L by its first differential and demand that
    d L = δ Q .
Within the FALCON framework originally introduced in [6], we can define a certain quantity Φ that we also wish to incrementally maximize at each epoch subject to the objectives defined above. Consequently the learning rule can be derived by solving the following constrained optimization problem:
Maximize Φ ( Φ = m a x ) w . r . t d w subject to d w T d w = ( δ P ) 2 d L = δ Q .
This constrained optimization problem can be solved analytically by a method similar to the constrained gradient ascent technique introduced in optimal control in [21], and leads to a generic update rule for w as follows:
First, we introduce suitable Lagrange multipliers λ 1 and λ 2 to take into account Equations (22) and (23) respectively. If δ P is small enough, the changes to Φ induced by changes in w can be approximated by the first differential d Φ . Thus, secondly, we introduce the function ϕ , whose differential is defined as
d ϕ = d Φ + λ 1 ( d L δ Q ) + λ 2 [ ( δ P ) 2 | | d w | | 2 ] .
On evaluating the differentials involved in the right hand side, we readily obtain
d ϕ = F · d w + λ 1 ( G · d w ) + λ 2 [ ( δ P ) 2 | | d w | | 2 ] ,
where G and F are given by
G = L / w , F = Φ / w .
To maximize d ϕ at each epoch, we demand that
d 2 ϕ = F λ 1 G 2 λ 2 d w · d 2 w = 0
and
d 3 ϕ = 2 λ 2 | | d 2 w | | 2 < 0 .
Hence, the factor multiplying d 2 w in Equation (27) should vanish, and therefore we obtain
d w = λ 1 2 λ 2 G + 1 2 λ 2 F .
Equation (29) constitutes the weight update rule, provided that λ 1 and λ 2 can be evaluated in terms of known quantities. This can be done as follows:
From Equations (22), (26) and (27) we obtain
λ 1 = I G F 2 λ 2 δ Q I G G
with I G G and I G F given by
I G G = | | G | | 2 , I G F = G · F .
It remains to evaluate λ 2 . To this end, we substitute (29) into (23) to obtain
4 λ 2 2 ( δ P ) 2 = I F F + λ 1 2 I G G 2 λ 1 I G F
where I F F is given by
I F F = | | F | | 2 .
Finally, we substitute (30) into (32) and solve for λ 2 to obtain
λ 2 = 1 2 I G G ( δ P ) 2 ( δ Q ) 2 I F F I G G I G F 2 1 / 2 .
Note that the positive square root value has been chosen for λ 2 in order to satisfy Equation (28).
Let us now discuss our choice for δ Q . This choice is dictated by the demand that the quantity under the square root in Equation (34) should be positive. It is easy to show that the term I F F I G G I G F 2 is always positive by the Cauchy–Schwarz inequality [5]. Now, since I G G = | | G | | 2 0 , it follows that care must be taken to ensure that I G G ( δ P ) 2 > ( δ Q ) 2 . The simplest way to achieve this is to select δ Q adaptively by setting δ Q = ξ δ P I G G with 0 < ξ < 1 . Consequently, the proposed generic weight update algorithm has two free parameters, namely δ P and ξ .

6. The SocialFALCON Algorithm

The most important aspect that we will now discuss is the definition of the quantity Φ that we wish to incrementally maximize at each epoch. Rather than adopting the approach of SocialMF which imposes a target condition on the latent feature vector of a user as the weighted average of the latent feature vectors of his direct neighbors, we will adopt and epoch-by-epoch approach so as to incrementally maximize at each epoch the alignment of the user’s feature vector update to the weighted average of the feature vectors updates of his direct neighbors at the immediately preceding epoch.
We thus introduce the quantity:
Φ = u = 1 N d β u t v N u T u v d β v t 1 + u = 1 N d U u t T v N u T u v d U v t 1
where d U u t is the user u feature vector update at the present epoch and v N u T u v d U v t 1 is the weighted average of the latent feature vector updates of his direct neighbors v at the immediately preceding epoch. Similarly, d β u t is the update of the bias of user u at the current epoch and v N u T u v d β v t 1 is the weighted average of the updates of the biases of the user’s direct neighbors at the immediately preceding epoch. Since within our constrained learning framework the whole weight vector w updates have constant moduli equal to δ P (by Equation (22)), maximization of Φ amounts to minimization of the angle between each user’s feature vector updates at the present epoch and the weighted average of the latent feature vector updates of his direct neighbors preceding that epoch.
Hence, for the factorization problem, at each epoch of the learning process we can restore from the weight vector w the user feature vectors U u using Equation (20), and update them according to Equations (29) and (26) as:
d U u = λ 1 2 λ 2 L U u + 1 2 λ 2 Φ U u = λ 1 2 λ 2 i = 1 M I u i · e u i · g ( θ u i ) · V i + 1 2 λ 2 v N u T u v d U v t 1 .
Similarly the user biases will be updated by:
d β u = λ 1 2 λ 2 i = 1 M I u i · e u i · g ( θ u i ) + 1 2 λ 2 v N u T u v d β v t 1 .
As far as the updates of the item feature vectors V i and item biases γ i are concerned, there are two possible alternatives. The first is to simply update them by:
d V i = λ 1 2 λ 2 L V i a n d d γ i = λ 1 2 λ 2 L γ i
since there are no dependencies of Φ as defined in Equation (35) from any V i or γ i .
An even more interesting option for updating the item feature vectors and biases is to supplement the quantity Φ with a term involving an incremental maximization target that we seek to achieve for each V i and γ i on an epoch-by-epoch basis. To this end, we adopt the approach that we have proposed in algorithm FALCON-M as described in [6]. According to that approach, the quantity that we wish to maximize is the alignment between the item feature vector updates at the present and immediately preceding epoch. Due to the constant moduli of the updates of the whole vector w (Equation (22), this also amounts to minimization of the angle between the vectors of the successive weight updates for each V i , and thus can suppress zig-zagging and allow learning to proceed along relatively smooth paths. In a sense this is equivalent to adding a momentum term in the updates of each V i (and γ i ), with the important difference that both the coefficients of the gradient (i.e., learning rate) and momentum are suitably adapted at each epoch of the learning process by the constained update rule of Equation (29).
Note that for this approach to work we should redefine Φ as:
Φ = u = 1 N d β u t v N u T u v d β v t 1 + u = 1 N d U u t T v N u T u v d U v t 1 + i = 1 M d γ i t d γ i t 1 + i = 1 M d V i t T d V i t 1
where d V i t and d V i t 1 are the item i feature vector updates at the present and immediately preceding epoch respectively (similarly for the item i bias updates d γ i t and d γ i t 1 ).
Again we can restore from the weight vector w the item feature vectors V i using Equation (20), and update them according to Equations (29) and (26) as:
d V i = λ 1 2 λ 2 L V i + 1 2 λ 2 Φ V i = λ 1 2 λ 2 u = 1 N I u i · e u i · g ( θ u i ) · U u + 1 2 λ 2 d V i t 1 .
Similarly the updates for the item biases will be given by:
d γ i = λ 1 2 λ 2 u = 1 N I u i · e u i · g ( θ u i ) + 1 2 λ 2 d γ i t 1 .
Equations (36), (40), (37), and (41) thus constitute our proposed SocialFALCON algorithm for updating the user features, item features, user biases, and item biases respectively.

6.1. Desirable Properties of SocialFALCON

As is the case with SocialMF, the SocialFALCON model makes the feature vector of each user to be dependent on the feature vectors of his direct neighbors. Since those neighbors are in turn connected to other users, recursively, indirect connections in the social network propagate their social infuence across the entire network. In addition, even for users that have expressed no ratings (cold start users) but have social connections, there still remains a social update term in their factor update rule, which means that latent features of those users will at least adapt towards those of their neighbors. Therefore, despite not having any expressed ratings, feature vectors for these users will be learned as well.
The SocialFALCON model has three further desirable properties. First, despite its seemingly complex derivation, the factor update rule for both users and items (and their biases) turns out to be quite simple since it only constitutes from a term proportional to the gradient of the objective function of Equation (21), and an extra term proportional to the gradient of Φ as given by Equation (39) with respect to each feature vector. Both of these terms are quite easy to compute as it can be seen from Equations (36), (37), (40) and (41). For the case of user factors and their biases, the extra term is a social term adapting the update towards the weighted average of the factors’ updates of the direct neighbors at the immediately preceding epoch, whereas for the item factors and biases the extra term acts as momentum.
Second, all the factor update rules can also be viewed as standard gradient descent with regularization. This is enforced by the hard constraint of Equation (22) that restricts the norm of the weight vector update d w to a hypersphere (which means that the compoments of the vector w cannot grow in an uncontrollable fashion), and can be more easily seen if we rewrite, for example, Equation (40) as:
d V i = λ 1 2 λ 2 u = 1 N I u i · e u i · g ( θ u i ) · U u + 1 2 λ 2 d V i t 1 = λ 1 2 λ 2 u = 1 N I u i · e u i · g ( θ u i ) · U u + 1 2 λ 2 ( V i V i t 1 ) .
The same argument applies, by expansion of the weighted average in the social term, to user factor updates. An important benefit here is that both the coefficient multiplying the gradient as well as the reguralization coefficient are automatically adapted at each epoch by the current values of λ 1 and λ 2 .
Finally, the algorithm has only two free parameters, namely δ P and ξ which, as we’ll discuss in the experimental results section, are not very sensitive to the exact setting of their values.

6.2. Complexity Analysis of SocialFALCON

The main overhead to complete an epoch is in evaluating the gradient of the objective function with respect to the feature vectors of users and items. As it can been seen from Equation (21), the number of operations per feature vector needed to complete this evaluation is proportional to the number of rating examples in the training set. In addition we need to calculate the gradient of Φ as given by Equation (39) with respect to the feature vectors. Following the notation used in [15], we assume that the average number of ratings per user is r ¯ , and the average number of direct neighbors per user is t ¯ . As reported in the same paper, the computational complexity of computing the gradients of SocialMF’s objective function L with respect to the number of users in the social rating network is O ( N r ¯ K + N t ¯ 2 K ) . The t ¯ 2 factor is justified if we take a closer look at Equation (18) which computes the derivative of Equation (17) with respect to each user feature vector U u so as to update it by gradient descent. In order to evaluate this derivative, for each user u, one needs to take the following steps: (a) Find his direct neighbors and aggregate their feature vectors, (b) find the users v in the social network to whichu is a direct neighbor, and (c) v find their direct neighbors w and aggregate their feature vectors.
In the SocialFALCON model the updates of the user feature vectors are given by Equation (36). As it can be readily seen from that equation, the evaluation of its second term only requires step (a) as described above. This makes the total omputational complexity of evaluating the gradients of (21) and (39) with respect to the number of users O ( N r ¯ K + N t ¯ K ) . Therefore SocialFALCON is r ¯ + t ¯ 2 r ¯ + t ¯ times faster than SocialMF in computing the gradient in each training epoch as it scales linearly with the number of user connections. Since usually the rating matrix R and trust matrix T are very sparse, t ¯ and r ¯ are relatively small, and therefore both SocialMF and SocialFALCON scale linearly with respect to the total number of users in the social rating network. However in a large social network where the number of average number of direct neighbors per user is large, the speedup factor provided by SocialFALCON becomes profound and crucially important in the ability to effectively train a recommender.
We should note here that once the gradient has been evaluated, a relatively small number of additional operations as given by Equations (31), (33), (34) and (30) (which is independent of the number of training examples) is needed to complete the update. In addition the updates of all user and item factors at the immediately preceding epoch should be stored since they are utilized by the SocialFALCON update rule. This additional computational burden however is very small compared to the calculation of the gradient since it just involves the evaluation of three inner vector products as given by Equations (31) and (33). The sizes of these vectors are equal to the size of vector w (Equation (20)), thus the number of operations involved is ( K + 1 ) ( N + M ) . This is confirmed by the actual CPU times measured in our experimental results.

7. Datasets

The availability of publicly available social rating network datasets is extremely limited, presumably due to the sensitive nature of social network data and the proprietary rights of social networks. To the best of our knowledge there are only very few public social rating network datasets: The Epinions.com dataset, and a dataset that was crawled from Flixster.com and has been made available by the authors of [15].
The Epinions dataset (http://www.trustlet.org/wiki/Downloaded_Epinions_dataset) that we used has 49,289 users, 664,824 ratings and 487,183 connections between users. Possible rating values are discrete integers in the range (1,5). The average number of ratings per user is 16.5 and each user has on average 14.3 direct neighbors. The social relations in the dataset are directed.
The Flixster dataset (http://www.cs.sfu.ca/$\sim$sja25/personal/datasets) dataset has 787,213 users, 8,196,077 ratings and 7,058,819 relations between users. Each user, on average, has rated 55.5 items and has 48.6 social relations. Even though the dataset was crawled by [15] as a directed network, the social relations are undirected. This means that either duplicates should be removed or be handled appropriately in the code (we opted for the later choice). Possible rating values are ten discrete numbers in the range (0.5,5), and in our implementation we scaled rating values by 2 to be integers in the range (1,10).
In order to evaluate the performance of all algorithms considered in the next section, we used five-fold cross-validation (CV). We used stratified sampling on user ratings so that, in each fold, 80% of the ratings of each user were chosen randomly and used for training and the remaining 20% of user’s ratings were used for evaluation. Due to this sampling scheme, users with only one rating in the dataset were inevitably excluded from the evaluation set. We have made publicly available (http://labs.fme.aegean.gr/ideal/socialfalcon-datasets/) both the database schema as well as the five-fold CV training/validation splits that we’ve used in the experiments.

8. Experimental Results

We compared the proposed SocialFALCON algorithm with four other factorization models: RegSVD and PMF as described in Section 2, SVD++ as described in Section 3 and SocialMF as described in Section 4. All algorithms were implemented in low-level C and were compiled with the GNU gcc 4.6.3 compiler. We should point out that the codebase for RegSVD, PMF and SVD++ has been carefully developed both in terms of minimizing their runtime and increasing their prediction accuracy, due to their prior utilization by one of the authors as member in “The Ensemble” team which was a runner up for the Netflix Prize competition (http://netflixprize.com/leaderboard). For the code implementation of SVD++, it should be noted that a naive implementation of gradient descent is very ineffective, because if the implicit items are updated at each training example then the training process will be very slow. In our implementation we have used the “looping trick” proposed in [22] for NSVD1, which is a predecessor of SVD++ originally introduced by [12]. This modification produces exactly the same results as the naive implementation of gradient descent, however it makes a significant difference in running time, especially for large scale problems.
To the best of our knowledge the authors of [15] have not provided any source code implementation of SocialMF. The only known publicly available code implementation of SocialMF is within MyMediaLite (http://www.mymedialite.net/), which is a lightweight, multi-purpose library of recommender system algorithms. An important bug fix to that code implementation has also been provided by one of the authors of the present paper (http://mymedialite.net/download/Changes), which was contributed while developing our own implementation. Nonetheless, the codes for all algorithms presented in this section have been rewritten from scratch and have been made publicly available from Github (https://github.com/namp/SocialFALCON). The experiments were run on a machine with an Intel Core2 Quad CPU (Q9400 @2.66GHz) with 4GB RAM, running Ubuntu 12.04 LTS.
The prediction error was measured in RMSE on the evaluation set of each split defined as:
R M S E = 1 | E | r u i E ( r u i r ^ u i ) 2
where | E | is the size of the evaluation set E, r u i is the actual rating and r ^ u i is the prediction.
Table 1 summarises the RMSE of RegSVD, PMF, SVD++, SocialMF and the proposed SocialFALCON algorithm on the Epinions and Flixster datasets. We report results for two choices of latent dimensions, specifically K = 5 and K = 10 as was also reported in [15], where it was shown that increasing K in Epinions did not improve the results, while increasing K in Flixster improved the results slightly. For all algorithms, each RMSE result is reported along with two numbers: The average number of epochs required to achieve it and the average time (in seconds) spent at each epoch. The RMSE numbers reported are the average of the five-fold CV splits. Note that for the Flixster dataset the average reported RMSE for all algorithms has been divided by 2 in order to compensate for reporting the error on the original rating values in the range (0.5,5). For fairness in the comparison, for all algorithms, each run was initialized from a common random seed, and their relevant parameters were carefully adjusted to achieve the best possible performance. These parameters were found by utilising grid search and are reported in Appendix A.
For the Epinions dataset, as highlighted in bold, SocialFALCON improves the RMSE of RegSVD, PMF, SVD++, and SocialMF on both choices of latent dimensions. In both cases, we notice that the RMSE of SocialFALCON is 0.6% lower compared to that of RegSVD, 0.62% lower than the RMSE of PMF, and 0.55% lower compared to the RMSE of SVD++. Empirical data on recommender systems research have shown that such levels of improvement are indeed significant [23]. For example during the Netflix Prize competition, yearly progress prizes of $50,000 US were offered for the best improvement of at least 1% over the previous year’s result [24]. The closely related SocialMF method seems to perform only marginally better than RegSVD and PMF, and marginally worse than SVD++. Naturally, as explained in the beginning of this section, the codes for RegSVD, PMF, and SVD++ have been very carefully implemented due to their utilization in the Netflix prize and thus their performance is expected to be quite good. However the same argument is also true for the code implementation of SocialMF. In addition all RMSEs are reported with optimal parameters found by grid search, and to this end, it remains unclear why we have not been able to confirm in our evaluation the much improved performance of SocialMF over PMF on the Epinions dataset as reported in [15]. A possible explanation for this is that in their experiments they used a different version of the Epinions dataset and, of course, our CV splits are different. Another explanation for the poor performance they reported for PMF is perhaps due to poor code implementation. Still, the results that we obtained by SocialMF for the Epinions dataset are better than those reported on their paper.
For the much larger Flixster dataset, for both choices of latent dimensions, SocialFALCON outperforms PMF and SocialMF, and performs as good as RegSVD (which apparently performed remarkably well on this task considering its model simplicity). SVD++ performs marginally better than SocialFALCON and RegSVD, which can be attributed to the large number of implicit ratings in the evaluation sets. In general, the RMSE results reported for all algorithms are better than the results for Epinions. An explanation for this has been given in [15] by noticing that the items in Epinions are heterogeneous (belonging to categories such as DVD players, cameras, printers, laptops, etc.), while the items in Flixster are from a single caterory, namely movies. Possibly this makes the rating signals quite more accurate than those of Epinions, and thus makes the factorization task easier in general. Despite the large number of social connections in the dataset, apparently SocialMF provided the worst RMSE score among all the algorithms tested. In this case also we have been unable to verify the claims of significant RMSE improvement over PMF for the Flixster dataset and, in general, the SocialMF’s results reported in [15]. It is also interesting to note how much SocialFALCON appears to improve the RMSE over SocialMF, which is 3.4% for K = 5 and 3.15% for K = 10 .
Overall, the proposed SocialFALCON algorithm performs well and it is competitive to state of the art matrix factorisation methods. More important however, is that it can effectively handle large datasets. This is supported by the other two numbers, apart from the RMSE, that are shown in each cell of Table 1. SocialFALCON and SocialMF are batch methods, and hence they require a much larger number of epochs than stochastic algorithms such as RegSVD, PMF and SVD++. The reason is that batch methods only have one chance to update the factors after each presentation of the whole training set, whereas stochastic methods make thousands of updates in one epoch, up to the N × M size of the user-by-item ratings matrix. This explains the significant difference, for example, between PMF and SocialFALCON or SocialMF in the average number of epochs to achieve the reported RMSE score. We should mention that SocialFALCON’s average number of epochs, as reported for the various tasks in Table 1, are higher than should be needed in practice. This is because its best scores were reported for small stepsizes δ P discovered by the grid search, as shown in Appendix A, even thought quite comparable RMSE scores where attained with larger stepsizes and hence fewer epochs. Similar performances were recorded with 0.5 < δ P < 2.0 and 0.7 < ξ < 0.9 , indicating that results are not very sensitive to the exact values of the parameters. Following the example of [25], we performed additional runs of SocialFALCON using common values ( δ P = 1.0 ) and ( ξ = 0.85 ) on the two datasets. Deterioration of the RMSE compared to the scores reported with the optimal parameters shown in Table 1, was in both cases never more than 2%, even though for the case of Flixster there was more than 30% reduction on the required number of epochs. The larger scale problem was more sensitive to the selection of δ P , whereas it was not so sensitive to the selection of ξ (for which a value around 0.85 worked well in both cases).
The third number shown in each cell of Table 1, reports the average time required to complete an epoch and is quite informative as it is a measure of the computational complexity of each algorithm. The numbers show that the complexity/epoch of the SocialFALCON algorithm is not significantly higher than that of RegSVD and PMF and even SVD++. This is true in spite of the fact that our implementation of SVD++ is very efficient as we have utilised the “looping-trick" mentioned in the beginning of this section, which significantly reduces SVD++’s computational burden. Note also that according to our discussion in Section 6.2, SocialFALCON should theoretically be r ¯ + t ¯ 2 r ¯ + t ¯ times faster than SocialMF. Since the Flixster data set is denser than the Epinions data set, the improvement over runtime efficiency for Flixster should be more prominent than that for Epinions. As we can see from Table 1, indeed SocialMF is much slower than SocialFALCON. The epoch of SocialFALCON is 2.9 times faster than that of SocialMF for Epinions and 14.5 times faster for Flixster.

Cold Start Users

Based on the findings of the previous section where, in general, the performance of SocialFALCON compared to SocialMF was significantly better, we also investigated the performance of these two closely related methods only on cold start users. As reported in [15], in both Flixster and Epinions more than 50% of users are cold start users and thus the efficiency of recommendations on this challenging class of users becomes very important. As cold start users we define users that have less than five ratings in each dataset and at least two, since users with only one rating were not included in the evaluation set of each fold. To this end, we created a subset of the evaluation set of each CV fold which contained only cold start users.
Table 2 summarises the RMSE of SocialMF and the proposed SocialFALCON algorithm for cold start users. For the Epinions dataset, as highlighted in bold, SocialFALCON improves the RMSE of SocialMF on both choices of latent dimensions. For the Flixster dataset, in both cases (especially for K = 10 ), SocialMF performs marginally better than SocialFALCON. On this dataset which has denser social relations, this can be attributed to the hard preset target distance metric incorporated into the SocialMF factorization model that, at the end of learning, the latent feature vector of each user is rendered equal to the weighted average of the latent feature vectors of his direct neighbors. On the other hand SocialFALCON seeks to minimize that distance incrementally by aligning the corresponding updates so as to render the latent feature vector of each user as close as possible to the weighted average of the latent feature vectors of his direct neighbors at the end of learning. Thus, for cold start users, social connections alone, might play a slightly more dominant role for providing recommendations in SocialMF than in SocialFALCON, which comes of course with a higher computational cost as was discuseed in the previous sections.

9. Conclusions

In this paper, we proposed an efficient constrained matrix factorization algorithm called SocialFALCON, for providing recommendations in social rating networks. The algorithm derives from the FALCON generic constrained matrix factorization which has been previously proposed by authors of this paper. The FALCON framework allows the incorporation of additional knowledge into the learning mechanism for determining the user and item factor matrices. In the case of SocialFALCON, this additional knowledge is embedded into mathematical constraints that, during learning, drive the feature vector of each user to be dependent on the feature vectors of his direct neighbors in the social network. Similarly to related proposed approaches, this allows the propagation of social influence within the factorization model, which has been shown to be an important factor in the social sciences, in social network analysis and in trust-based recommendations. The propagation of social infuence also allows for providing recommendations to cold start users that have not expressed many ratings, since their feature vectors can be learned through their social relations. However, unlike similar approaches the proposed algorithm has reduced computational complexity, can be implemented easily, and thus be utilized more frequently in social recommendation setups.
Experimental results on two publicly available datasets showed that the algorithm improves on baseline and state of the art factorization methods as well as on previously proposed related approaches in terms of convergence speed and recommendation accuracy. To elucidate on the transparency of our results, we have made publicly available the source code of the proposed algorithm, as well as the codes of the algorithms against which it was compared, along with the datasets used in the experiments.
One of the most attractive features of this algorithm is its potential for suggesting several interesting directions for further improvements. In the same framework for constrained matrix factorization, it is possible to augment it with further information about learning in social rating networks, including methods for embedding graph theory measures and indices into the social trust values and incorporating implicit direct neighbor ratings as utilized in SVD++. It is the concerted incorporation of such detailed information into the same algorithm that will hopefully lead to increasingly efficient matrix factorization training schemes combining fast learning, good scalability properties, and powerful generalization capabilities on predicted ratings for unseen items.

Author Contributions

Conceptualization, N.A.; Methodology, N.A. and T.E.; Software, N.A. and T.E.; Validation, T.E.; Writing–original draft, N.A. and T.E.; Writing–review & editing, F.S.

Funding

F.S. is supported by a PhD Scholarship by the State Scholarships Foundation (IKY), Greece.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Algorithms’ Parameters

In this section we list the learning parameters (discovered by grid search) with which the algorithms achieved their best performance on each dataset, as presented in the results tables (Table 1 for all users and Table 2 for cold start users only).
In order to boost the performance of RegSVD, PMF, SVD++ and SocialMF we used different learning rates for the user features and biases ( η u and η b u respectively) and for the item features and biases ( η v and η b v respectively). SVD++ also requires two further parameters for the training of the implicit item factors, namely η y and λ y which are the learning rate and regularisation parameter respectively.
SocialFALCON has only two free parameters, namely δ P and ξ .
  • RegSVD (Epinions): η u =0.003, η v =0.00005, η b u =0.003, η b v =0.003, λ u =0.015, λ v =0.1, λ β =0.015, λ γ =0.015
  • RegSVD (Flixster): η u =0.02, η v =0.005, η b u =0.02, η b v =0.005, λ u =0.1, λ v =0.1, λ β =0.01, λ γ =0.01
  • PMF (Epinions): η u =0.005, η v =0.005, η b u =0.005, η b v =0.005, λ u =0.1, λ v =0.1, λ β =0.01, λ γ =0.01
  • PMF (Flixster): η u =0.015, η v =0.005, η b u =0.015, η b v =0.005, λ u =0.01, λ v =0.01, λ β =0.01, λ γ =0.01
  • SVD++ (Epinions): η u =0.003, η v =0.003, η b u =0.003, η b v =0.003, η y =0.003, λ u =0.015, λ v =0.5, λ β =0.015, λ γ =0.015, λ y =1.0
  • SVD++ (Flixster): η u =0.003, η v =0.003, η b u =0.003, η b v =0.003, η y =0.003, λ u =0.015, λ v =0.5, λ β =0.015, λ γ =0.015, λ y =1.0
  • SocialMF (Epinions): η u =0.0001, η v =0.0001, η b u =0.0001, η b v =0.0001, λ u =0.1, λ v =0.1, λ β =0.1, λ γ =0.1, λ t =0.1
  • SocialMF (Flixster): η u =0.0001, η v =0.0001, η b u =0.0001, η b v =0.0001, λ u =0.1, λ v =0.1, λ β =0.1, λ γ =0.1, λ t =0.1
  • SocialFALCON (Epinions): δ P =1.0, ξ =0.8
  • SocialFALCON (Flixster): δ P =0.3, ξ =0.95

References

  1. Takács, G.; Pilászy, I.; Németh, B.; Tikk, D. On the Gravity Recommendation System. In Proceedings of the 13th ACM International Conference on Knowledge Discovery and Data Mining, San Jose, CA, USA, 12–15 August 2007; pp. 22–30. [Google Scholar]
  2. Koren, Y. The BellKor Solution to the Netflix Grand Prize. Available online: http://www.netflixprize.com/assets/GrandPrize2009_BPC_BellKor.pdf (accessed on 11 August 2019).
  3. Toscher, A.; Jahrer, M.; Bell, R. The Big Chaos Solution to the Netflix Grand Prize. Available online: http://www.netflixprize.com/assets/GrandPrize2009_BPC_BigChaos.pdf (accessed on 11 August 2019).
  4. Piotte, M.; Chabbert, M. The Pragmatic Theory Solution to the Netflix Grand Prize. Available online: http://www.netflixprize.com/assets/GrandPrize2009_BPC_PragmaticTheory.pdf (accessed on 11 August 2019).
  5. Perantonis, S.J.; Ampazis, N.; Virvilis, V. A Learning Framework for Neural Networks Using Constrained Optimization Methods. Ann. Oper. Res. 2000, 99, 385–401. [Google Scholar] [CrossRef]
  6. Ampazis, N.; Emmanouilidis, T. FALCON: A matrix factorization framework for recommender systems using constrained optimization. Intell. Decis. Technol. 2015, 9, 221–232. [Google Scholar] [CrossRef]
  7. He, X.; Liao, L.; Zhang, H.; Nie, L.; Hu, X.; Chua, T.S. Neural Collaborative Filtering. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 April 2017; pp. 173–182. [Google Scholar] [CrossRef]
  8. Wasserman, S.; Faust, K. Social Network Analysis: Methods and Applications; Cambridge University Press: Cambridge, UK, 1994; Volume 8. [Google Scholar]
  9. Crandall, D.; Cosley, D.; Huttenlocher, D.; Kleinberg, J.; Suri, S. Feedback Effects Between Similarity and Social Influence in Online Communities. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, NV, USA, 24–27 August 2008; pp. 160–168. [Google Scholar]
  10. Tang, J.; Hu, X.; Liu, H. Social recommendation: A review. Soc. Netw. Anal. Min. 2013, 3, 1113–1133. [Google Scholar] [CrossRef]
  11. Yang, X.; Guo, Y.; Liu, Y.; Steck, H. A survey of collaborative filtering based social recommender systems. Comput. Commun. 2014, 41, 1–10. [Google Scholar] [CrossRef]
  12. Paterek, A. Improving regularized singular value decomposition for collaborative filtering. In Proceedings of the KDD Cup and Workshop, San Jose, CA, USA, 12 August 2007. [Google Scholar]
  13. Salakhutdinov, R.; Mnih, A. Probabilistic Matrix Factorization. In Proceedings of the 20th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 3–6 December 2007; pp. 1257–1264. [Google Scholar]
  14. Koren, Y. Factorization meets the neighborhood: A multifaceted collaborative filtering model. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, NV, USA, 24–27 August 2008; ACM: New York, NY, USA, 2008; pp. 426–434. [Google Scholar]
  15. Jamali, M.; Ester, M. A matrix factorization technique with trust propagation for recommendation in social networks. In Proceedings of the Fourth ACM Conference on Recommender Systems, Barcelona, Spain, 26–30 September 2010; ACM: New York, NY, USA, 2010; pp. 135–142. [Google Scholar]
  16. Koren, Y.; Bell, R.; Volinsky, C. Matrix Factorization Techniques for Recommender Systems. Computer 2009, 42, 30–37. [Google Scholar] [CrossRef]
  17. Cullum, J.; Willoughby, R. Lanczos Algorithms for Large Symmetric Eigenvalue Computations: Volume 1, Theory; Classics in Applied Mathematics, Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 2002. [Google Scholar]
  18. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. Acm Trans. Intell. Syst. Technol. 2011, 2, 27:1–27:27. [Google Scholar] [CrossRef]
  19. Ma, H.; King, I.; Lyu, M.R. Learning to recommend with social trust ensemble. In Proceedings of the 32nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Boston, MA, USA, 19–23 July 2009; ACM: New York, NY, USA, 2009; pp. 203–210. [Google Scholar]
  20. Yang, X.; Steck, H.; Liu, Y. Circle-based Recommendation in Online Social Networks. In Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Beijing, China, 12–16 August 2012; ACM: New York, NY, USA, 2012; pp. 1267–1275. [Google Scholar] [CrossRef]
  21. Bryson, A.E.; Denham, W.F. A Steepest-Ascent Method for Solving optimum Programming Problems. J. Appl. Mech. 1962, 29, 247–257. [Google Scholar] [CrossRef]
  22. Takács, G.; Pilászy, I.; Németh, B.; Tikk, D. A unified approach of factor models and neighbor based methods for large recommender systems. In Proceedings of the 2008 First International Conference on the Applications of Digital Information and Web Technologies (ICADIWT), Ostrava, Czech Republic, 4–6 August 2008; pp. 186–191. [Google Scholar]
  23. Feuerverger, A.; He, Y.; Khatri, S. Statistical Significance of the Netflix Challenge. Stat. Sci. 2012, 27, 202–231. [Google Scholar] [CrossRef]
  24. Bennett, J.; Lanning, S. The Netflix Prize. In Proceedings of the KDD-Cup and Workshop at the 13th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Jose, CA, USA, 12–15 August 2007. [Google Scholar]
  25. Jacobs, R.A. Increased rates of convergence through learning rate adaptation. Neural Netw. 1988, 1, 295–307. [Google Scholar] [CrossRef]
Table 1. Experimental results for Epinions and Flixster. In each table cell we report the triplet: RMSE/average number of epochs/time spent in each epoch (in secs).
Table 1. Experimental results for Epinions and Flixster. In each table cell we report the triplet: RMSE/average number of epochs/time spent in each epoch (in secs).
Features (K)AlgorithmEpinionsFlixster
5RegSVD1.049009861.60.0649460.8577111166.42.1827098
PMF1.0492854540.09630920.86856714.42.3057142
SVD++1.0485256620.1452210.85430181442.9196802
SocialMF1.0488033180.61.01239980.885620997955.8514706
SocialFALCON1.0428028216.60.31948680.85544561040.23.8049924
10RegSVD1.049000261.60.09414020.8483767187.82.2727028
PMF1.04924653.80.13747660.859007215.22.8283334
SVD++1.048397261.20.22867220.845011180.44.134036
SocialMF1.04875363179.61.68760020.8760801983.672.0185296
SocialFALCON1.0427142216.20.64288720.848409410825.0869388
Table 2. Experimental results for Epinions and Flixster, cold start users only.
Table 2. Experimental results for Epinions and Flixster, cold start users only.
FeaturesAlgorithmEpinionsFlixster
5SocialMF1.12650161.0904493
SocialFalcon1.11010221.1067439
10SocialMF1.1266481.1094216
SocialFalcon1.11023821.109602

Share and Cite

MDPI and ACS Style

Ampazis, N.; Emmanouilidis, T.; Sakketou, F. A Matrix Factorization Algorithm for Efficient Recommendations in Social Rating Networks Using Constrained Optimization. Mach. Learn. Knowl. Extr. 2019, 1, 928-944. https://doi.org/10.3390/make1030053

AMA Style

Ampazis N, Emmanouilidis T, Sakketou F. A Matrix Factorization Algorithm for Efficient Recommendations in Social Rating Networks Using Constrained Optimization. Machine Learning and Knowledge Extraction. 2019; 1(3):928-944. https://doi.org/10.3390/make1030053

Chicago/Turabian Style

Ampazis, Nicholas, Theodoros Emmanouilidis, and Flora Sakketou. 2019. "A Matrix Factorization Algorithm for Efficient Recommendations in Social Rating Networks Using Constrained Optimization" Machine Learning and Knowledge Extraction 1, no. 3: 928-944. https://doi.org/10.3390/make1030053

Article Metrics

Back to TopTop