Next Article in Journal
A Coalition Formation Game Approach for Efficient Cooperative Multi-UAV Deployment
Previous Article in Journal
The Optical Barcode Detection and Recognition Method Based on Visible Light Communication Using Machine Learning
 
 
Correction published on 19 July 2021, see Appl. Sci. 2021, 11(14), 6633.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Attention-Based Recommender System to Predict Contextual Intent Based on Choice Histories across and within Sessions

1
School of Computer Science, Beijing University of Posts & Telecommunications, Beijing 100876, China
2
Leavey School of Business, Santa Clara University, Santa Clara, CA 95053, USA
*
Author to whom correspondence should be addressed.
Appl. Sci. 2018, 8(12), 2426; https://doi.org/10.3390/app8122426
Submission received: 31 October 2018 / Revised: 23 November 2018 / Accepted: 26 November 2018 / Published: 29 November 2018

Abstract

:
Recent years have witnessed the growth of recommender systems, with the help of deep learning techniques. Recurrent Neural Networks (RNNs) play an increasingly vital role in various session-based recommender systems, since they use the user’s sequential history to build a comprehensive user profile, which helps improve the recommendation. However, a problem arises regarding how to be aware of the variation in the user’s contextual preference, especially the short-term intent in the near future, and make the best use of it to produce a precise recommendation at the start of a session. We propose a novel approach named Attention-based Short-term and Long-term Model (ASLM), to improve the next-item recommendation, by using an attention-based RNNs integrating both the user’s short-term intent and the long-term preference at the same time with a two-layer network. The experimental study on three real-world datasets and two sub-datasets demonstrates that, compared with other state-of-the-art methods, the proposed approach can significantly improve the next-item recommendation, especially at the start of sessions. As a result, our proposed approach is capable of coping with the cold-start problem at the beginning of each session.

1. Introduction

In recent years, we have witnessed the rapid growth of neural networks for a variety of applications in computer vision, natural language processing, speech recognition, etc. Neural network provides four key benefits: (1) It works with unlabeled data; (2) It learns low-level features from minimally processed raw data; (3) It detects complex interactions among features; and (4) It works with large class memberships [1,2]. Recent achievements have shown that it is also of great use to combine different neural networks into recommender systems. There are many works which employed neural networks to improve the recommendation in the domain of recommender systems [3,4,5,6,7] and in the context of web-based decisions [8,9,10,11,12]. The success of these recommendation approaches based on neural networks demonstrates its capability.
Session-based recommender systems are good examples. Although not a novel research topic, session-based recommender systems are largely under-investigated [3,4,13]. Compared with traditional recommendation methods, a session-based recommender system is more appropriate for capturing dynamic and sequential user behavior. The Recurrent Neural Network (RNN) plays an important role in session modeling, and it seems to have promise for further improvement. An RNN, compared to many other recommendation models, takes advantage of the ordered sequence in a very natural way [3,5,7].
Automated and personalized recommendations of the type “You may also be interested in” are common on e-commerce sites nowadays, and there exists sufficient evidence that such item related recommendations can measurably impact businesses [14,15,16,17,18,19]. Undoubtedly, one of the representative business scenarios is in e-commerce. Modern online shops are not static catalogs of items anymore but tend to fulfil the user’s interests by providing personalized product recommendations. In the best case, these recommendations should match both the users’ long-term preference as well as their current shopping goals. On Amazon.com, for example, each product page contains multiple personalized recommendation lists with different purposes. Though the Amazon recommender system employs collaborative filtering approach [20], it also uses solo-user recommendations as well. It provides alternatives to the product currently viewed, remind the user of items recently viewed, or render recommendations that should appeal to the general taste of the user, as shown in Figure 1.
As a result, the content displayed on the page not only depends on the features of the currently viewed article (i.e., the product category), but can also be influenced by an integration of the user’s historical shopping behavior (long-term preference) and his most recent interactions of viewing different products (short-term shopping goals). In this case, we assume that user’s preference is composed of two components: (1) the long-term preference which reflects the considerably stable interests of the users based on their online activities, and (2) the short-term intent which represents the users’ current interests. There are many existing techniques that investigate the combination of the user’s long-term preference and the short-term intent [13,21,22,23,24,25,26,27]. In addition, the user’s preference for items constantly evolves over time [28], while these works implicitly assume the user preference to be stationary, which seems to be an unrealistic assumption in many scenarios, particularly in the news or shopping related scenarios [5,6,7,13,29]. Therefore, such methods perform poorly when preferences, in fact, are context-sensitive and transient. There are session-based approaches [7,29,30,31,32] that focus on the interest drifts, to solve the problem of the static user preference and intent, while these methods assume that each item in one session has the same influence, which is usually an invalid assumption, especially when considering a user’s interactions with items of diverse characteristics. One of the main attractions of search-engine advertising is that search engines take into account the “current interest” of the user and thereby can deliver the right message, ad, or product just at the right time. The enormous growth in spending going to search advertising (about $100b in the US in 2018 [33]) is a testament to the power of contextual “interest-based” advertising. To deal with the problem of the static weights being assigned to each item, researchers proposed several approaches [34,35,36,37]. Our proposed model goes a step further, not only taking the user’s interest drift into consideration, but also regarding each item in one session as “the one and only” item, which means each one has a unique influence on the user and therefore the user may pay different attention to each one.
In this work, we propose a novel approach named Attention-based Short-term and Long-term Model (ASLM), to solve the next-item recommendation problem. It has been shown that the attention mechanism is able to assign different weights according to the importance of the items, and there have been works that employed the attention mechanism to form the characteristics of the user [38,39,40]. In order to capture the variation in user preference and generate the precise contextual user profile, we build two different RNN-based layers: one attentive layer to model the user’s short-term intent, and the other one to model his long-term preference. The main contributions of this paper include:
  • We introduce the attention mechanism to investigate the variation in a user’s short-term intent for session-based recommendations.
  • Taking the user’s previous sessions into account, we model the user’s long-term preference and concatenate that with short-term intent to account for more comprehensive contextual user outcomes.
  • We conduct empirical evaluations and validate the effectiveness of the model on three real-world datasets and two sub-datasets.
  • We demonstrate the effectiveness of the approach in resolving the cold-start problem.

2. Related Work

Collaborative Filtering (CF) is the most commonly used and studied technology [20,41]. CF allows users to give ratings about a set of elements (e.g., videos, songs, films, etc. in a CF-based website) in such a way that when enough information is stored on the system, we can make recommendations to each user based on information provided by other users we consider to have the most in common with them [42]. The most commonly used algorithm for CF is the k Nearest Neighbors (kNN) [41,43,44], which offers several mechanisms to identify similar users and pool recommendations. The other main approach for CF is matrix factorization (MF) [42,45], which regards recommendation as a dimensionality reduction problem. However, these methods are far from researchers’ satisfaction with several major challenges:
  • Resulting from the high level of sparsity [46,47] in recommender system databases, user similarity measures often encounter processing and cold-start problems from inadequate mutual ratings for a comparison of users and items [48,49,50].
  • The CF-based recommendation lacks diversity, because it only recommends popular items from similar users but cannot recommend unique tastes of a user.
  • When numbers of existing users and items grow tremendously, traditional CF algorithms will suffer serious scalability problems, with computational resources going beyond practical or acceptable levels [45].
  • Most importantly, they are not qualified to capture the temporal context and sequential patterns of user behavior (e.g., the evolving user long-term preference and short-term intent) [51]. In other words, when given the data which does not have ratings involved but pure user-item interaction (or choice) logs, CF-based methods may not work.
To solve the problem of dealing with the sequential patterns of user behavior and temporal aspects of recommendations, there are a variety of works that focus on session-based recommendation which features sequence modeling. In academia, sequential recommendation problems are typically regarded as the task of predicting the next user action. Experimental evaluations are usually based on larger, time-ordered logs of user actions (e.g., the users’ item viewing and purchasing activities on an e-commerce shop), which contains both the implicit and explicit feedback from users and provides the recommender system the capability of capturing the sequential patterns of users. Zimdars et al. [25] framed CF as a sequence prediction problem and used a simple Markov model for web-page recommendation. Mobasher et al. [26] proposed a similar approach, using sequential pattern mining. Both showed the superiority of methods based on sequence over nearest-neighbor approaches. Park et al. [27] presented a modified User-based CF method, called Session-Based CF (SSCF), that uses information from similar sessions to capture sequence and repetitiveness in the listening process. Most recently, researchers have successfully adopted deep learning techniques in recommender systems. In particular, RNN, which is capable of learning models from sequentially ordered data, is a "natural choice" for sequence modeling [4]. Hidasi et al. [13] first introduced the idea of using RNN for session-based recommendation to deal with sparsity problems and proposed the GRU4REC architecture. Tan et al. [21] improved the model performance proposed in [13] by proposing the data augmentation and the method to account for shifts in the input data distribution. Jannach et al. [22] also improved the model proposed in [13] by proposing a model combining the kNN approach with GRU4REC. Dong et al. [23] improved the RNNs [52] and the MF model [53] by combining them in a multi-task learning framework, where they performed joint optimization with shared model parameters enforcing the two parts to regularize each other. Liang et al. [24] employed RNN to extract the item and user profiles from user-defined items’ tags and their tagging behaviors. However, these session-based methods regard user preference for items as a stationary property and therefore cannot capture the changing contextual interactions between user and items.
To solve the problem of the static user preference and intent, Jing et al. [30] focused on the variation in user intent at different times and proposed an LSTM-based “Just-In-Time” approach to recommend the right item at the right time. Wu et al. [31] proposed an LSTM-based approach that implicitly captures various known temporal patterns in movie ratings data without explicit inclusion in the model, to learn dynamic embeddings of users and movies. The next-basket recommendation model proposed in [32] learned the varying representation of a user and captured global sequential features among baskets to reflect the user’s dynamic intent at different times and interactions of all baskets of the user over time. Ruocco et al. [7] improved the next-item recommendation and cold-start problem by implementing two RNNs to extract user preference from recent sessions and the current session, respectively. Hu et al. [29] designed an efficient personalized session-based recommender system with shallow wide-in-wide-out networks over relaxed ordered user-session contexts. These methods assume that each item in one session has the same influence. However, this assumption is usually invalid, especially when considering a user’s interactions with items of diverse characteristics. In the meantime, not only his preference for items changes among different sessions, but also his intent on items within one session.
To deal with the problem of the static weights being assigned to each item, Cheng et al. [34] developed an attention-based model to capture the changing attention that a user pays to different text reviews and ratings. Wang et al. [35] designed an attention-based transaction embedding model to weight each item in a transaction without assuming order. To solve these two issues at the same time in a unified way, we follow the idea in [7] but make the following contributions: (1) We propose an Attention-based Short-term Intent Layer to assign different weights for items in the current session to capture the contextual property of user intent. (2) We propose a Long-term Preference Layer to take the user’s previous sessions into account, to provide information to the attention-based short-term intent layer and enable it to combine both the user’s short-term intent and long-term preference in the attention process.

3. The ASLM Architecture

In this section, we first formulize the description of the problem as explained above and the main idea of Attention-based Short-term and Long-term Model (ASLM) proposed in this paper. Afterwards, we elaborate by diagram the architecture of ASLM.

3.1. Problem Formulation

Let U denote a set of users and V denote a set of items, where | U | and | V | are the total numbers of users and items, respectively. In this work, we concentrate on extracting the user contextual intent and preference from sequential user-item interaction events (e.g., commenting to a thread or listening to music). For each user u U , the sequential sessions are denoted as S u = { S 1 u , S 2 u , , S T u } , where T is the total number of time steps and S t u S u ( t [ 1 , T ] ) represents the item set related to the interaction of user u at time step t. In each session, there is a set of events { e t , i u R m | i = 1 , 2 , , | S t u | } , where e t , i u represents the event i in the session S t u . For each event, the user u interacts with an item v i V . Both S t u and e t , i u S t u are ordered by time sequence. For a time step t, the session S t u implies the u’s short-term intent at time t, while the sessions before time step t, defined as L t 1 u = S 1 u S 2 u S t 1 u , implies user u’s long-term preference. As in [7], our task is to predict each successive item in a session S t u . That is, for a sub-session { e t , 1 u , e t , 2 u , , e t , i u , , e t , n u } of S t u , the system is to predict e t , i + 1 u . This is repeated for i = 1 , 2 , , | S t u | 1 . A recommendation R i is a rank ordered list of k recommended items. The goal of the system is to make the next item, e t , i + 1 u , be located as close to the top of the ranked list as possible.

3.2. Main Idea

Consider the following example of a user shopping scenario: On May 15th, John purchased a printer and a smartphone. On July 10th, when the ink cartridge ran out, he wanted another to replace it. At that point in time, he would have been happy to see an ad for any ink cartridge and would have bought another brand if the system had recommended it. Also, since John was in the purchasing process, he might also have been attracted to a cross-sell recommendation of a phone-protector case. However, he would have been much less interested, in July, about recommendations for another brand-new printer or smartphone.
In the ASLM model, there are two layers: the Long-term Preference Layer and the Attention-based Short-term Intent Layer. Our model is illustrated in Figure 2, where two mutually exclusive methods are shown in one figure to conserve space: (a) ASLM-AP, and (b) ASLM-LHS. In other words, only one method is used separately at a time, and has no connection between each other. The two ASLM models only differ by how they feed information to the Attention-based Short-term Intent Layer, either using average pooling (ASLM-AP) or the last hidden state (ASLM-LHS). Based on the experiment results discussed in Section 5, researchers can feel free to use either of these two models since their performances are at the same level.
The task of the Long-term Preference Layer is to supply knowledge of the user’s long-term preference to the Attention-based Short-term Intent Layer before the short-term intent layer investigates the current session at the beginning. The task of the Attention-based Short-term Intent Layer is to produce recommendations by taking the user’s long-term preference into consideration when the short-term layer is processing the sequence of items in the current session to extract his short-term intent.
We discuss the work flow of these two layers in detail in Section 3.3. Particularly, the Attention-based Short-term Intent Layer’s procedure of capturing the characteristics of each step in the user behavior is thoroughly discussed at the part of decoder in Section 3.3.2. Our model is proposed according to the following characteristics of user preference. (1) User preference changes at different time steps. (2) Within one time step, the user may have another intent on a different category of items, which has a different impact on the future interaction of items. (3) Since there are different preferences across users, the importance of the same item may differ, which results in different influences on the prediction of the next item for different individuals. The main idea of our approach is to:
(1)
use the GRU (Gated Recurrent Unit) [13] in the Long-term Preference Layer to process a user’s recent sessions, i.e., a user’s long-term preference, which is similar to the Inter-session RNN in [7].
(2)
enhance the recommendations by employing an augmented attention-based encoder-decoder in a structured framework named Attention-based Short-term Intent Layer. We use the encoder to process the events within a session, i.e., a user’s short-term intent, and make the decoder receive the user’s long-term preference representation P L u from the user’s history session representations as its initial hidden state. As a result, the model enables the Attention-based Short-term Intent Layer to combine both the user’s long-term preference and short-term intent in the attention process. Eventually, for ASLM-LHS, the decoder stores its output state in session representations for the Long-term Preference Layer to update the states for future recommendation. For ASLM-AP, we put the average of current session’s embedded vector representations of the items into session representations.

3.3. Model Description

In this section, we discuss the ASLM model’s entire recommendation process and the main system components with the adopted technologies we introduced.
As illustrated in Figure 2, ASLM integrates the modeling of the user’s Long-term and augmented Attention-based Short-term Intent Layers into one single architecture. The Long-term Preference Layer is evolved from the Inter-session RNN used in [7], and the model proposed in it is regarded as a baseline to compare with ASLM model.

3.3.1. Long-Term Preference Layer

For the short-term intent layer, it learns about the user’s intent throughout the session, but discards all the information at the end of the session. Similar to the model proposed in [7], we employ a GRU to take the user’s k most recent history sessions into account. At the beginning of each session, the GRU models the user’s long-term preference representation P L u from the user u’s history session representations and provides it for the GRU-based decoder in the attention-based short-term preference layer as the initial hidden state.
The graph inside the dark green box in Figure 2 illustrates the work flow of the Long-term Preference Layer. For each session S t u in u’s interaction history S u , let s t u be an embedded vector representation of that session. The Long-term Preference Layer takes a series of vector representation of u’s k most recent history sessions { s t k u , s t k + 1 u , , s t 1 u } as input, where s t 1 u is the most recent history session. Afterwards, the layer produces the initial hidden state, H 0 , of the Attention-based Short-term Intent Layer before the short-term intent layer starts making predictions. As a result, it enables the Attention-based Short-term Intent Layer to combine both the user’s short-term intent and long-term preference at the attention process.

3.3.2. Attention-Based Short-Term Intent Layer

For predicting next items, a user’s short-term preference is important. A user’s short-term preference consists of different kinds of short-term intent on different categories of items. Previously, Rendle et al. [54] has done some research on combining long- and short-term preference for sequential recommendation. However, in the earlier works, researchers regard the user’s short-term preference as a stationary property, and therefore items are given the same weights and the variation in short-term intent is not evaluated properly. To resolve this problem, we employ an attention-based encoder-decoder architecture, to assign weights to both long- and short-term session representations, to form the characteristics of user u comprehensively. In addition, we employ bi-directional LSTM and GRU as the encoder and decoder, respectively. One advantage of bi-directional LSTM over uni-directional LSTM is that it uses each event e t , i u S t u to predict recommendations based on each event’s past and future contexts, rather than just its past ones, which can be sensitive to the variation in the user’s short-term intent from both sides. Attention mechanism has been widely and successfully applied in many tasks, e.g., information retrieval [38], computer vision [39] and recommendation [40]. The key idea of attention is to learn to assign attentive weights (normalized by sum to 1) for a set of features: higher weights indicate that the corresponding features are informative for the task. It computes the importance of each item both in the short-term and long-term item sets of a given user and integrates these items’ embeddings to build the representations of the user’s intent and preference.
The graph inside the purple box in Figure 2 illustrates the work flow of the Attention-based Short-term Intent Layer. In the following paragraphs, we state the functionality and the working procedure of each component in the Attention-based Short-term Intent Layer.
Encoder. A bi-directional LSTM is employed as the encoder. We use item embeddings to represent each item. At the beginning of each session, each event e t , i u which consists of the embedded item representation is sent through a bi-directional LSTM to be processed together with its past and future contexts, to investigate the sequence patterns of the current session. Afterwards, the bi-directional LSTM produces its output vector, namely the source hidden states vector [ h ¯ 1 h ¯ 2 h ¯ i ] where h ¯ i is a source hidden state regarding the item v i , and includes the latent information of v i . Finally, they are sent to the architecture of attention mechanism.
Decoder. A GRU is employed as the decoder. The procedure of capturing the characteristics of each step in the user behavior is stated below. At the beginning of each session, the GRU receives the user’s long-term preference representation P L u from the Long-term Preference Layer as the initial hidden state. In the meantime, the GRU takes the context vector c j produced by the Attention procedure into consideration. In other words, c j provides the distribution of every event e t , i u ’s influence upon the current target item prediction in the current session, from which it enables the GRU to find out which events are more important to the current choice. Apparently, the more important events are highly related to the user u’s current intent. Therefore, when the time step t increases and the previously important events become less important, the u’s intent probably has changed. The GRU is ready for capturing the variation of the user u’s current intent. In other words, our proposed model is aware of the variation of the user u’s short-term intent. Moreover, it makes the recommendation with the help of not only the short-term intent, but also the long-term preference. Finally, the GRU produces the output vector, namely the target hidden states vector [ h 1 h 2 h j ] where h j is a target hidden state regarding the current target item prediction y j . The target hidden states vector, together with the context vector c j stated below, is used for the calculation of the target item prediction.
Architecture of Attention Mechanism. The main part of the attention mechanism is illustrated in Figure 3. The context vector c j is used to capture relevant source-side information to help provide the current target item prediction y j . Given the target hidden state h j and the context vector c j calculated in Equation (2), we use a concatenation to integrate both vectors to produce an attentional hidden state:
a j = f ( c j , h j ) = t a n h ( W c [ c j ; h j ] ) ,
where
c j = i α j i h ¯ i ,
where W c is the model parameter, and c j is computed by the sum of the product of each attentive weight α j i and source hidden state h ¯ i . A dense layer is used to fully connect the neurons in the encoder and the decoder. α j i is the amount of attention that the output prediction of h j should pay to h ¯ i , which can be calculated by the SoftMax function:
α j i = e x p ( α j i ) i e x p ( α j i ) ,
where α j i is used to compare the target hidden state h j with each of the source hidden state h ¯ i :
α j i = h j W h ¯ i
Denote s t u as the embedded vector representation of the session S t u . After the calculation of the attentional vector a j , it is fed through another SoftMax layer to produce the predictive distribution. Given the predictions y < j = { y 1 , y 2 , , y j 1 } , s t u and the weight W i of a j , the predictive distribution of items is formulated as:
p ( y j | y < j , s t u ) = s o f t m a x ( W i a j )
We apply two methods to generate the session representations s t u similar to that in [7]. One is to generate s t u with the average of the embedded vector representations of the items in the session, as illustrated in Part Method (a): ASLM-AP of Figure 2. The other is to use the last hidden state of attention-based short-term preference layer as the session representation, illustrated in Part Method (b): ASLM-LHS of Figure 2. On one hand, although there is more useful information for the Attention-based Short-term Intent Layer to learn from the final hidden state, it is more a representation of the end of the session, rather than the entire session. On the other hand, since the average of the embeddings is irrelevant to the sequence order of the items, it can represent the entire session.

4. Experimental Setup

We conduct experiments divided into two groups with different datasets to investigate the effectiveness of our model: (1) Group A with abundant user interactions for each user: Reddit and Last.fm datasets. (2) Group B with scarcely enough user interactions for each user: 1/10 Reddit, 1/80 Last.fm and Tmall dataset.

4.1. Datasets

We experimented with three different datasets and two sub-datasets: Reddit (Subreddit dataset: https://www.kaggle.com/colemaclean/subreddit-interactions), Last.fm (Last.fm dataset: https://blog.csdn.net/hopygreat/article/details/96444827), Tmall (Tmall dataset: https://tianchi.aliyun.com/datalab/dataSet.html?spm=5176.100073.0.0.40236fc1kIb4f8&dataId=42), 1/10 Reddit and 1/80 Last.fm. The descriptive statistics for the datasets are shown in Table 1. The distribution histograms of the number of sessions by the frequency of users for Reddit, Last.fm and Tmall datasets are illustrated in Figure 4 and Figure 5. Moreover, another set of distribution histograms, namely the one of the number of events by the frequency of sessions for Reddit, Last.fm and Tmall datasets, is given in Figure 6 and Figure 7. Table 2 and Table 3 show the standard deviation and the skewness (We used the skew function provided by SciPy to calculate the skewness. See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.skew.html) of them. To conserve space, only the standard deviation and the skewness but not the histograms of the two distributions for 1/10 Reddit and 1/80 Last.fm datasets are given in Table 2 and Table 3.
Reddit dataset. Reddit contains a log of user interactions on different subreddits (sub-forums) at different timestamps. Since the dataset does not split the interactions into sessions, we did this by setting a time limit for inactivity and integrate the actions that occurred within one hour into a session as in [7]. Also, as in [7], we replicated their model against ours on a training set of 80% and holdout set of 20% using the same Recall and MRR (Mean Reciprocal Rank) metrics as they did [7]. We discuss the two evaluation metrics in Section 4.3. Our results replicate theirs (not shown here to minimize space). Additionally, for a deeper assessment as advocated in [55], we split the sessions into a training set (60%), a validation set (20%) and a test set (20%) to establish the performance of the alternative models on an identical basis.
Last.fm dataset. For Last.fm dataset, we split logs of every user into sessions the same way as that for Reddit. Instead, we chose a 30-minute limit as in [7]. The same procedure for assessment was used on Last.fm dataset as stated above. Afterwards, we split the sessions similar to that for Reddit dataset.
1/10 Reddit dataset. For the 1/10 Reddit dataset, we select 10% random samples from the original Reddit dataset. The same procedure for assessment was used on 1/10 Reddit dataset as stated above.
1/80 Last.fm dataset. For the 1/80 Last.fm dataset, we select 1/80 random samples from the original Last.fm dataset. The same procedure for assessment was used on 1/80 Last.fm dataset as stated above.
Tmall dataset. The Tmall dataset contains anonymized users’ shopping logs in the past 6 months before and on the “Double 11” day (11 November). We only kept the purchase data. We chose a one-day limit as in [29]. Only the training set (80%) and the test set (20%) are used because of the lack of enough user interactions for each user in Tmall dataset.

4.2. Baselines

We compare our model with the following baseline algorithms.
Most Popular. This baseline always recommends the absolute most popular items in the training set. All items are sorted by their number of occurrences in the training set, and the top K items are recommended at each time step. In other words, the same items, which are the most popular ones among all the users, are recommended to each user. Although it is a simple baseline, it is usually a strong baseline especially in recommendation domains.
II-RNN-AP [7]. This baseline uses two RNNs to extract information from the user’s current session and history sessions, respectively. II-RNN-AP is the II-RNN Model with average pooling to create session representations from items. The model’s configurations are the same with that in [7].
II-RNN-LHS [7]. II-RNN-LHS is the II-RNN Model where the last hidden state of the Intra-session RNN is stored as the session representation. We use the same configurations in [7].
SWIWO [29]. This full model is a three-layer network that predicts the relevant item based on the user-session context. We use the same configurations in [29].
SWIWO-I [29]. This baseline is the simplified SWIWO model which only models item-session contexts without considering users. We use the same configurations in [29].

4.3. Evaluation Metrics and Hyperparameters

To evaluate the performance of each method, we employ two commonly used metrics Recall@K and MRR@K with K = 5, 10, 20 to evaluate all models. The first metric evaluates the fraction of ground truth items that have been rightly ranked over top-K items in all testing sessions, while the second metric, MRR, is a statistic measure for evaluating any process that produces a list of possible responses to a sample of queries, ordered by probability of correctness. The reciprocal rank of a query response is the multiplicative inverse of the rank of the first correct answer. For hyperparameters, we experimented with different ones to find the best configurations of our method for each dataset. The configurations of our model are summarized in Table 4.

5. Results

In this section, the model performances of the proposed models are shown, according to the experimental settings we stated in the previous section.

Model Performance

Table 5, Table 6 and Table 7 show two groups’ test set performances of all methods under the metric of Recall@K and MRR@K: (1) For Reddit and Last.fm, we set K = 5, 10 and 20. (2) For 1/10 Reddit and 1/80 Last.fm, we set the same K as the full versions of Reddit and Last.fm. For Tmall, to keep consistent with the evaluation in [29], we use Recall@10, Recall@20 and MRR@20 to evaluate methods, and cited the same evaluation results of SWIWO and SWIWO-I in [29]. Relative scores are given compared to the strongest baselines. For all datasets, ASLM-LHS outperforms all baselines with moderate margins. Table 8 shows the validation set performances of three different methods under the metric of Recall@K and MRR@K in Reddit and Last.fm datasets. Please note that, for the Reddit dataset, the validation set performance of ASLM-LHS is better than that of ASLM-AP, and, for Reddit and Last.fm datasets, the validation set performances of ASLM-AP and ASLM-LHS are all better than the test set performances with moderate margins, respectively. Table 9 shows the validation set performances of three different methods in 1/10 Reddit and 1/80 Last.fm datasets. The validation set performances of the two ASLM models in the 1/10 Reddit dataset are all better than the test set performances, while, in the 1/80 Last.fm dataset, the majority of the validation set performances of the two ASLM models are better than the test set performances.

6. Discussion

In this section, we discuss the superiority of the proposed models through comparison of performance and demonstrate their effectiveness on the session cold-start problem.

6.1. Comparison of Performance

For Group A, in Table 5, we observe that ASLM-LHS consistently outperforms all the baselines under all measurements for testing cases in both Reddit and Last.fm datasets with moderate margins; In most cases, ASLM-AP outperforms all the baselines in both datasets, except in Recall@10 and Recall@20 scores in the Reddit test set. Specifically, ASLM-LHS improves 9.12% and 22.10% in Recall@5 and MRR@5 scores compared with the II-RNN-LHS method for test cases in the Reddit dataset, respectively. ASLM-AP improves 4.36% and 7.33% in Recall@5 and MRR@5 compared with the II-RNN-AP method for test cases in the Last.fm dataset, respectively. The reason ASLM-AP’s Recall@10 and Recall@20 scores are lower than those of II-RNN-LHS in the Reddit dataset is that the average pooling method only utilizes the average of the embedding of each event e t , i u S t u as the representation of the current session S t u , and the important context information, e.g., the sequential patterns and the user’s intent captured by the attention mechanism, will be lost when this representation is stored in the user’s long-term session representations. It results in the fact that the abundant context information cannot be fully utilized by ASLM-AP, and thus its performance declined. However, the performance of ASLM-AP is still better than II-RNN-AP, which also utilizes the average pooling method. The good performance of ASLM-LHS is attributed to the attention mechanism and the bidirectional-LSTM we employed in the attention-based layer and the last hidden state method. One of the main characteristics of the attention mechanism is that it calculates the importance of each given event and captures the user’s short-term intent with it. The bidirectional-LSTM can extract the sequential patterns from each given event in both the forward and the backward directions. Both ASLM-LHS and II-RNN-LHS utilize the last hidden state method which retains the context information from the current session and regards it as one part of the user’s long-term session representations, compared with the average pooling method. Therefore, the model’s subsequent training process benefits from this method by receiving the valuable context information from history sessions. In addition, in Table 1 we observe that, in the Reddit dataset, the number of sessions per user (namely, 62.15) and the average number of events in a session (namely, 3.00) are much smaller than those in the Last.fm dataset (namely, 645.62 and 8.10), which shows that ASLM-LHS can perform better when the user’s history information is less adequate. Table 5 and Table 8 show the results that the performance of ASLM for validation cases and that for testing cases in Reddit and Last.fm datasets are at the same level, which assure us the validity of our model.
For Group B, similar results can be seen in Table 6 and Table 7 that ASLM-LHS and ASLM-AP outperform all the baselines under all measurements in 1/10 Reddit, 1/80 Last.fm, and Tmall datasets. For the Tmall dataset, the evaluation result of ASLM-LHS shown in Table 7 improves 70.86%, 49.06%, and 114.50% in R@10, R@20, and MRR@20 scores compared with the SWIWO-I method, respectively. Please note that, for the Tmall dataset, as shown in Table 1, there is scarcely enough sessions for each user (namely, 3.99 per user) and number of average session length (2.99, per user) compared with those in Reddit and Last.fm. For 1/10 Reddit and 1/80 Last.fm, since the user history information is reduced dramatically compared with the full Reddit and Last.fm datasets, most of the models’ performances significantly decrease accordingly. However, both ASLM models still outperform the strongest baselines, as shown in Table 6. It demonstrates that ASLM can reach better performance even when it is severely short of both long-term and short-term user behavior.
When K changes from 5 to 20, for the Reddit and 1/10 Reddit datasets, there is a downward trend in the relative scores of ASLM-LHS and ASLM-AP. For the Last.fm dataset, there is an upward trend in those of ASLM-LHS. The reason for the upward trend is that there is more abundant user history information in the Last.fm dataset, and, with the increase of K, ASLM-LHS still has the potential to further capture the context information implied by the Last.fm dataset, which results in some space to improve its performance.
In terms of the two versions of ASLM (ASLM-AP using average pooling and ASLM-LHS using the last hidden state), Table 5, Table 6 and Table 7 show that, for the Reddit dataset, the test performance of ASLM-LHS has an obvious advantage compared to that of ASLM-AP, while they are at the same level for the other four datasets. As mentioned above, the model’s subsequent training process benefits from the last hidden state method by receiving the valuable context information from history sessions.

6.2. Effectiveness on Session Cold-Start Problem

In addition to the overall recommendations we reported in Section 5 and Section 6.1. We evaluate the effectiveness of ASLM-LHS on recommendations for the first n time steps, for n = 1, …,5, L, where L is the maximum session length and L = 19 in our experiment.
Figure 8 shows the comparison of performances of Recall@5 at the first n recommendations of each session for Reddit and Last.fm datasets between ASLM and II-RNN.
Notice that, for Reddit and Last.fm, where abundant user history behavior is given, our attention-based ASLM-LHS model achieves R @ 5 > 0.3 and R @ 5 > 0.099 when n = 1 , respectively, and, under most circumstances, each R @ 5 score of the ASLM-LHS model increases throughout the session. ASLM-LHS has an advantage over II-RNN-LHS and II-RNN-AP at the start of a new session in Reddit and Last.fm datasets, with the improvements of 7.00% and 3.12%, respectively. When more events have been investigated in the current session, for the Reddit dataset, the advantage becomes more obvious, while, for the Last.fm dataset, since the performances of ASLM-LHS and II-RNN-AP are identical when n = 2 , the advantage becomes greater when n increases from 3.
Similar results are shown in Figure 9 for the Tmall dataset. In terms of the Tmall dataset, which is severely short of both long-term and short-term user behavior, ASLM-LHS still provides 0.5063 in R @ 5 with just a slight difference compared with the overall R @ 5 (i.e., 0.5074). Due to the lack of history information from each user in the Tmall dataset, its performance is not stable at the start of a session when n is between 1 and 3, however it gets better when n increases from 3. As a result, the attention-based short-term intent layer can thoroughly reveal the variation in the user’s short-term intent between events from the current session S t u . To sum up, ASLM significantly alleviates the cold-start problem.

6.3. Implications of the Work

This system is designed to address situations where only past decisions of the users are available (e.g., there are no ratings by the users of the items). This is a very common situation, and yet recommendations are still needed and can be produced by estimating the interest drift of user preferences from past choices across items, and the pattern and sequence of past choices across sessions and particularly in the current session.
The ASLM can be applied by recommender systems in the scenarios of news, music, and e-commerce (e.g., Reddit, Last.fm and Tmall, respectively) where the user’s interest drift is the main topic since it takes place frequently. The model depends on the data availability because it is a data-driven method, and continuously requires the user’s session data to process when running. During the experiment, it took about 28 min for the ASLM to run every epoch on Reddit dataset which contains 1,135,488 user sessions with a NVIDIA GTX 1080 Ti graphic card. In other words, it took about 1.5 ms for each session, and that is appropriate for real-time recommendation. Therefore, the system can be employed online and executed in real-time, although we conducted the experiment offline.

7. Conclusions

In this work, we proposed an attention-based RNNs named ASLM for next-item recommendation problems. By integrating both the user’s short-term intent and long-term preference at the same time with a two-layer network, our model makes the best use of the awareness of the variation in the user’s contextual preference, especially the short-term intent, to cope with the cold-start problem within sessions. We have conducted experiments on three real-world datasets and two sub-datasets and demonstrated considerable improvements over strong baselines. Moreover, our proposed model can provide accurate recommendation at the start of sessions, which remarkably mitigates the cold-start problem.

Author Contributions

Conceptualization, R.H., S.M., M.S. and H.E.; Methodology, R.H.; Software, R.H.; Validation, R.H., S.M., M.S., H.E. and Z.O.; Formal Analysis, R.H.; Investigation, R.H.; Resources, R.H., S.M., M.S. and H.E.; Data Curation, R.H.; Writing—Original Draft Preparation, R.H.; Writing—Review & Editing, R.H., S.M., M.S., H.E. and Z.O.; Visualization, R.H.; Supervision, S.M., M.S., H.E. and Z.O.; Project Administration, Z.O.; Funding Acquisition, Z.O.

Funding

This research was funded by National Natural Science Foundation of China (Grant No. 61702046).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript (in alphabetical order):
APAverage Pooling
ASLMAttention-based Short-term and Long-term Model
GRUGated Recurrent Unit
II-RNNInter- and Intra-session Recurrent Neural Network
LHSLast Hidden State
LSTMLong Short-Term Memory
RNNRecurrent Neural Network
SWIWOSession-based Wide-In-Wide-Out

References

  1. Deep Learning: A Brief Guide for Practical Problem Solvers. Available online: https://www.infoworld.com/article/3003315/big-data/deep-learning-a-brief-guide-for-practical-problem-solvers.html (accessed on 28 November 2018).
  2. Karthik, S.; Paul, A.; Karthikeyan, N. Deep Learning Innovations and Their Convergence with Big Data; IGI Global: Hershey, PA, USA, 2017. [Google Scholar]
  3. Zhang, S.; Yao, L.; Sun, A. Deep learning based recommender system: A survey and new perspectives. arXiv, 2017; arXiv:1707.07435. [Google Scholar]
  4. Ludewig, M.; Jannach, D. Evaluation of Session-based Recommendation Algorithms. arXiv, 2018; arXiv:1803.09587. [Google Scholar]
  5. Quadrana, M.; Cremonesi, P.; Jannach, D. Sequence-aware recommender systems. arXiv, 2018; arXiv:1802.08452. [Google Scholar]
  6. Quadrana, M.; Karatzoglou, A.; Hidasi, B.; Cremonesi, P. Personalizing session-based recommendations with hierarchical recurrent neural networks. In Proceedings of the 11th ACM Conference on Recommender Systems, Como, Italy, 27–31 August 2017; pp. 130–137. [Google Scholar]
  7. Ruocco, M.; Skrede, O.S.L.; Langseth, H. Inter-Session Modeling for Session-Based Recommendation. In Proceedings of the 2nd Workshop on Deep Learning for Recommender Systems, Como, Italy, 27–31 August 2017; pp. 24–31. [Google Scholar]
  8. Kraus, M.; Feuerriegel, S. Decision support from financial disclosures with deep neural networks and transfer learning. Decis. Support Syst. 2017, 104, 38–48. [Google Scholar] [CrossRef] [Green Version]
  9. Bountris, P.; Kotronoulas, G.; Tagaris, T.; Haritou, M.; Spathis, A.; Karakitsos, P.; Koutsouris, D. CxCaDSS: A web-based clinical decision support system for cervical cancer. In Proceedings of the 6th European Conference of the International Federation for Medical and Biological Engineering, Dubrovnik, Croatia, 7–11 September 2014; Volume 45, pp. 757–760. [Google Scholar]
  10. Zeng, Y.; Wang, L.; Chen, T.; Lu, Y. A hybird intelligent decision support system for spare parts inventory control using neural network and gene algorithms approach. In Proceedings of the IASTED International Conference on Power, Energy, and Applications, PEA 2006: Science and Technology for Development in the 21st Century, Gaborone, Botswana, 11–13 September 2006; pp. 32–36. [Google Scholar]
  11. Lai, K.K.; Yu, L.; Wang, S. A neural network and web-based decision support system for forex forecasting and trading. In Data Mining and Knowledge Management; Springer: Berlin, Germany, 2005; pp. 243–253. [Google Scholar]
  12. Zavadskas, E.K.; Kaklauskas, A.; Gulbinas, A. Multiple criteria decision support web-based system for building refurbishment. J. Civ. Eng. Manag. 2004, 10, 77–85. [Google Scholar] [CrossRef]
  13. Hidasi, B.; Karatzoglou, A.; Baltrunas, L.; Tikk, D. Session-based Recommendations with Recurrent Neural Networks. In Proceedings of the International Conference on Learning Representations, San Juan, Puerto Rico, 2–4 May 2016; pp. 1–10. [Google Scholar]
  14. Cremonesi, P.; Garzotto, F.; Turrin, R. Investigating the persuasion potential of recommender systems from a quality perspective: An empirical study. ACM Trans. Interact. Intell. Syst. 2012, 2, 11. [Google Scholar] [CrossRef]
  15. Garcin, F.; Faltings, B.; Donatsch, O.; Alazzawi, A.; Bruttin, C.; Huber, A. Offline and online evaluation of news recommender systems at swissinfo.ch. In Proceedings of the 8th ACM Conference on Recommender Systems, Foster City, CA, USA, 6–10 October 2014; pp. 169–176. [Google Scholar]
  16. Gomez-Uribe, C.A.; Hunt, N. The netflix recommender system: Algorithms, business value, and innovation. ACM Trans. Manag. Inf. Syst. 2016, 6, 13. [Google Scholar] [CrossRef]
  17. Jannach, D.; Hegelich, K. A case study on the effectiveness of recommendations in the mobile internet. In Proceedings of the Third ACM Conference on Recommender Systems, New York, NY, USA, 23–25 October 2009; pp. 205–208. [Google Scholar]
  18. Kirshenbaum, E.; Forman, G.; Dugan, M. A live comparison of methods for personalized article recommendation at forbes.com. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Riva del Garda, Italy, 19–23 September 2012; Springer: Berlin, Germany, 2012; pp. 51–66. [Google Scholar]
  19. Lerche, L.; Jannach, D.; Ludewig, M. On the value of reminders within e-commerce recommendations. In Proceedings of the 2016 Conference on User Modeling Adaptation and Personalization, Halifax, NS, Canada, 13–17 July 2016; pp. 27–35. [Google Scholar]
  20. Herlocker, J.L.; Konstan, J.A.; Terveen, L.G.; Riedl, J.T. Evaluating collaborative filtering recommender systems. ACM Trans. Inf. Syst. 2004, 22, 5–53. [Google Scholar] [CrossRef] [Green Version]
  21. Tan, Y.K.; Xu, X.; Liu, Y. Improved recurrent neural networks for session-based recommendations. In Proceedings of the 1st Workshop on Deep Learning for Recommender Systems, Boston, MA, USA, 15–19 September 2016; pp. 17–22. [Google Scholar]
  22. Jannach, D.; Ludewig, M. When recurrent neural networks meet the neighborhood for session-based recommendation. In Proceedings of the 11th ACM Conference on Recommender Systems, Como, Italy, 27–31 August 2017; pp. 306–310. [Google Scholar]
  23. Dong, D.; Zheng, X.; Zhang, R.; Wang, Y. Recurrent Collaborative Filtering for Unifying General and Sequential Recommender. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 3350–3356. [Google Scholar]
  24. Liang, N.; Zheng, H.T.; Chen, J.Y.; Sangaiah, A.K.; Zhao, C.Z. TRSDL: Tag-Aware Recommender System Based on Deep Learning–Intelligent Computing Systems. Appl. Sci. 2018, 8, 799. [Google Scholar] [CrossRef]
  25. Zimdars, A.; Chickering, D.M.; Meek, C. Using temporal data for making recommendations. In Proceedings of the Seventeenth Conference on Uncertainty in Artificial Intelligence, Seattle, WA, USA, 2–5 August 2001; pp. 580–588. [Google Scholar]
  26. Mobasher, B.; Dai, H.; Luo, T.; Nakagawa, M. Using sequential and non-sequential patterns in predictive web usage mining tasks. In Proceedings of the IEEE International Conference on Data Mining, Maebashi City, Japan, 9–12 December 2002; pp. 669–672. [Google Scholar]
  27. Park, S.E.; Lee, S.; Lee, S.G. Session-based collaborative filtering for predicting the next song. In Proceedings of the IEEE 2011 First ACIS/JNU International Conference on Computers, Networks, Systems and Industrial Engineering (CNSI), Jeju Island, Korea, 23–25 May 2011; pp. 353–358. [Google Scholar]
  28. Moore, J.L.; Chen, S.; Turnbull, D.; Joachims, T. Taste Over Time: The Temporal Dynamics of User Preferences. In Proceedings of the 14th International Society for Music Information Retrieval Conference, Curitiba, Brazil, 4–8 November 2013; pp. 401–406. [Google Scholar]
  29. Hu, L.; Cao, L.; Wang, S.; Xu, G.; Cao, J.; Gu, Z. Diversifying personalized recommendation with user-session context. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 1858–1864. [Google Scholar]
  30. Jing, H.; Smola, A.J. Neural survival recommender. In Proceedings of the 10th ACM International Conference on Web Search and Data Mining, Cambridge, UK, 6–10 February 2017; pp. 515–524. [Google Scholar]
  31. Wu, C.Y.; Ahmed, A.; Beutel, A.; Smola, A.J.; Jing, H. Recurrent recommender networks. In Proceedings of the 10th ACM International Conference on Web Search and Data Mining, Cambridge, UK, 6–10 February 2017; pp. 495–503. [Google Scholar]
  32. Yu, F.; Liu, Q.; Wu, S.; Wang, L.; Tan, T. A Dynamic Recurrent Model for Next Basket Recommendation. In Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval, Pisa, Italy, 17–21 July 2016; pp. 729–732. [Google Scholar]
  33. Digital Search Advertising Revenue Growth in the United States from 2017 to 2022. Available online: https://www.statista.com/statistics/460955/search-advertising-revenue-growth-digital-market-outlook-usa/ (accessed on 26 November 2018).
  34. Cheng, Z.; Ding, Y.; He, X.; Zhu, L.; Song, X.; Kankanhalli, M.S. A^3NCF: An Adaptive Aspect Attention Model for Rating Prediction. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 3748–3754. [Google Scholar]
  35. Wang, S.; Hu, L.; Cao, L.; Huang, X.; Lian, D.; Liu, W. Attention-based transactional context embedding for next-item recommendation. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 2532–2539. [Google Scholar]
  36. Tay, Y.; Tuan, L.A.; Hui, S.C. Multi-Pointer Co-Attention Networks for Recommendation. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 1–10. [Google Scholar]
  37. Zhou, C.; Bai, J.; Song, J.; Liu, X.; Zhao, Z.; Chen, X.; Gao, J. ATRank: An Attention-Based User Behavior Modeling Framework for Recommendation. arXiv, 2017; arXiv:1711.06632. [Google Scholar]
  38. Xiong, C.; Callan, J.; Liu, T.Y. Word-Entity Duet Representations for Document Ranking. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, Tokyo, Japan, 7–11 August 2017; pp. 763–772. [Google Scholar]
  39. Chen, L.; Zhang, H.; Xiao, J.; Nie, L.; Shao, J.; Liu, W.; Chua, T.S. Sca-cnn: Spatial and channel-wise attention in convolutional networks for image captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6298–6306. [Google Scholar]
  40. Xiao, J.; Ye, H.; He, X.; Zhang, H.; Wu, F.; Chua, T.S. Attentional factorization machines: Learning the weight of feature interactions via attention networks. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 3119–3125. [Google Scholar]
  41. Adomavicius, G.; Tuzhilin, A. Toward the next generation of recommender systems: A survey of the state-of-the-art and possible extensions. IEEE Trans. Knowl. Data Eng. 2005, 17, 734–749. [Google Scholar] [CrossRef]
  42. Bobadilla, J.; Ortega, F.; Hernando, A.; Gutiérrez, A. Recommender systems survey. Knowl. Based Syst. 2013, 46, 109–132. [Google Scholar] [CrossRef]
  43. Bobadilla, J.; Hernando, A.; Ortega, F.; Bernal, J. A framework for collaborative filtering recommender systems. Expert Syst. Appl. 2011, 38, 14609–14623. [Google Scholar] [CrossRef] [Green Version]
  44. Schafer, J.B.; Frankowski, D.; Herlocker, J.; Sen, S. Collaborative filtering recommender systems. In The Adaptive Web; Springer: Berlin, Germany, 2007; pp. 291–324. [Google Scholar]
  45. Su, X.; Khoshgoftaar, T.M. A survey of collaborative filtering techniques. Adv. Artif. Intell. 2009, 2009. [Google Scholar] [CrossRef]
  46. Bobadilla, J.; Serradilla, F. The effect of sparsity on collaborative filtering metrics. In Proceedings of the Twentieth Australasian Conference on Australasian Database, Wellington, New Zealand, 1 January 2009; Volume 92, pp. 9–18. [Google Scholar]
  47. Luo, X.; Xia, Y.; Zhu, Q. Incremental collaborative filtering recommender based on regularized matrix factorization. Knowl. Based Syst. 2012, 27, 271–280. [Google Scholar] [CrossRef]
  48. Bobadilla, J.; Ortega, F.; Hernando, A.; Bernal, J. A collaborative filtering approach to mitigate the new user cold start problem. Knowl. Based Syst. 2012, 26, 225–238. [Google Scholar] [CrossRef] [Green Version]
  49. Kim, H.N.; El-Saddik, A.; Jo, G.S. Collaborative error-reflected models for cold-start recommender systems. Decis. Support Syst. 2011, 51, 519–531. [Google Scholar] [CrossRef]
  50. Schein, A.I.; Popescul, A.; Ungar, L.H.; Pennock, D.M. Methods and metrics for cold-start recommendations. In Proceedings of the 25th annual international ACM SIGIR conference on Research and development in information retrieval, Tampere, Finland, 11–15 August 2002; pp. 253–260. [Google Scholar]
  51. Devooght, R.; Bersini, H. Collaborative filtering with recurrent neural networks. arXiv, 2016; arXiv:1608.07400. [Google Scholar]
  52. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef] [PubMed]
  53. Koren, Y.; Bell, R.; Volinsky, C. Matrix factorization techniques for recommender systems. Computer 2009, 42, 30–37. [Google Scholar] [CrossRef]
  54. Rendle, S.; Freudenthaler, C.; Schmidt-Thieme, L. Factorizing personalized markov chains for next-basket recommendation. In Proceedings of the 19th International Conference on World Wide Web, Raleigh, NC, USA, 26–30 April 2010; pp. 811–820. [Google Scholar]
  55. What Is the Difference between Test and Validation Datasets? Available online: https://machinelearningmastery.com/difference-test-validation-datasets/ (accessed on 26 November 2018).
Figure 1. An example of Amazon recommender system.
Figure 1. An example of Amazon recommender system.
Applsci 08 02426 g001
Figure 2. ASLM architecture. Two mutually exclusive methods are shown: (a) ASLM-AP, and (b) ASLM-LHS. Only one method is used at a time.
Figure 2. ASLM architecture. Two mutually exclusive methods are shown: (a) ASLM-AP, and (b) ASLM-LHS. Only one method is used at a time.
Applsci 08 02426 g002
Figure 3. The main part of attention mechanism.
Figure 3. The main part of attention mechanism.
Applsci 08 02426 g003
Figure 4. Distribution histograms of the number of sessions by the frequency of users for Reddit and Last.fm datasets.
Figure 4. Distribution histograms of the number of sessions by the frequency of users for Reddit and Last.fm datasets.
Applsci 08 02426 g004
Figure 5. Distribution histogram of the number of sessions by the frequency of users for Tmall dataset.
Figure 5. Distribution histogram of the number of sessions by the frequency of users for Tmall dataset.
Applsci 08 02426 g005
Figure 6. Distribution histograms of the number of events by the frequency of sessions for Reddit and Last.fm datasets.
Figure 6. Distribution histograms of the number of events by the frequency of sessions for Reddit and Last.fm datasets.
Applsci 08 02426 g006
Figure 7. Distribution histogram of the number of events by the frequency of sessions for Tmall dataset.
Figure 7. Distribution histogram of the number of events by the frequency of sessions for Tmall dataset.
Applsci 08 02426 g007
Figure 8. Comparison of cold-start effects at the start of a new session for Reddit and Last.fm datasets between ASLM and II-RNN.
Figure 8. Comparison of cold-start effects at the start of a new session for Reddit and Last.fm datasets between ASLM and II-RNN.
Applsci 08 02426 g008
Figure 9. Cold-start effect at the start of a new session for the Tmall dataset. Please note that no data are provided for the cold-start problem by the SWIWO method.
Figure 9. Cold-start effect at the start of a new session for the Tmall dataset. Please note that no data are provided for the cold-start problem by the SWIWO method.
Applsci 08 02426 g009
Table 1. Descriptive statistics for the datasets.
Table 1. Descriptive statistics for the datasets.
RedditLast.fmTmall1/10 Reddit1/80 Last.fm
Number of users18,27197750,1977405612
Number of sessions1,135,488630,774200,28648,4729099
Number of items27,45294,28452,20771707137
Average number of events in a session3.008.102.992.242.09
Sessions per user62.15645.623.996.5514.87
Table 2. Standard deviation and skewness of the number of sessions by the frequency of users for the datasets.
Table 2. Standard deviation and skewness of the number of sessions by the frequency of users for the datasets.
RedditLast.fmTmall1/10 Reddit1/80 Last.fm
Standard deviation48.33678.591.153.7918.85
Skewness0.572.0612.161.544.49
Table 3. Standard deviation and skewness of the number of events by the frequency of sessions for the datasets.
Table 3. Standard deviation and skewness of the number of events by the frequency of sessions for the datasets.
RedditLast.fmTmall1/10 Reddit1/80 Last.fm
Standard deviation2.086.141.450.700.42
Skewness4.080.833.056.6113.63
Table 4. Configurations for ASLM.
Table 4. Configurations for ASLM.
RedditLast.fmTmall1/10 Reddit1/80 Last.fm
Embedding size5010010050100
Learning rate0.00010.00010.00010.00010.0001
Dropout rate00.2000.2
Max. recent session representations2020202020
Mini-batch size100100100100100
Table 5. Group A: Recall and MRR scores for the ASLM models and the baselines for the test sets in Reddit and Last.fm datasets.
Table 5. Group A: Recall and MRR scores for the ASLM models and the baselines for the test sets in Reddit and Last.fm datasets.
R@5R@10R@20MRR@5MRR@10MRR@20
Most Popular0.12410.17990.24990.08170.08900.0940
II-RNN-AP0.29430.37170.45270.19200.20240.2080
II-RNN-LHS0.31120.39040.47090.20250.21310.2186
ASLM-AP0.31620.38790.46430.22770.23730.2426
(+1.62%)(−0.63%)(−1.40%)(+12.44%)(+11.36%)(+10.95%)
ASLM-LHS0.33960.41000.48430.24730.25670.2619
(+9.12%)(+5.04%)(+2.85%)(+22.10%)(+20.48%)(+19.79%)
(a) Reddit Dataset (Test).
R@5R@10R@20MRR@5MRR@10MRR@20
Most Popular0.05690.06890.08660.04700.04870.0499
II-RNN-AP0.10090.13410.17840.06820.07250.0755
II-RNN-LHS0.08550.11420.15550.05880.06250.0654
ASLM-AP0.10530.13870.18370.07320.07760.0806
(+4.36%)(+3.46%)(+3.01%)(+7.33%)(+6.99%)(+6.75%)
ASLM-LHS0.10420.14150.19140.06930.07420.0776
(+3.34%)(+5.52%)(+7.31%)(+1.66%)(+2.34%)(+2.78%)
(b) Last.fm Dataset (Test).
Table 6. Group B: Recall and MRR scores for the ASLM models and the baselines for the test sets in 1/10 Reddit and 1/80 Last.fm datasets.
Table 6. Group B: Recall and MRR scores for the ASLM models and the baselines for the test sets in 1/10 Reddit and 1/80 Last.fm datasets.
R@5R@10R@20MRR@5MRR@10MRR@20
Most Popular0.14510.21550.28820.09660.10600.1111
II-RNN-AP0.14840.21610.29280.09690.10580.1112
II-RNN-LHS0.16140.22920.30620.10610.11510.1204
ASLM-AP0.19040.25070.32450.12560.13360.1387
(+17.96%)(+9.38%)(+5.98%)(+18.37%)(+16.07%)(+15.17%)
ASLM-LHS0.18860.24880.32510.12340.13140.1366
(+16.83%)(+8.55%)(+6.19%)(+16.27%)(+14.13%)(+13.48%)
(a) 1/10 Reddit Dataset (Test).
R@5R@10R@20MRR@5MRR@10MRR@20
Most Popular0.03710.04930.05960.02900.03080.0314
II-RNN-AP0.03490.04480.06040.02820.02940.0305
II-RNN-LHS0.03610.04590.06080.02850.02980.0308
ASLM-AP0.03880.04960.06560.03080.03230.0334
(+4.38%)(+0.57%)(+7.95%)(+6.18%)(+4.93%)(+6.13%)
ASLM-LHS0.03880.05030.06740.03090.03240.0336
(+4.38%)(+1.92%)(+10.97%)(+6.52%)(+5.37%)(+6.98%)
(b) 1/80 Last.fm Dataset (Test).
Table 7. Group B: Recall and MRR scores for the ASLM models and the baselines in the Tmall dataset.
Table 7. Group B: Recall and MRR scores for the ASLM models and the baselines in the Tmall dataset.
R@10R@20MRR@20
Most Popular0.02340.04200.0123
SWIWO-I0.31770.38100.1903
SWIWO0.30820.37030.1885
ASLM-AP0.54120.56610.4037
(+70.34%)(+48.57%)(+112.12%)
ASLM-LHS0.54280.56790.4082
(+70.86%)(+49.06%)(+114.50%)
Table 8. Group A: Recall and MRR scores for the ASLM models and the baselines for the validation sets in Reddit and Last.fm datasets.
Table 8. Group A: Recall and MRR scores for the ASLM models and the baselines for the validation sets in Reddit and Last.fm datasets.
R@5R@10R@20MRR@5MRR@10MRR@20
Most Popular0.12690.18300.25290.08350.09080.0957
II-RNN-AP0.30240.38270.46530.19710.20790.2136
II-RNN-LHS0.31910.40130.48370.20730.21820.2239
ASLM-AP0.32550.39910.47680.23360.24340.2488
(+2.02%)(−0.54%)(−1.43%)(+12.72%)(+11.55%)(+11.10%)
ASLM-LHS0.34890.42180.49820.25270.26240.2677
(+9.34%)(+5.12%)(+3.00%)(+21.90%)(+20.26%)(+19.56%)
(a) Reddit Dataset (Validation).
R@5R@10R@20MRR@5MRR@10MRR@20
Most Popular0.05630.06850.08620.04600.04770.0489
II-RNN-AP0.10820.14600.19620.07030.07530.0787
II-RNN-LHS0.09210.12490.17050.06140.06580.0689
ASLM-AP0.11470.15260.20270.07760.08260.0860
(+6.07%)(+4.50%)(+3.33%)(+10.44%)(+9.69%)(+9.36%)
ASLM-LHS0.11790.16130.21860.07600.08180.0857
(+9.00%)(+10.48%)(+11.42%)(+8.21%)(+8.63%)(+8.94%)
(b) Last.fm Dataset (Validation).
Table 9. Group B: Recall and MRR scores for the ASLM models and the baselines for the validation sets in 1/10 Reddit and 1/80 Last.fm datasets.
Table 9. Group B: Recall and MRR scores for the ASLM models and the baselines for the validation sets in 1/10 Reddit and 1/80 Last.fm datasets.
R@5R@10R@20MRR@5MRR@10MRR@20
Most Popular0.15160.22060.29130.10040.10950.1145
II-RNN-AP0.15510.22150.29660.10260.11130.1165
II-RNN-LHS0.17020.23660.31160.11080.11960.1248
ASLM-AP0.19820.25630.32820.13150.13920.1442
(+16.47%)(+8.33%)(+5.32%)(+18.68%)(+16.42%)(+15.58%)
ASLM-LHS0.19350.25490.33020.12750.13560.1408
(+13.71%)(+7.76%)(+5.96%)(+15.04%)(+13.38%)(+12.82%)
(a) 1/10 Reddit Dataset (Validation).
R@5R@10R@20MRR@5MRR@10MRR@20
Most Popular0.03600.04730.06260.02840.02990.0310
II-RNN-AP0.03430.04110.05630.02860.02950.0306
II-RNN-LHS0.03720.04540.06410.03170.03270.0340
ASLM-AP0.03800.04760.06630.03180.03300.0343
(+1.97%)(+0.55%)(+3.54%)(+0.21%)(+0.92%)(+0.88%)
ASLM-LHS0.03980.04820.06440.03210.03320.0343
(+6.89%)(+1.75%)(+0.57%)(+1.26%)(+1.43%)(+0.88%)
(b) 1/80 Last.fm Dataset (Validation).

Share and Cite

MDPI and ACS Style

Huang, R.; McIntyre, S.; Song, M.; E, H.; Ou, Z. An Attention-Based Recommender System to Predict Contextual Intent Based on Choice Histories across and within Sessions. Appl. Sci. 2018, 8, 2426. https://doi.org/10.3390/app8122426

AMA Style

Huang R, McIntyre S, Song M, E H, Ou Z. An Attention-Based Recommender System to Predict Contextual Intent Based on Choice Histories across and within Sessions. Applied Sciences. 2018; 8(12):2426. https://doi.org/10.3390/app8122426

Chicago/Turabian Style

Huang, Ruo, Shelby McIntyre, Meina Song, Haihong E, and Zhonghong Ou. 2018. "An Attention-Based Recommender System to Predict Contextual Intent Based on Choice Histories across and within Sessions" Applied Sciences 8, no. 12: 2426. https://doi.org/10.3390/app8122426

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop