Next Article in Journal
Multi-Mode Compound Grasping Robot Finger Driven by Linkage
Previous Article in Journal
A New Macroscopic Traffic Flow Characterization Incorporating Traffic Emissions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Personalized Hybrid Recommendation Algorithm for MOOCs Based on Learners’ Dynamic Preferences and Multidimensional Capabilities

School of Economics and Management, Tongji University, Shanghai 200092, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(9), 5548; https://doi.org/10.3390/app13095548
Submission received: 24 February 2023 / Revised: 24 April 2023 / Accepted: 26 April 2023 / Published: 29 April 2023

Abstract

:
In the MOOCs context, learners experience information overload. Thus, it is necessary to improve personalized recommendation algorithms for learners. The current recommendation algorithm focuses mainly on the learners’ course ratings. However, the choice of courses is not only based on the learners’ interests and preferences. It is also affected by learners’ knowledge domains and learning capabilities, all of which change dynamically over time. Therefore, this study proposes a personalized hybrid recommendation algorithm combining clustering with collaborative filtering. First, data on learners’ course rating preferences, course attribute preferences, and multidimensional capabilities that match course traits are used based on multidimensional item response theory. Second, considering that learners’ preferences and multidimensional capabilities change dynamically over time, the Ebbinghaus forgetting curve is introduced by integrating memory weights to improve the accuracy and interpretation of the proposed recommendation algorithm for MOOCs. Finally, the performance of the proposed recommendation algorithm is investigated using data from Coursera, an internationally renowned MOOCs platform. The experimental results show that the proposed recommendation algorithm is superior to the baseline algorithms. Accordingly, relevant suggestions are proposed for the development of MOOCs.

1. Introduction

Massive open online courses (MOOCs) have become popular recently. Learners can select MOOCs from different platforms based on their learning interests [1,2]. According to a report from Class Central, by December 2021, the number of MOOC learners reached 220 million, and the number of courses reached 19,400 [3].
Since 2011, Udacity, edX, and Coursera have become well-known international MOOC platforms. On the one hand, courses continue to increase, making it difficult for learners to find suitable courses [4,5]. Some high-quality course resources are also difficult for target learners to find. On the other hand, the massive learning data generated by MOOC platforms contain rich learning information. Therefore, it is necessary to dig deeply into the characteristics of learners and courses so that the learning characteristics and learning experiences of MOOC learners can be incorporated into improved and personalized course recommendations.
Currently, MOOCs’ recommendation algorithms usually focus on data mining learners’ course ratings. However, learners’ course choices are also based on their personal interests, preferences, knowledge domains, and learning capabilities. For example, knowledge graph enhanced recommendation algorithms were proposed to estimate learners’ potential interests automatically and iteratively [6] In fact, their interests, preferences, knowledge domains, and capabilities dynamically change with time [7,8]. Notably, owing to the large-scale data in the MOOC context, runtime efficiency is low when using a collaborative filtering algorithm to measure learners’ similarities. Therefore, based on multidimensional item response theory in psychometrics and the Ebbinghaus forgetting curve, this study proposed a hybrid recommendation algorithm for MOOCs by integrating clustering and collaborative filtering algorithms. These algorithms mine learners’ preferences based on course ratings, course attributes, and multidimensional capabilities by integrating memory weights.
Coursera, an internationally renowned MOOC platform, was selected to verify the effectiveness of the proposed recommendation algorithm. On the theoretical side, this study can enrich and develop personalized recommendations in the research field of MOOCs. In practice, it can promote the effective knowledge dissemination of high-quality course resources to improve learners’ learning experiences in the MOOC context.

2. Literature Review on the Personalized Recommendation in the MOOC Context

Owing to developments in information technology, MOOCs have flourished since 2008. Especially during the COVID-19 pandemic, MOOC platforms provided learners with stable learning resources and environments. Moreover, owing to the expanding scale of MOOC platforms and learners, personalized recommendations for MOOCs have become a research focus recently [9]. The current mainstream recommendation algorithms can be summarized as content-based, collaborative filtering-based, machine learning model-based, and hybrid model-based.

2.1. Content-Based Recommendation Algorithms

Content-based recommendation algorithms primarily center on information screening and filtering mechanisms. A direct correlation between items or content was found, and based on user preferences and historical behavior records, relevant items or contents were recommended to users.
Content-based personalized recommendations on MOOC platforms are based on learners’ course preferences embedded in both the relevant content searched and course learning records. Courses with high relevance are then recommended to matched learners. An advantage of this recommendation algorithm is its easy implementation. A content-based personalized recommendation only requires a learner’s course information, learning records, and search content. The proposed content-based MOOC recommendation algorithm has a higher accuracy rate than random recommendations. Notably, the richer the course information, the more accurate the recommendations [10,11].
However, content-based recommendation algorithms have some obvious disadvantages. This is because the recommendations based on learners’ historical behaviors are likely to be too consistent, resulting in a high degree of similarity between the recommended courses and the courses that learners have already studied. Consequently, exploring new interests using this content-based recommendation algorithm is difficult.

2.2. Recommendation Algorithm Based on Collaborative Filtering

Collaborative filtering is another widely used method. It is mainly based on the assumption that similar things or people fit well together. Correspondingly, learners who like the same items are more likely to have the same interests and preferences in general [12,13]. At present, collaborative filtering algorithms are mainly of two types, namely, user- and item-based.
User-based collaborative filtering has been applied to MOOCs’ personalized recommendations to determine target learners’ preferences based on their course ratings to group together learners that are similar to some extent (hereafter referred to as similar learner groups). On this basis, personalized recommendations to target learners are provided according to the courses preferred by similar learner groups.
Item-based collaborative filtering (Item-CF) has been applied to personalized recommendations to find similar courses by modeling courses and calculating course similarities to recommend similar courses for learners. Thus, the recommendation algorithm based on collaborative filtering is easy to understand and implement. Personalized recommendation research in the MOOC context has shown that an adaptive recommendation scheme based on collaborative filtering and time series can improve the accuracy of personalized recommendations [14].
However, the essence of the collaborative filtering algorithm for MOOCs is to make recommendations based on learners’ historical data, leading to certain unavoidable disadvantages. For newly registered MOOC learners and newly started courses, there may be a cold-start problem, resulting in sparse data, thus affecting recommendation outcomes.

2.3. Recommendation Algorithms Based on Machine Learning Models

With the development of big data technology, recommendation algorithms based on machine learning models have been created to predict and rank learners’ interests and preferences. Common machine learning models include matrix decomposition, topic, clustering, and neural network models. Although the training time of recommendation algorithms based on machine learning models is longer, they have obvious advantages in feature extraction. The recommendation results are more accurate, especially when based on a deep learning model. Data can be accurately represented in a high-level form.
Recommendation algorithms based on machine learning models have been applied to personalized recommendation for MOOCs. For example, a deep learning personalized recommendation algorithm for MOOCs based on multiple attention mechanisms was proposed by analyzing learning behaviors [15]. The personalized recommendation algorithm for MOOCs based on deep belief networks was proposed in function approximation, feature extraction, and prediction classification [16]. A personalized recommendation algorithm for MOOCs based on a priori clustering was proposed by revealing the course association rules [17].
Although recommendation algorithms based on machine learning models have gained increasing attention owing to their high accuracy rate, they also have some disadvantages. First, because the recommendation process of this algorithm is similar to that of the black box, the interpretability of the recommendation results is weak. Second, to ensure the model has better recommendation results, manual intervention is required to screen the data attributes and adjust and correct the parameters, requiring more time and resources.

2.4. Recommendation Algorithms Based on Mixed Models

In practical applications, it is usually rare to directly use one algorithm alone for a recommendation because each of the aforementioned recommendation algorithms has certain shortcomings. Therefore, many personalized recommendation systems use a combination of different algorithms to obtain a more accurate recommendation performance.
Given the characteristics of MOOC platforms, the resource recommendation model is constructed by integrating a deep belief network model. As a result, the contrast divergence algorithm has higher accuracy [18]. The matrix decomposition model of a multi-social network framework based on a joint probability distribution helps improve the runtime efficiency [19]. Furthermore, a collaborative filtering algorithm based on a dichotomous graph context can effectively solve the cold-start problem in collaborative filtering [20]. In addition, both the hybrid recommendation model integrating the collaborative filtering algorithm and deep-learning algorithm [21], and the personalized recommendation method based on deep-learning and multi-view fusion [22] have better recommendation results.
Consequently, recommendation algorithms based on hybrid models, which combine the advantages of multiple algorithms, have gradually become popular in the MOOC context.

2.5. Comments on Current Research

Collaborative filtering algorithms are still widely popular owing to their simplicity and efficiency. However, the problems of data sparsity and cold-start must be solved. Currently, hybrid recommendations are the mainstream recommendation algorithms. For example, for the movie dataset scenario, clustering users based on their personal attributes or their rating matrix [23,24], followed by applying collaborative filtering recommendations, help improve the accuracy of collaborative filtering algorithms.
In the MOOC context, learners often choose courses according to the information provided by the platform. Accordingly, course information reflects learners’ interests in and preferences for a course to a certain extent. Although current recommendation algorithms for MOOCs focus on course ratings, they do not sufficiently consider learners’ knowledge domains and learning levels. Thus, this study expanded course rating data to learners’ course rating preferences, course attribute preferences, and multidimensional capabilities matched with course traits.
To reiterate, learners’ preferences and capabilities change dynamically over time in the MOOC context. Although some algorithms recommend based on learners’ learning sequences [14], current recommendation algorithms seldom consider dynamic changes in these attributes. Moreover, the runtime efficiency of the collaborative filtering algorithm was not high when directly measuring the learners’ similarities in the MOOC context.
Therefore, this study proposes a hybrid recommendation algorithm for MOOCs by integrating clustering and collaborative filtering based on multidimensional item response theory (MIRT) and the Ebbinghaus forgetting curve. MIRT can be viewed as an extension of IRT for better interpreting the interaction between item characteristics and individual behaviors due to the focus on multidimensional traits [25]. By integrating memory weight into learners’ course rating preferences, course attribute preferences, and multidimensional capabilities that match course traits, the proposed recommendation algorithm aims to solve the data sparsity and singularity problems in the collaborative filtering algorithm.

3. Methods

Based on the general architecture of a personalized recommendation system, the hybrid recommendation model framework proposed in this study is presented in Figure 1.

3.1. Input Module

Data on the learners’ course ratings, course attribute preferences, and multidimensional capabilities were obtained from expanding course rating data. Then, based on multidimensional item response theory in psychometrics, the learners’ multidimensional capabilities that match course traits were acquired. Subsequently, the Ebbinghaus forgetting curve that incorporates memory weights was introduced in Step 1 to capture the dynamic changes in learners’ course ratings, course attribute preferences, and multidimensional capabilities. This summarizes the data preparation.

3.1.1. Learners’ Course Ratings with Memory Weight

The Ebbinghaus forgetting curve represents the capacity to forget and remember as a time-dependent function [26]. The rate at which people forget descends until it becomes stable, as shown in Equation (1). In Equation (1), f t denotes the memory retention ratio, f 0 denotes the initial value of memory, t denotes the interval between the initial memory time and the current time, and c and k are control constants. According to repeated experiments, the amount of memory left is a logarithmic function curve, and f 0 = 100 ,   c = 1.25 ,   k = 1.84 [23].
f t = f 0 k l g t c + k = 100 0 1.84 l g t 1.25 + 1.84
Therefore, the memory-related characteristics of MOOC learners should be incorporated into the model [27]. According to the timeframes of learners’ course ratings, the memory retention ratio of a course should be incorporated as a factor in a learner’s course ratings.
First, a two-dimensional matrix of course ratings is constructed, in which the number of matrix rows denotes the number of MOOC learners, the number of matrix columns denotes the number of MOOC courses, and the element of the matrix rate i j represents the value of learner u i ’s rating of course j . Then, the memory retention ratio in Equation (1) is introduced to construct course rating dynamic preferences r a t e i j , as in Equation (2).
r a t e i j = r a t e i j f t

3.1.2. Learners’ Course Attribute Preferences with Memory Weight

First, the learners’ course ratings can be used to construct a two-dimensional matrix of course attributes. The number of rows in the matrix denotes the number of learners, the number of columns in the matrix denotes the number of course attributes, and the element in the matrix attr i j = c = 1 n n c j n denotes the average value of learner u i ’s course attribute j in n rating courses by learner u i ; n c j represents the value of attribute j of course c . The course ratings with memory weights in Equation (2) is introduced to construct course attribute dynamic preferences a t t r i j , as in Equation (3).
a t t r i j = attr i j r a t e i j

3.1.3. Learners’ Multidimensional Capabilities with Memory Weight

The knowledge domain, teaching language (the necessary language requirement), and course level were treated as course traits in this study. Thus, multidimensional capabilities that match course traits were explored based on multidimensional item response theory (MIRT). Although MIRT is widely used in assessment and evaluation of educational and psychological tests, it is seldom applied in the MOOC context [28]. In this study, it modeled learners’ multidimensional capabilities that match course traits.
First, the course traits ratings were obtained from the learners’ course rating data. The data were converted into a numerical matrix using one-hot coding to construct a learner-multidimensional capability matrix. The number of rows in the matrix denotes the number of learners, the number of columns in the matrix denotes the number of course traits, and the element of the matrix c a p a b i l i t y i j = c = 1 n n c j n represents the average value of learner u i ’s course traits j in n rating courses by learner u i ; n c j represents the value of trait j of course c . Then, the course rating with memory weight in Equation (2) is introduced to model multidimensional dynamic capabilities that match course traits c a p a b i l i t y i j , as in Equation (4).
cap a b i l i t y i j = c a p ability i j r a t e i j

3.2. Recommending Module

In the MOOC context, course rating data are usually sparse. Therefore, in this study, learners were clustered by combining course attribute preferences and multidimensional capabilities. Then, for each learner group, a collaborative filtering algorithm was used. This algorithm comprehensively considers the similarity of course rating preferences, course attribute preferences, multidimensional capability matching course traits, and similarities among learners (thus creating similar learning groups). Consequently, the efficiency and accuracy of the recommendation can be improved.

3.2.1. Clustering Algorithm

As MOOCs have many course attributes, feature extraction and feature selection are required to reduce attribute redundancy so that course attributes can be better applied to course attribute preferences to improve the clustering effect. Therefore, principal component analysis (PCA) was employed to reduce the dimension of course attributes before clustering learners based on course attributes and multidimensional capabilities in Step 2.
As an unsupervised machine learning algorithm, the clustering algorithm divides the overall sample into several non-overlapping groups. It is based on a distance measurement method so that the individuals within each group are relatively more similar, and the individuals from different groups differ. Consequently, each group is clearly differentiated. Although there are many clustering algorithms, K-means has the advantage of being simple and efficient because it divides different individuals into different groups through iterative calculations, which is suitable in big data analysis. Therefore, by combining course attribute preferences and multidimensional capabilities that match course traits, this study selected K-means clustering to divide learners into K clusters (similar learning groups).

3.2.2. Collaborative Filtering Algorithm

As the learning choice of courses by MOOC learners is largely based on personal interests and capabilities, diversity is an important evaluation indicator in personalized recommendations. Therefore, in Step 3, for each cluster, a learner-based collaborative filtering algorithm (Learner-CF) is used to calculate the comprehensive similarity among learners using the k nearest neighbor algorithm.
(1)
Learner similarity in course rating preferences
The improved cosine similarity calculation of course ratings is shown in Equation (5). The larger the absolute value of s i m r a t e u i , u j , the higher the similarity of the course ratings between learner u i and learner u j . In Equation (5), I u i , u j denotes the intersection of courses rated by learner u i and learner u j ; x i k , x j k denote the rating of course k by learner u i and learner u j , respectively; x ¯ k denotes the average rating of course k .
s i m r a t e u i , u j = k I u i , u j x i k x ¯ k x j k x ¯ k k I u i , u j x i k x ¯ k 2 k I u i , u j x j k x ¯ k 2
(2)
Learner similarity in course attribute preference
The cosine similarity calculation of the course attribute preferences is shown in Equation (6). The larger the value of sim a t t r u i , u j , the higher the similarity of course attribute preferences between learner u i and learner u j . In Equation (6), A u i , u j denotes the attribute intersection of courses rated by learner u i and learner u j ; x i k , x j k denote the number of rating times for courses with attribute value k by learner u i and learner u j , respectively.
s i m a t t r u i , u j = k A u i , u j x i k x j k k A u i , u j x i k 2 k A u i , u j x j k 2
(3)
Learner similarity in multidimensional capabilities
The matching course traits of the cosine similarity calculation of multidimensional capabilities are shown in Equation (7). The larger the value of s i m a b i l i t y u i , u j , the higher the similarity of the multidimensional capabilities between learner u i and u j . In Equation (7), H u i , u j denotes the intersection of multidimensional capabilities that match the course traits of learner u i and learner u j ;   x i k , x j k denote the value of learner u i ’s capability k and learner u j ’s capability k , respectively.
s i m a b i l i t y u i , u j = k H u i , u j x i k x j k k H u i , u j x i k 2 k H u i , u j x j k 2
(4)
Comprehensive similarity among users
According to Equations (5)–(7), s i m r a t e u i , u j denotes learner similarity of course rating preferences, s i m a t t r u i , u j denotes learner similarity of course attribute preferences, and s i m a b i l i t y u i , u j denotes learner similarity of multidimensional capabilities that match course traits. Therefore, the comprehensive similarity between learner u i and learner u j is shown in Equation (8); α ,   β ,   λ are in the range of [0, 1] and satisfy α + β + λ = 1 to balance the influence of similarity in learners’ course rating preferences, course attribute preferences, and multidimensional capabilities that match course traits. It is usually necessary to determine the optimal value of the balance factors α ,   β ,   λ by testing.
sim u i , u j = α sim r a t e u i , u j + β sim a t t r u i , u j + λ sim a b i l i t y u i , u j
According to Equation (8), a symmetric similarity matrix with an m × m dimension is formed by calculating the similarities among learners, where m is the number of learners, and the elements of matrix s i j denote the similarity between learner u i and learner u j . The learners are ranked according to their similarities to other learners. The top k most similar learners are selected as the nearest neighbors; their similarity exceeds a pre-specified criterion.

3.3. Output Module

3.3.1. Prediction and Recommendation

First, we take the courses that the top k nearest neighbors have learned but that the target learner has not learned as candidate courses.
Moreover, some learners rate courses more conservatively, while others rate courses in an opposite way. As a second step, to ensure the consistency of learner course ratings, the predicted rating of unlearned course c by the target learner u i can be predicted, as shown in Equation (9). Here, r i ¯ denotes the average rating for learned courses by learner u i ; k i denotes the set of target learner’s top k nearest neighbor learners who rated course c in candidate courses; u j denotes learner u j in k i ; r j c denotes the rating of course c by learner u j ; r j ¯ denotes the average rating of learned courses by learner u j ; s i m u i , u j shown in Equation (8) denotes the comprehensive similarity between learner u i and learner u j . Considering the influence of the similarity between learner u i and learner u j , the predicted value of learner u i ’s course c rating r i c ^ can be calculated based on r i ¯ and j k i s i m u i , u j ( r j c r j ¯ ) j k i s i m u i , u j .
r i c ^ = r i ¯ + j k i s i m u i , u j ( r j c r j ¯ ) j k i s i m u i , u j
Finally, the top n courses with the highest predicted ratings were recommended for the target learners in Step 4.

3.3.2. Analysis of Performance Indicators

Offline evaluation is one of the most commonly adopted evaluation methods for recommendation systems. The holdout test is a basic offline evaluation method, which uses indicators of the precision rate, recall rate, F1 score, and mean average error (MAE) [14,18,19] to analyze and compare the performance of the recommendation results.
The precision or accuracy rate indicates the proportion of correctly predicted positive samples to the total number of samples predicted to be positive. In this study, this rate is shown in Equation (10). It describes the proportion of learners’ rated courses in the list of recommended courses. The recall rate is the proportion of correctly predicted positive samples to the actual positive samples, as shown in Equation (11); it describes the proportion of learners’ rated courses in the list of recommended courses to the overall learner-rated courses in this study. L u denotes the list of recommended courses for learner u; B u denotes the list of courses rated by learner u ; n denotes the total number of learners; U denotes the set of learners; P L denotes the overall recommendation precision rate; and R L denotes the overall recommendation recall rate.
P L = 1 n u U L u B u L u
P L = 1 n u U L u B u B u
To comprehensively reflect the recommendation performance, the F1 score reflects the precision rate and recall rate simultaneously, as shown in Equation (12).
F 1 = 2 · P L · R L P L + R L
Moreover, the mean absolute value error (MAE) can describe the error between the learners’ actual and predicted course ratings. As shown in Equation (13), y is the learner’s actual course rating, y ^ is the predicted one, and N t e s t is the number of the actual rating courses.
MAE = y y ^ N t e s t

4. Results

4.1. Data Collection

Coursera is a world-renowned MOOCs platform. As of December 2021, Coursera had over 92 million registered learners. It has over 250 university and industry partners. It offers more than 2000 instructional programs that provide hands-on learning. It has more than 6000 course resources covering the following 11 subject categories: arts and humanities, business, computer science, data science, information technology, health, math and logic, personal and development, physical science and engineering, social science, and language learning. Each category of discipline is subdivided into multiple second-level categories. Moreover, course discussion and course review forums are set up for teachers and students to discuss and share information. Therefore, Coursera was selected as the research object to collect course ratings, course attribute, and course trait data from January 2016 to December 2021.

4.2. Data Preprocessing

The raw data were preprocessed in the following manner. First, fields and records with null values were eliminated. Second, the data items were normalized and unified. Third, to improve the efficiency of the proposed recommendation algorithm, the preprocessed dataset was screened for courses rated by at least ten learners [29].
The data volumes of the original dataset, the preprocessed dataset, and the experiment dataset are listed in Table 1. There were 1849 learners who rated at least ten courses. They learned 2951 courses, accounting for 77% of the total number of courses following data preprocessing. The number of ratings and reviews of the 2951 experiment courses was 42,485 and 41,741, respectively.
Based on the experimental dataset, the statistics of 11 course attributes are shown in Table 2. The number of clicks and the number of views had the largest variances. These were followed by the number of registrations and the number of ratings. These indicated that courses on Coursera vary in popularity.

4.3. Recommendation Process on Coursera

4.3.1. Learners’ Dynamic Preferences and Multidimensional Capabilities

According to Step 1, as shown in Figure 1, Equations (1)–(4) with memory weight were applied to calculate learners’ course rating preferences, course attribute preferences, and multidimensional capabilities.
Table 3 shows a subset of data on learners’ course rating preferences with memory weight calculated by Equation (2). In this table, 0 means a learner did not rate any of the ten courses in the first row.
Table 4 shows a subset of data on learners’ course attribute preferences with memory weight calculated by Equation (3). It displays the average values of learners’ course attribute preferences. For example, courses rated by Learner 8164 had 54 materials, one teaching assistant, and 2554 reviews, on average.
Table 5 shows a subset of data on learners’ multidimensional capabilities (that match course traits) with memory weight calculated using Equation (4). Course traits on Coursera consist of teaching language (i.e., English, French, Russian, Spanish, and Portuguese), subject categories (i.e., computer science, data science, information technology, math and logic, business, personal development, life science, physical science and engineering, arts and humanities, social science, and language learning), and course-level (i.e., beginner, intermediate, and advanced). For example, Learner 8164 learned courses in English, French, Russian, and Spanish in the knowledge domain of computer science and data science at beginner and intermediate levels.

4.3.2. Feature Extraction of Course Attributes and Learner Clustering

Although course attribute preferences can be obtained, there may be a certain correlation between the 11 course attributes, as shown in Table 2. It is necessary to reduce the dimensions of the course attributes before clustering the learners. According to Step 2 shown in Figure 1, principal component analysis was applied.
The results of the KMO and Bartlett tests for course attribute preferences are presented in Table 6. The KMO measure was 0.877, which is greater than 0.8. The Bartlett spherical test was significant, at p < 0.05, indicating that the data were suitable for principal component analysis [30].
Based on the principal component analysis, the factors retained 86% of the component information. The principal components and the course attributes are listed in Table 7.
After dimension reduction and combing course attribute preferences with multidimensional capabilities that matched course traits, learners were clustered using the K-means clustering algorithm. This involved evaluating the clustering results by minimizing the distance between samples within each cluster. The smaller the sum of the squared distances between the class centers of each class and the sample points within the class, the more similar the class members. As the number of clusters increased, the error sum decreased until it reached a critical value and remained relatively stable; therefore, this point could be considered the optimal number of clusters.
In this study, the elbow method, a heuristic used in determining the number of clusters in a dataset, was applied to find the appropriate inflection point quickly and effectively [31]. Accordingly, K = 5 was determined as the optimal number of learner clusters. The characteristics of the clustering sample are shown in Table 8. Apparently, the number of samples in each cluster was balanced. The characteristics included course attributes and course traits based on course attribute preferences and multidimensional capabilities matching the course traits.

4.3.3. Calculating Learner Similarity

Using both the course attribute preferences and multidimensional capability to cluster learners (into similar learning groups), better clustering performance was shown. The same weight was given to the course attribute preferences and multidimensional capabilities in Equation (8), that is, β = γ . Consequently, according to Step 3, shown in Figure 1, an example of the learner similarity matrix is illustrated in Table 9 below. The diagonal elements represent the similarity of each learner to himself/herself, thus taking the value of 1. For example, the similarity between Learner 8064 and Learner 10,269 was 0.21.
Moreover, to calculate the comprehensive similarity of learners so that the top k similar learners in clusters could be found, the optimal values of the course rating similarity weight coefficient α in Equation (8) and the top k nearest neighbors were determined through sensitivity analysis.
In this study, the mean absolute error (MAE) between predicted and actual course ratings was obtained when the value of the weight coefficient α changed by iterating the nearest top k neighbors. Usually, MAE changes smoothly with the nearest top k neighbors. The step size k is generally set to 5 or 10 [26,32]. This study started with k = 10 , and a change in MAE was observed, as shown in Figure 2.
As indicated in Figure 2, when the number of the nearest neighbors k is fixed and the course rating similarity weight coefficient α increases from 0 to 1, MAE fluctuates; it decreases at first and then it increases after α = 0.8 . Finally, when the course rating similarity weight coefficient α is 0.8 and the number of the nearest neighbor k is 20, the model obtains the smallest MAE. Therefore, the course rating similarity weight coefficient α and the top k nearest neighbors were determined to be 0.8 and 20, respectively, through sensitivity analysis. Accordingly, based on Equation (8), both the course attribute similarity weight coefficient β and multidimensional capability similarity weight coefficient γ took the value of 0.1. Consequently, sorted by comprehensive similarity, the top 20 nearest neighbors were selected for target learners.

4.3.4. Prediction and Recommendation

According to Step 4 in Figure 1, a target learner’s predicted course rating, as shown in Equation (9), is obtained based on the top k nearest neighbors. Then the top n courses are recommended to the target learner in the ascending order of rating.
To obtain the optimal top n , a sensitivity analysis was performed by changing the number of recommended courses, as shown in Figure 3. As the number of recommended courses increased from 5 to 30, the precision rate of the algorithm decreased from 66% to 22%, whereas the recall rate increased from 24% to 45%. Therefore, the F1 score (Equation (12)) was introduced as the comprehensive evaluation criterion to determine the optimal number of recommend courses. Finally, when the number of recommended courses was 10, the F1 score reached a maximum of 39%. Therefore, according to the course rating by the top 2 0 nearest neighbors, the value of the target learner’s course rating was predicted so that the top 10 courses with the highest predicted rating were selected (recommended).

4.4. Model Performance Evaluation

First, when using the Holdout test method, 80% of the course rating dataset was randomly used as the training dataset. The remaining 20% was used as the testing dataset [33].
Second, the learner-based collaborative filtering algorithm (Learner-CF), the item-based collaborative filtering algorithm (Item-CF), and the proposed hybrid recommendation algorithm removing clustering (NoCluster-CF) were set as baseline algorithms. Then, the performance of the baseline algorithms was compared with the proposed hybrid recommendation algorithm (Cluster-CF).
Third, according to the top 20 nearest neighbors and the top 10 recommended courses from the previous sensitivity analysis, the precision rate, recall rate, and F1 score of the four algorithms were compared and analyzed, as shown in Figure 4. Compared with the Learn-CF, Item-CF, and NoCluster-CF, the precision and recall rates of Cluster-CF proposed in this study improved. Specifically, the F1 score (comprehensive evaluation index) was increased by 12%, 17%, and 7%, respectively.
Although the F1 score in the Cluster-CF improved by only less than 10%, its runtime efficiency improved by nearly 50% compared to NoCluster-CF. This shows that Cluster-CF is superior to the baseline recommendation algorithms with better runtime efficiency. This is because relatively homogenous learners belonging to the similar learning groups can be summarized by a single cluster representative, which enables data reduction.

5. Discussion

The development of MOOCs is important for promoting equity and education. However, with the massive growth of MOOC platforms, learners are confronted with information overload when searching for learning resources or choosing courses. This hampers their learning efficiency and learning experiences. It also hampers the platforms’ knowledge sharing. Consequently, course recommendations are essential in the MOOC context.
At present, many recommendation algorithms mainly focus on learners’ course ratings (ratings of previous courses). However, learners’ choices of courses depend not only on their previous course preferences. They also depends on their course attribute preferences and multidimensional capabilities that match course traits. Moreover, their course attribute preferences and multidimensional capabilities change dynamically over time. Therefore, based on multidimensional item response theory in psychometrics and the Ebbinghaus forgetting curve, this study addresses the need for better recommendation algorithms in the MOOC context. To this end, this study mines course ratings, course attribute preferences, and multidimensional capabilities, which integrate memory weights. Moreover, in order to improve the recommendation algorithm’s runtime efficiency, a clustering algorithm must be used (clustering learners before the collaborative filtering algorithm is applied) to measure learners’ similarities. Therefore, a hybrid recommendation algorithm is proposed in this paper to improve the runtime efficiency by combining a clustering algorithm based on collaborative filtering. According to this algorithm, learners are divided into clusters. A similarity calculation is done within each cluster to solve the problem of dealing with high-dimensional sparse data and to further improve the performance of the recommendation algorithm.
Coursera, the world-renowned MOOCs platform, was selected for this experiment. The learner-based collaborative filtering algorithm, the item-based collaborative filtering algorithm, and the proposed hybrid recommendation algorithm removing clustering were set as baseline algorithms. According to the experimental results, the proposed algorithm is superior to the baseline recommendation algorithms, with better runtime efficiency.

6. Conclusions

Theoretically, this study contributes the following.
First, we used multidimensional item response theory in psychometrics to expand course rating data to learners’ course rating preferences, course attribute preferences, and multidimensional capabilities that match the course traits. This solves the problem of singularity and sparse data sources arising from only using course rating data in the MOOC context.
Second, integrating memory weights from the Ebbinghaus forgetting curve is useful for depicting the dynamic characteristics of learners’ course rating preferences, course attribute preferences, and multidimensional capabilities matching the course traits. Consequently, the performance and interpretation of the recommendation algorithm proposed in this study can be improved.
Third, integrating Cluster-CF (a clustering algorithm with a collaborative filtering recommendation algorithm) improves the runtime efficiency of calculating learners’ similarities in the MOOC context that has a large amount of data.
This study used a dataset from Coursera. Cluster-CF (the proposed hybrid recommendation algorithm) was compared to three baseline algorithms. Cluster-CF has improved accuracy and runtime efficiency. In terms of practice, this study makes the following contributions.
First, since learners’ course rating preferences, course attribute preferences, and multidimensional capabilities that match course traits are the basis for course recommendations, data on the basic course information and comments provided on MOOCs platform should be exhaustive and elaborate.
Second, since learners’ course rating preferences, course attribute preferences, and multidimensional capabilities that match course traits have dynamic characteristics, courses provided on MOOC platforms should be diversified in language, knowledge domain, and knowledge levels. Consequently, diversified courses can better meet learners’ learning needs and competency development.
Third, Cluster-CF is recommended for MOOC platforms. This will help improve the dissemination of high-quality MOOC resources, knowledge sharing, knowledge exploration, and knowledge learning in the MOOC context.
However, for future research, other MOOC platforms can be selected, especially those capable of providing learners’ personal attribute information. This can make the data more exhaustive and diverse to improve the recommendation algorithm. Moreover, in the process of model experiments, the selection of the number of clusters, similarity weight coefficients, and the number of recommended courses can be combined with other methods in future research, such as using the parameter optimization function in machine learning, to make a more refined comparison for obtaining the best combination of parameters.

Author Contributions

Conceptualization, B.W.; methodology, B.W.; software, B.W. and L.L.; validation, B.W. and L.L.; formal analysis, B.W.; investigation, B.W. and L.L.; resources, B.W.; data curation, B.W. and L.L.; writing—original draft preparation, B.W. and L.L.; writing—review and editing, B.W.; visualization, B.W.; supervision, B.W.; project administration, B.W.; funding acquisition, B.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Chinese National Social Science Fund “Thirteenth Five-Year Plan” education topic (BFA180064).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chi, Z.X.; Zhang, S.; Shi, L. Analysis and Prediction of MOOC Learners’ Dropout Behavior. Appl. Sci. 2023, 13, 1068. [Google Scholar] [CrossRef]
  2. Iniesto, F.; Rodrigo, C.; Hillaire, G. A Case Study to Explore a UDL Evaluation Framework Based on MOOCs. Appl. Sci. 2023, 13, 476. [Google Scholar] [CrossRef]
  3. By The Numbers: MOOCs in 2021. Available online: https://www.classcentral.com/report/mooc-stats-2021/ (accessed on 1 December 2021).
  4. Luik, P.; Suviste, R.; Lepp, M.; Palts, T.; Tonisson, E.; Sade, M.; Papli, K. What motivates enrolment in programming MOOCs. Br. J. Educ. Technol. 2019, 50, 153–165. [Google Scholar] [CrossRef]
  5. Basnet, R.B.; Johnson, C.; Doleck, T. Dropout prediction in MOOCs using deep learning and machine learning. Educ. Inf. Technol. 2022, 27, 11499–11513. [Google Scholar] [CrossRef]
  6. Zhang, H.; Shen, X.; Yi, B.; Wang, W.; Feng, Y. KGAN: Knowledge Grouping Aggregation Network for course recommendation in MOOCs. Expert Syst. Appl. 2023, 211, 118344. [Google Scholar] [CrossRef]
  7. Deng, Y.Z. Construction of higher education knowledge map in university libraries based on MOOC. Electron. Libr. 2019, 37, 811–829. [Google Scholar] [CrossRef]
  8. Baron, C.; Daniel-Allegro, B. About adopting a systemic approach to design connected embedded systems: A MOOC promoting systems thinking and systems engineering. Syst. Eng. 2020, 23, 261–280. [Google Scholar] [CrossRef]
  9. Estrada-Molina, O.; Fuentes-Cancell, D.R. Engagement and desertion in MOOCs: Systematic review. Comunicar 2022, 30, 111–124. [Google Scholar] [CrossRef]
  10. Zhang, H.; Huang, T.; Lv, Z.H.; Liu, S.Y.; Yang, H. A Highly Accurate Resource Recommendation Model for Use in MOOC Environments. Mob. Netw. Appl. 2019, 24, 34–46. [Google Scholar] [CrossRef]
  11. Wang, R.R.; Shi, Z.J. Personalized Online Education Learning Strategies Based on Transfer Learning Emotion Classification Model. Secur. Commun. Netw. 2021, 2021, 5441631. [Google Scholar] [CrossRef]
  12. Elahi, E.; Halim, Z. Graph attention-based collaborative filtering for user-specific recommender system using knowledge graph and deep neural networks. Knowl. Inf. Syst. 2022, 64, 2457–2480. [Google Scholar] [CrossRef]
  13. Kaya, T.; Kaleli, C. A novel top-n recommendation method for multi-criteria collaborative filtering. Expert Syst. Appl. 2022, 198, 16695. [Google Scholar] [CrossRef]
  14. Sanchez-Moreno, D.; Zheng, Y.; Moreno-Garcia, M.N. Time-Aware Music Recommender Systems: Modeling the Evolution of Implicit User Preferences and User Listening Habits in A Collaborative Filtering Approach. Appl. Sci. 2020, 10, 5324. [Google Scholar] [CrossRef]
  15. Fan, J.; Jiang, Y.; Liu, Y.; Zhou, Y. Interpretable MOOC recommendation: A multi-attention network for personalized learning behavior analysis. Internet Res. 2021, 32, 588–605. [Google Scholar] [CrossRef]
  16. Zhang, H.; Yang, H.; Huang, T.; Zhan, G.Q. DBNCF: Personalized Courses Recommendation System Based on DBN in MOOC Environment. In Proceedings of the 2017 International Symposium on Educational Technology (ISET), Hong Kong, China, 27–29 June 2017; pp. 106–108. [Google Scholar] [CrossRef]
  17. Zhang, H.; Huang, T.; Lv, Z.H.; Liu, S.Y.; Zhou, Z.L. MCRS: A course recommendation system for MOOCs. Multimed. Tools Appl. 2018, 77, 7051–7069. [Google Scholar] [CrossRef]
  18. Elghomary, K.; Bouzidi, D.; Daoudi, N. Design of a Smart MOOC Trust Model: Towards a Dynamic Peer Recommendation to Foster Collaboration and Learner’s Engagement. Int. J. Emerg. Technol. Learn. 2022, 17, 36–56. [Google Scholar] [CrossRef]
  19. Bok, K.; Lee, S.; Choi, D.; Lee, D.; Yoo, J. Recommending personalized events based on user preference analysis in event based social networks. Electron. Commer. Res. 2021, 21, 707–725. [Google Scholar] [CrossRef]
  20. Pang, Y.X.; Liu, W.Y.; Jin, Y.Y.; Peng, H.W.; Xia, T.; Wu, Y.H. Adaptive recommendation for MOOC with collaborative filtering and time series. Comput. Appl. Eng. Educ. 2018, 26, 2071–2083. [Google Scholar] [CrossRef]
  21. Wu, L.L. Collaborative Filtering Recommendation Algorithm for MOOC Resources Based on Deep Learning. Complexity 2021, 2021, 5555226. [Google Scholar] [CrossRef]
  22. Bastani, H.; Harsha, P.; Perakis, G.; Singhvi, D. Learning Personalized Product Recommendations with Customer Disengagement. MSOM-Manuf. Serv. Oper. Manag. 2021, 24, 2010–2028. [Google Scholar] [CrossRef]
  23. Binbusayyis, A. Deep Embedded Fuzzy Clustering Model for Collaborative Filtering Recommender System. Intell. Autom. Soft Comput. 2022, 33, 501–513. [Google Scholar] [CrossRef]
  24. Anwar, T.; Uma, V.; Hussain, M.I.; Pantula, M. Collaborative filtering and kNN based recommendation to overcome cold start and sparsity issues: A comparative analysis. Multimed. Tools Appl. 2022, 81, 35693–35711. [Google Scholar] [CrossRef]
  25. Chang, H.H.; Wang, C. Multidimensional Item Response Theory; Springer: Berlin/Heidelberg, Germany, 2011; Volume 76, pp. 504–506. [Google Scholar] [CrossRef]
  26. Li, T.Y.; Jin, L.L.; Wu, Z.B.; Chen, Y. Combined Recommendation Algorithm Based on Improved Similarity and Forgetting Curve. Information 2019, 10, 130. [Google Scholar] [CrossRef]
  27. Gan, Z.G.; Rao, S.; Jiang, X.N.; Yu, C.H.; Yu, B.X.; Li, C.H. Improved FC-LFM Algorithm Integrating Time Decay Factor. Arab. J. Sci. Eng. 2021, 46, 8629–8639. [Google Scholar] [CrossRef]
  28. Tian, X.T.; Liu, F. Capacity Tracing Enhanced Course Recommendation in MOOCs. IEEE Trans. Learn. Technol. 2021, 14, 313–321. [Google Scholar] [CrossRef]
  29. Hussein, A.H.; Kharma, Q.M.; Taweel, F.M.; Abualhaj, M.M.; Shambour, Q.Y. A Hybrid Multi-Criteria Collaborative Filtering Model for Effective Personalized Recommendations. Intell. Autom. Soft Comput. 2022, 31, 661–675. [Google Scholar] [CrossRef]
  30. Albelbisi, N.A. Development and validation of the MOOC success scale (MOOC-SS). Educ. Inf. Technol. 2020, 25, 4535–4555. [Google Scholar] [CrossRef]
  31. Shi, C.M.; Wei, B.T.; Wei, S.L.; Wang, W.; Liu, H.; Liu, J.L. A quantitative discriminant method of elbow point for the optimal number of clusters in clustering algorithm. EURASIP J. Wirel. Commun. Netw. 2021, 2021, 31. [Google Scholar] [CrossRef]
  32. Magron, P.; Fevotte, C. Neural content-aware collaborative filtering for cold-start music recommendation. Data Min. Knowl. Discov. 2022, 36, 1971–2005. [Google Scholar] [CrossRef]
  33. Rahmani, H.A.; Deldjoo, Y.; di Noia, T. The role of context fusion on accuracy, beyond-accuracy, and fairness of point-of-interest recommendation systems. Expert Syst. Appl. 2022, 205, 117700. [Google Scholar] [CrossRef]
Figure 1. Framework of hybrid recommendation for MOOCs.
Figure 1. Framework of hybrid recommendation for MOOCs.
Applsci 13 05548 g001
Figure 2. Sensitivity analysis of weight coefficient α and the k nearest neighbors.
Figure 2. Sensitivity analysis of weight coefficient α and the k nearest neighbors.
Applsci 13 05548 g002
Figure 3. Sensitivity analysis of the number of recommended courses.
Figure 3. Sensitivity analysis of the number of recommended courses.
Applsci 13 05548 g003
Figure 4. Comparative analysis of algorithm performance.
Figure 4. Comparative analysis of algorithm performance.
Applsci 13 05548 g004
Table 1. Data volume in datasets.
Table 1. Data volume in datasets.
DatasetsNumber of
UsersCoursesRatingsReviews
Original605,69341551,192,9271,169,109
Preprocessed30,9573816245,496241,085
Experiment1849295142,48541,741
Table 2. Statistics of course attributes (n = 2951).
Table 2. Statistics of course attributes (n = 2951).
AverageMedianMinMax
Number of materials67.665.226.5134.9
Price296.8302.4130.8689.5
Duration13.412.83.630.6
Number of weeks4.84.71.39.5
Number of teaching assistants1.81.71.16.5
Number of clicks520,539.8375,045.77927.43,622,115.7
Number of views638,246.8420,610.116,648.83,551,475.9
Number of registrations16,640.9141,213.75666.4805,146.5
Number of ratings8799.85751.768.262,734.2
Number of reviews1773.91278.118.311,812.9
Course ratings4.94.94.55
Table 3. A subset of learners’ course rating preferences with memory weight.
Table 3. A subset of learners’ course rating preferences with memory weight.
User_idIndigenous-CanadaVisual-Elements-User-Interface-DesignFundamentals-of-Graphic-DesignGlobal-DiplomacyMoral-PoliticsPhilosophyTerrorismIntro-Common-LawModern-Art-IdeasCraft-of-Plot
81640000000000
102690000000000
146540000000000
155980000000000
263070000000000
281850000000000
477280000000000
603770000000000
887890000000000
907440000000000
Table 4. A subset of data on learners’ course attribute preferences with memory weight.
Table 4. A subset of data on learners’ course attribute preferences with memory weight.
User_idMaterialsPriceDurationWeeksAssistantsClicksViewsRegistrationsRating VolumeReviewsRating
816454 319 13 4 1 562407 894006 281359 12905 2554 5
1026961 331 12 4 2 182944 486581 67486 3152 611 5
1465464 237 10 5 2 145783 719196 114781 3423 917 5
1559878 434 17 4 2 140287 76550 45004 1834 392 5
2630798 390 19 6 2 236303 816381 156626 7623 1023 5
2818586 284 19 6 1 85564 142817 73284 1847 383 5
4772858 311 17 4 2 500932 1697855 201391 18999 2604 5
6037771 336 15 4 2 238566 2433070 437778 34573 6151 5
8878969 364 9 5 1 80350 51049 38149 1058 288 5
9074474 299 16 5 1 156801 306587 102930 3985 878 5
Table 5. A subset of data on learners’ multidimensional capabilities with memory weight.
Table 5. A subset of data on learners’ multidimensional capabilities with memory weight.
User_idEnglishFrenchRussianSpanishPortuguese (Brazilian)Computer-ScienceData-ScienceInformation-TechnologyMath-and-LogicBusinessPersonal-DevelopmentLife-SciencesPhysical-Science-and-EngineeringArts-and-HumanitiesSocial-SciencesLanguage-LearningBeginnerIntermediateAdvanced
81642221069210000000470
102694222113235217400000251515
1465463342513021120001010
155981111061006001000930
2630752242730721100001030
2818574454620140024101510
477283222268011000000460
6037711110911000000000380
8878965555501071002101910
9074421111700010040001010
Table 6. KMO and Bartlett test.
Table 6. KMO and Bartlett test.
KMO Sampling Suitability Quantity0.877
Bartlett’s sphericity testapproximate chi-square29,090.65
degrees of freedom276
significance0.000
Table 7. Interpretation and naming of principal components.
Table 7. Interpretation and naming of principal components.
Course AttributesPrincipal Components after Dimension Reduction
PopularityPriceResourceTimeRating
Number of materials
Price
Course duration
Number of weeks
Number of teaching assistants
Number of clicks
Number of views
Number of registrations
Rating volume
Number of reviews
Course rating
Table 8. Characteristics of clusters.
Table 8. Characteristics of clusters.
ClusterNumberPercentageCharacteristics
132917.79%High price, short class hours, and advanced level
239221.20%Low price, unpopular course, and beginner level
328815.58%Popular and high rating
425613.85%Few materials and low rating
558431.58%Few materials, short class hours, and niche
Table 9. A subset of data on learners similarity.
Table 9. A subset of data on learners similarity.
ta8164102691465415598263072818547728603778878990744
81641.000.210.030.010.02−0.240.420.33−0.17−0.02
102690.211.00−0.050.010.07−0.070.200.06−0.11−0.12
146540.03−0.051.00−0.20−0.14−0.050.15−0.080.010.06
155980.010.01−0.201.000.100.16−0.10−0.010.220.35
263070.020.07−0.140.101.000.190.150.03−0.110.05
28185−0.24−0.07−0.050.160.191.00−0.15−0.270.150.46
477280.420.200.15−0.100.15−0.151.000.40−0.20−0.09
603770.330.06−0.08−0.010.03−0.270.401.00−0.29−0.11
88789−0.17−0.110.010.22−0.110.15−0.20−0.291.000.12
90744−0.02−0.120.060.350.050.46−0.09−0.110.121.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, B.; Liu, L. Personalized Hybrid Recommendation Algorithm for MOOCs Based on Learners’ Dynamic Preferences and Multidimensional Capabilities. Appl. Sci. 2023, 13, 5548. https://doi.org/10.3390/app13095548

AMA Style

Wu B, Liu L. Personalized Hybrid Recommendation Algorithm for MOOCs Based on Learners’ Dynamic Preferences and Multidimensional Capabilities. Applied Sciences. 2023; 13(9):5548. https://doi.org/10.3390/app13095548

Chicago/Turabian Style

Wu, Bing, and Lixue Liu. 2023. "Personalized Hybrid Recommendation Algorithm for MOOCs Based on Learners’ Dynamic Preferences and Multidimensional Capabilities" Applied Sciences 13, no. 9: 5548. https://doi.org/10.3390/app13095548

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop