Next Article in Journal
Hopf Bifurcation in a Delayed Equation with Diffusion Driven by Carrying Capacity
Next Article in Special Issue
LDA-CNN: Linear Discriminant Analysis Convolution Neural Network for Periocular Recognition in the Wild
Previous Article in Journal
Design and Experience of Mobile Applications: A Pilot Survey
Previous Article in Special Issue
Predicting Change in Emotion through Ordinal Patterns and Simple Symbolic Expressions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

One Aggregated Approach in Multidisciplinary Based Modeling to Predict Further Students’ Education

by
Milan Ranđelović
1,
Aleksandar Aleksić
2,
Radovan Radovanović
3,
Vladica Stojanović
4,
Milan Čabarkapa
5 and
Dragan Ranđelović
2,*
1
Science Technology Park, 18000 Niš, Serbia
2
Faculty of Diplomacy and Security, University Union-Nikola Tesla Belgrade, 11000 Beograd, Serbia
3
Department of Forensic Engineering, University of Criminal Investigation and Police Studies, 11000 Beograd, Serbia
4
Department of Information Technology, University of Criminal Investigation and Police Studies, 11000 Beograd, Serbia
5
Faculty of Electrical Engineering, University of Belgrade, 11000 Belgrade, Serbia
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(14), 2381; https://doi.org/10.3390/math10142381
Submission received: 31 May 2022 / Revised: 13 June 2022 / Accepted: 30 June 2022 / Published: 6 July 2022

Abstract

:
In this paper, one multidisciplinary-applicable aggregated model has been proposed and verified. This model uses traditional techniques, on the one hand, and algorithms of machine learning as modern techniques, on the other hand, throughout the determination process of the relevance of model attributes for solving any problems of multicriteria decision. The main goal of this model is to take advantage of both approaches and lead to better results than when the techniques are used alone. In addition, the proposed model uses feature selection methodology to reduce the number of attributes, thus increasing the accuracy of the model. We have used the traditional method of regression analysis combined with the well-known mathematical method Analytic Hierarchy Process (AHP). This approach has been combined with the application of the ReliefF classificatory modern ranking method of machine learning. Last but not least, the decision tree classifier J48 has been used for aggregation purposes. Information on grades of the first-year graduate students at the Criminalistics and Police University, Belgrade, after they chose and finished one of the three possible study modules, was used for the evaluation of the proposed model. To the best knowledge of the authors, this work is the first work when considering mining closed frequent trees in case of the streaming of time-varying data.

1. Introduction

In order to clearly present the content of the research covered in this paper, the introduction is divided into two subchapters: ‘Background’, where the problem and the importance and main goal of the solution are presented and ‘Related studies’, where the current state of the research in the field related to the subject of this paper as well as the key publications are given.
The rest of this paper is organized as follows. Section 1 presents the ‘Introduction’, where a description of the considered problem, the background, objectives and existing research gap, as well as our contribution, including the organization of the paper, are given. It gives the authors’ review of the state-of-the-art literature that deals with the topic of interest together with the motivation of the authors to work on this paper. After that, we present Section 2, ‘Materials and Methods’, as a part of the paper that introduces the problem that the authors solve in this paper, together with the used data material and the methodology that is proposed and adopted in this solution. In Section 3, ‘Results and Findings’, the obtained results are presented using the proposed new model in a concrete case study performed in the Criminal and Police Academy in Belgrade, Serbia. In Section 4, ‘Discussion and Suggestions’, the authors discuss the possibilities of the proposed approach. The end of this paper deals with the concluding remarks in Section 5, which is the ‘Conclusion’.

1.1. Background

Quantification of the importance of relevance of each factor (attribute, criteria) which is included in a usually complex, multidimensional model which describes any problem in many fields of human life, is the subject of research in many fields of sciences. For instance, the scientific disciplines of prediction of the importance of these factors are classification in modern data mining and machine learning, multicriteria decision making and regression as well as correlation analysis in classic mathematics and statistics. It is especially important to pay attention to the need to reduce the dimensionality of these problems to achieve the highest possible accuracy of the model that solves them.
Supervised machine learning, which is used in our proposed model, can be observed as the process where the system improves its performance on the assigned task without additional programming [1]. Supervised learning of understandable knowledge or learning of the concepts (hypothesis, models, classifiers and rules) deals with the induction of logical lawfulness understandable to humans. Since it is also necessary to evaluate the worthiness of implemented knowledge, usually, the set of examples used for supervised learning is divided into two parts: the learning set used for the learning process and the test set used for the testing of acquired knowledge. The basic measurement of the successfulness of acquired (understandable) knowledge is predictive accuracy. It represents the percentage of classification success of new and unconverted examples by using learning rules. Development of some applications of supervised machine learning, above all in construction of the systems based on knowledge and revealing knowledge in information data mining brought to emphasize request for understanding the scientific knowledge. Data mining is the wide-ranging application of technology items relying on several other things such as database technology, statistical analysis as well as artificial intelligence. It gives great commercial value to the penetration of other occupations such as the retail, insurance, telecommunication and power industries use [2].
However, decision making is the process of selection between alternative operational directions to achieve goals and tasks. A multi-criteria decision that we will use in our proposed model which we will propose is one of the most famous branches in decision making with wide application in solving real problems. The presence of various criteria, of which some should be maximized and some minimized, means that the decisions are brought under conflict conditions. These instruments that have to be applied here are more flexible than strict mathematical techniques related to pure optimization.
Analytic hierarchy process-AHP is a mathematical method [3,4] that, compared to the other methods and decision-making techniques, offers the possibility for decision-makers to compare the individual importance of an alternative compared to the others within decision making on an important criterion. The first two phases of the aforementioned process define targets wishing to be achieved by the selection, attributes (criteria) as a base to perform alternative worthiness, determining the importance of the attributes, and a set of available alternatives for selecting the best one. Afterward, the decision maker evaluates available alternatives compared to the selected attributes. Regression analysis is one of the most exploited statistical methods that has great application in the economy and other humanities. To explain, regression as a statistical method [5,6] allows estimations and assessment of a phenomenon based on the value of some other phenomenon or group of phenomena. Therefore x1, x2, … xk are undependable variables influencing and conditioning the size worthiness of dependable variable y. In experimental theory, undependable variables are called factors, while dependable variables are the results of the experiment. In regression analysis, there is the possible use of such units for determining performances, which are using one input or providing one output.
The presence of irrelevant and redundant attributes harms supervised learning performances; hence, the optimum set of attributes for learning makes fairly relevant non-redundant attributes and strongly relevant attributes. Due to the necessity of attributes selection method analysis, the subject of the research in this paper is the possibility of applying regression analysis as a classical statistical technique and AHP as a method of operational research in the process of supervised learning using the ReliefF classificatory modern ranking method. The method of previous learning shall determine sub-set of attributes based on accuracy estimation by applying decision tree classifiers after applying wrapper methods. WEKA [7] data mining tools were used for the application of selected classifiers, including intensive pattern discovery, which has been frequently studied during the last decade.
Evaluation of the proposed methodology was conducted using a case study predicting the choice of one of three possible modules for further study based on the assessment of subjects from the joint first year of study. The information on grades of the first-year graduate students at the Academy of Criminalistics and Police Studies in Belgrade was used as a dataset in this paper. Modules are presented as the attributes set. Analysis of the aforementioned sets should determine the prediction of the further orientation of students after the first year of studies. For each orientation, the meaningfulness of given attributes was determined by the application of traditional and smart methods to achieve an optimal sub-set that is relevant for further students’ orientation. We use two traditional methods in our proposed model-AHP and regression analysis because, in general, we propose aggregation as methodology and because we first aggregate these two methods, which, as it is known, belong to two different types of methodology, i.e., subjective and objective. The comparison of accuracy prediction estimation is presented in the paper. It was performed by application of decision tree classifiers using wrapper methods and a set of attributes acquired by the application of ReliefF classifiers. Attributes whose combination acquires a satisfying prediction model were abstracted by the method of negative elimination; thus, these models were represented as relevant attributes present in both sets. Thereby, this paper presents an algorithm for extracting a minimal set of attributes that represent relevant data for further orientation in this case study. We propose novel label and unlabeled approaches for adaptively mining closed-rooted trees from data streams that change over time. It should be noted that closed patterns are a powerful type of frequent patterns due to the fact that they eliminate unnecessary information.

1.2. Related Studies

The use of statistical regression methods to predict student achievement can be found in the literature based on their success in graduation or entrance exam [8,9]. Moreover, statistical methods are used to determine the parameters for assessing whether a student will withdraw from studies or continue their studies based on the first results [10]. Regression models for predicting student academic performance in an engineering dynamics course are the considering subject of the solution with the same name [11]. Another paper [12] deals with academic progress using models based on linear and logistic regression, employing prior success and demographic factors as predictors.
The multi-criteria AHP method is used in higher education for the selection of candidates for teaching positions [13] where the AHP method is used to determine joint action as well as the priority of individual selection criteria. The selection of doctoral studies, depending on the goals that the applicant wants in his career, was done with the help of AHP [14]. Perspectives are set as a pseudo level in the hierarchy and for each of them, doctoral programs are offered and selected by the candidates depending on their preferences. The paper [15] gives a knowledge management system that recommends an orientation program to students who should choose it, and therefore, helps him be the most efficient in further studies. In the paper [16], comparative analyzes of DEA and AHP were performed, as well as their aggregation according to the importance of the first-year study subjects for selecting an appropriate study program on the example of the Faculty of Organizational Sciences in Belgrade. Significance of the weights of individual subjects was used to rank the benefits of study programs for a particular student, based on the success achieved in subjects that are common to all students in the first year of study. Moreover, very interesting papers are [17,18], in which the effects of the first-year course on student retention and the prediction of student success based on the achieved results are considered, respectively. The time dependence of predictions and ANOVA implementation in fuzzy AHP is discussed in [19].
Machine learning algorithms in [20,21] and data mining techniques in [22,23] are considered as one overview of student performance prediction modeling for further education in both pairs of these papers. Application of machine learning in predicting performance for computer engineering students is subject of the manuscripts [24,25] and data mining techniques are applied in predicting further education for students that study medical curriculum [26]. Additionally, supervised learning application is proposed in one model for student performance prediction [27]. Moreover, one comprehensive consideration of the application of supervised machine learning techniques was conducted in the prediction of students’ performance using different parameters including demographics and social interest [28].
A machine learning approach that uses two techniques, logistic regressions and decision trees, to predict student dropout at the Karlsruhe Institute of Technology is considered in [29]. We could also find such approach in [30], where a decision tree algorithm combined with the linear regression of data classification for students’ evaluation purposes is applied in Turkey.

2. Materials and Methods

These two subchapters are written to allow readers to replicate and build on published results. The proposed model is described in detail in the subchapter Methods, while a detailed description of the data set used for its development and evaluation is described before in the following subchapter Materials.

2.1. Materials

The dataset used in this case study for the evaluation of the proposed model in this paper and used data are taken from the students’ service database at the Academy of Criminalistics and Police Studies-CPA in Belgrade from 2006 to 2016. The analysis encompassed several graduated students in three departments with mutual modules in year one. Out of 290 graduated students at the Academy, 83 graduated from the Department of Safety (which is labeled as SD), 114 students graduated from the Criminology Department (labeled as CD), and 93 graduated from the Police Department (labeled as PD). For each student, we have compiled information on grades from 10 modules (i.e., subjects) in year one. These numbers are presented, along with their labels, in the following Table 1.
As can be seen, descriptive statistics with the average grade of the students at the CPA in the above modules at year one, observed for each department, are presented. For these purposes, the group of statistical methods, based on the methods of descriptive statistical analysis of the observable data, were used. The values of minima (MIN), maxima (MAX), average grade (MEAN), median grade (MEDIAN), mode grade (MODE), standard deviation (STDEV), sample variance (VAR), kurtosis (KURT) and skewness (SKEW) are set in rows of the following Table 2, Table 3 and Table 4.
As can be seen from the presented tables, the minimum and maximum grades of students in all subjects are equal to MIN = 6 and MAX = 10. The next three rows show the so-called dominant characteristics (MEAN, MEDIAN and MODE). Obviously, the students showed the best success in the subject ‘English language’ (module S2), while other subjects have lower values of these statistical indicators. In the next two rows, the values of the so-called variability measure, standard deviations (STDEV) and variance (VAR) are given. In all three sets of observed data, i.e., for all three departments above, slightly statistically-significant differences in the variability of the observed subjects can be seen.
The last two rows of Table 2, Table 3 and Table 4 show the values of kurtosis, i.e., the elongation coefficient (KURT), as well as the skewness coefficient (SKEW). One can notice that most of the observed data (except, for instance, subject S3 in the Police Department) have negative values (KURT < 0). This means that the observed data sets have the so-called platykurtic properties, which means that they are ‘flatter’ compared to a normal distribution. Finally, the obtained values of the skewness coefficient are, in most cases, positive (SKEW > 0), which means a higher frequency (and consequently mode values) of lower student grades is present.

2.2. Methods

The model, which is proposed in this paper, uses a traditional method of regression analysis which is first aggregated with the AHP method and, in this way, the obtained method has been combined using the application of the ReliefF classificatory ranker method of machine learning and the decision trees classifiers J48 has been used for the aggregation purpose. All of these methods and the proposed aggregated model are studied in detail in four subsections of this subchapter.

2.2.1. Traditional Techniques

The problem of prediction in the choice of study direction is observed in this paper, from the point of view of the mathematical approach. The solution to the problem is to determine the impact of individual subjects in the first year as criteria that affect the final success of the student, which is the observed goal and later allows for the creation of a model that is based on the success of individual subjects in the first year and can predict the choice of appropriate study. Practically, this problem, when it comes to the traditional mathematical and statistical approach, is reduced to the well-known in the literature problem of determining the weights of criteria in a multicriteria problem and a linear regression model.
The classification of mathematical methods for determining the weights of criteria is not uniform. Namely, the division between these methods is made following the authors’ concept and the need to solve a practical problem. There are different divisions of methods, such as: statistical and algebraic, direct and indirect, holistic and decomposed, and compensatory and non-compensatory, which can be seen more in [31].
The most important groups of objective and subjective methodologies for determining the weight of the criteria are discussed in [32], as follows:
  • Standard methods of statistical analysis, including the most commonly used regression and correlation analysis as well as variance and factor analysis. These methods require a statistically relevant sample as well as meeting the strict condition of the relationship between the number of members of the observed sample and the number of criteria and imply a normal distribution, which is eliminated by using so-called non-parametric statistical methods;
  • Methods of operational research, most often methods of multicriteria analysis, which are mostly algebraic but also statistically-based methods for determining the weight of the criteria and whose main task is in fact to choose the optimal number of alternatives and optimal solutions. Methods of operational research that are less demanding than statistical, as well as non-parametric statistical methods, are used where the condition of normal distribution in the sample is not met as well as the required sample size and its relationship with the number of considered criteria. When it is necessary to choose alternatives and their ranking and quantitative positioning.
Methods of operational research can be divided into two basic groups [32]: subjective methods, which depend on the influence of decision makers on the difficulty of criteria and objective methods, which are based on the application of objective mathematical, primarily statistical, apparatus to the information contained in the matrix decision making [31,33].
Objective methods from the essential information of each of the evaluated criteria determine their weight according to two principles [34,35,36]:
  • The criterium that has the least variation in the considered cases, i.e., alternatives have the least impact, i.e., weight; this is the principle of contrast intensity within when special methods appear as those in which entropy, standard deviation, variation, etc. are most often used as a measure of contrast intensity;
  • The criterium that is in conflict with a large number of others is more important, and that is the principle of conflict character in which the most well-known method of correlation is the correlation of the coefficient between pairs of criteria.
In the literature, one can find the application of statistical regression methods to predict students’ success [16] based on their success in graduation or the entrance exam [8,19]. Moreover, statistical methods are used to determine the parameters for assessing whether a student will withdraw from their studies or continue their studies based on the first results [10].
Supervised learning is considered through its main aim to obtain the best prediction results in [37,38] and as one overview in [39]. In [24,40], the applications of machine learning in predicting performance for computer engineering students and disease prediction were considered, respectively.

Regression Analysis

Regression analysis is a method for examining the influence of more different independent variables on one dependent variable to determine the analytic form of this connection, i.e., the model which will be used in analytical and predictive applications. It has a deterministic model in which each value of independent variables exists exactly one value of the dependent variable and which can be given in the form:
y = a + b 1 x 1 + b 2 x 2 + b 3 x 3 + + b p x p
where bi are partial coefficients of regression.
In the case when we have a non-explained variation of the dependent variable in the form of error which we notate as a consequence of the influence of random action or non-including all existing independent variables, we can present multiple regression in the form:
    y = a + b 1 x 1 + b 2 x 2 + b 3 x 3 + + b p x p + e i
The calculation of parameters a and bi in Equation (2) can be executed with the method of least squares with minimization of the sum of squares of the residuals, and in a general case, notated as:
i = 0 n ( y i a b 1 x 1 i b 2 x 2 i Λ b p x p i ) 2
In practice, an algebraic algorithm for solving arising system of equations is rarely in use compared with the well-known Gaussian method of multiplication or with the analysis of variance-ANOVA that is mostly used in empiric research.
One of the particularly important problems that may occur in the multiple regression model is a multi-correlation. It manifests as an intercorrelation between explanatory variables and usually leads to erroneous results of regression analysis [41,42]. In order to check the multi-correlation in the regression model that we obtained here, two diagnostic tools were used and described in detail below (please see Section 3 ‘Results and Findings’). Firstly, the correlation coefficients between all predictor variables were computed, and the weak or moderate correlation was noticeable. In addition, for the purpose of double-checking, another measure of multi-correlation, a well-known variance inflation factor (VIF), was also used.

Multi-Criteria Decision Making

Multi-criteria decision making is a field that is receiving great importance in the last two decades since each process requires contemplation of numerous criteria that are often in conflict or stated in different measurement units. Since the 1960s, further on, a various number of multi-criteria analyses have been developed and could be classified on several grounds. One of the most important multi-criteria classification methods of decision making was conducted by Hwang and Yoon [43], who had classified 17 different methods per type and relevant features of information brought by the decision makers.
According to the type of information, all stated methods are divided into two groups:
  • Methods without information on attributes,
  • Methods requesting certain attribute information.
There are several ways to enable attribute transformation and adjust them to the models of multi-criteria decision making: conversion of the attributes to the scale interval, normalization of the attributes and assembling of the appropriate weight set. Another way of transformation is the assignment of an appropriate weight set, and it is mostly used in situations when it is necessary to determine the relevant importance of certain attributes. For the N criterion, the weight set is:
t T = ( t 1 , t 2 , , t j , , t n )           where               j = 0 n t j = 1
Among numerous estimation techniques of the relative importance of certain abstracted attributes are methods of natural vectors, method of weight in smallest quadrants, entropy method, etc. The algorithm of the AHP method can be described as a structural analysis of one complex decision-making problem that contains more criteria, more alternatives, even to have more decision makers (decision-making group), determining relevant weight criteria and alternatives per level and forming a final alternative outcome (rang alternative). This process is one of the well-known methods of multi-criteria decision making, which is mostly used in cases when there is the possibility of a hierarchical structure of relevant criteria. This method was created in the 1970s by Thomas Saaty [44].
It is important to note that, since the main part of the AHP method is pairs’ comparison of hierarchy elements and formation of appropriate local reciprocal numerical matrices, from which the weights of compared elements are determined, these matrices with the calculated element’s weights carry the measurable information on consistency used by the consistency ratio (CR). A model that provides CR < 0.1 is considered a good one.

2.2.2. Knowledgeable Data Analysis

The basic property differing deep data analysis from the traditional or ‘ordinary’ analysis is the application of machine learning. Intelligent data analysis is another term for deep data analysis. Adjective ‘intelligent’ emphasizes that this data analysis is based on artificial intelligence proceedings. Attribute selection is the field that was developed within the framework of sample recognition in mathematical statistics [45], knowledge revelation [46], machine learning [47,48], especially neuron grids, and many other fields. The fundamental task of the attributes’ selection is the reduction of spatial dimensionality and removal of redundant, irrelevant and disrupted data that speeds up the operation of the learning algorithm, improves the data quality and increases the accuracy of taught knowledge. When selecting attributes, there are two approaches that are different in what they need to achieve. The first approach is finding and ranking the sub-set of the attributes useful for the construction of the quality model. Another approach is finding or ranking all potentially important attributes. Using a sub-set of potentially important attributes for model construction is presented as a sub-optimal approach with the construction of some models [49]. Furthermore, the issue of usefulness versus the importance of the attributes could be very interesting [50].

Attribute Selection Using ReliefF Algorithm

The complexity of group correlation analysis is a result of a huge number of combinations of the attributes, which relations should be considered as O ( 2 N ) , where N is the number of attributes in the model [51]. Due to the usual great complexity, the approximation is applied. For example, only partial analysis of individual attributes with the class O ( N ) , or analysis of only some of the possible combinations (interaction of length 2 or 3 attributes) is performed. Entropy is a commonly used measure in information theory [52], which characterizes the purity of an arbitrary collection of examples. The entropy measure is considered a measure of the system’s unpredictability. The entropy of Y is:
  H ( Y ) = y Y p ( y ) l o g 2 ( p ( y ) )
where p ( y ) is the marginal probability density function for the random variable Y . If the observed values of Y in the training data set S are partitioned according to the values of a second feature X , and the entropy of Y concerning the partitions induced by X is less than the entropy of Y before partitioning, then there is a relationship between features Y and X . The entropy of Y after observing X is then:
H   ( Y X ) = x X p ( x ) y Y p ( y | x ) l o g 2 p ( y | x )
where p ( y | x ) is the conditional probability of y for given x .
ReliefF is a filtering method with proceedings of attribute ranking. It is based on the procedure of the k-nearest neighbors (k-NN). Figure 1 shows the ReliefF algorithm. This algorithm estimates and ranks each attribute with the global grade function [−1,…, 1]. Weight calculation could be performed based on the probability of the nearest neighbors from two different classes with different values of the attributes as well as on the probability of two neighbors from the same class having the same value of the attributes. The function diff (Attribute; Instance1; Instance2) calculates the difference between the values of the attribute for two instances.
When considering discrete attributes, the difference is either 1 (when values are different) or 0 (when values are the same), while in continuous attributes case, the difference is an actual difference normalized to the interval [0, 1]. Kononenko notes in [53] that the higher value of m (the number of sampled instances), the more reliable the ReliefF’s estimates are. Of course, it should be noted that the increasing m rises the running time.

2.2.3. Classification Technique and Algorithms

Classification is the task of data mining that administers the separation of data set examples into previously determined classes of output variables based on the value of input variables [54]. Classification of some objects is based on finding similarities with priory-determined objects belonging to different classes, whereas the similarity of two objects is determined by analyzing their characteristics. The task is to design a model based on the characteristics of the objects with a previously known classification. That model shall represent the ground to perform the classification of new objects. In the problem of classification, several classes are known in advance and limited. By testing, the classifier performs classification of the test set examples in priory determined classes of attributes. If the classifier makes some errors in test data or if there is a higher percentage of wrongly classified examples, the conclusion is that a wrong and unstable model is created. In that case, it is necessary to perform improvements to the model by modifying the applied classification process. Up-to-date research shows that most applied classifiers include Bayes networks, decision trees, neural networks, support vector machines, and K–nearest neighbors [55].
Decision trees are well-known classification techniques considering that they include several ways of creation of easily interpreted trees used for the classification of categories and numerical values of the attributes. These classification methods perform the division of data to nodes and leaves until the entire set of data is being analyzed. The well-known algorithms are ID3 [56] and C4.5 [57]. Algorithm C4.5 for decision-making tree induction is developed on the basis of the ID3 algorithm, with multiple significant improvements compared to the basic algorithm: imperfect operation with continual attributes and emitted values of the attributes, new estimation of gain ration quality, and simplification of the taught tree due to the increasing classification accuracy of new examples. It is available as an independent program as the objective module (library MLC++) for usage within the other systems for supervised learning and intelligent data analysis. J48 algorithm of Weka software is a popular machine learning algorithm based upon Quilan’s C4.5 algorithm [58].

Classifiers Estimation

Classifiers estimation enables the prediction of performances, selection of the best and behavioral assessment of multiple different classifiers. Cross-validation is a statistical method of assessing and comparing learning algorithms by dividing data into two segments: one is used to learn or train a model, and the other one is used for model validation. In standard cross-validation, the training and validation sets must cross over in successive rounds such that each data point has a chance of being validated.
The basic form of cross-validation is k-fold cross-validation. When we talk about the k-fold cross-validation, the whole data is partitioned into k equally (or nearly equal) sized segments/folds. Afterward, k iterations of training and validation are completed in such a way that within each iteration, a different fold of the data is held-out for validation while the remaining k = 1 folds are used for learning. In most cases in data mining and machine learning, 10-fold cross-validation (k = 10) is the most common one. J48 classifier with the corresponding parameters is used and evaluated with this mostly used type of cross-validation.
The biggest number of measurements for classifying models’ estimation is related to classification issues with two classes. That does not represent a specific limitation for the application of those measurements, taking into consideration that the issues with the higher number of classes can be shown in the form of a sequence of problems with two classes. Each of them is separately abstracted for one of the classes as the target ones, a the data set is divided into a positive and negative example of the target class, but the negative ones are the examples of all other classes. Information Retrieval (IR) measures community has developed three measures in everything according to Figure 2.
In Figure 2 so called a 2 × 2 confusion matrix which enables calculation formulas is presented:
TP + TN + FN + FP = N
where N is the total number of members in the considered set which will be classified.
It is necessary to notice that these numbers are counts, i.e., integers, not ratios or fractions.
The accuracy, precision, recall and F1 measure can be respectively calculated as:
Accuracy = ( TP + TN ) / N
Precision = TP / ( TP + FP )  
Recall ( Sensitivity ) = TP / ( TP + FN )  
F-measure = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
An evaluation of the prediction performance of a classifier is carried out using the method called Receiver Operating Characteristic (ROC) curves, which represents the rate of false-positive cases and the rate of true-positive cases on the OX and the OY axes, respectively. ROC curves and parameter AUC (area under the curve) enable best-way evaluation for checking each classification model in Figure 3 [59].
In theory, it is known that for the values of AUC = 70%, it is said that the classification is good and sometimes leads to the highest possible values of 100%.

2.2.4. Proposed Model for Selection of the Relevant Attributes

In the phase of data preparation, in the data set for three departments, the Attribute Department was added, and its value was appointed as binominal. For the Attribute Department, the (yes, no) values were determined. Attribute Department is marked as class and response variable, respectively. Educational information is mostly transparent since they do not contain wrong values collected automatically (log files, database with grades, database of learning system). As the data with the average grade of the year one module at the CPA on all three bases were in numerical form for each graduated student, it was not necessary to apply some of the standard discretization filters during the data preparation.
After the data were acquired and extracted into one unique direction selection, the data mining algorithm was applied as a method and used to estimate and select the attributes. The aim is to extract irrelevant and redundant attributes from the learning data set. In order to keep the data learning set with the desirable attributes only, the measure of attributes evaluation compared to the classification problem is necessary. Filtering methods encompass techniques for attribute values assessment relying on heuristics based on general features of data.
The proposed model selects relevant attributes by combining acquired aggregation results obtained using the ReliefF algorithm, shown in Figure 4.
A selection of the attributes using various filtering techniques that are performing ranking according to the importance estimation is carried out in the WEKA system so that the attributes are ranked and assessed using the current training set. For the importance assessment of the attributes, the ReliefF algorithm with the ranker method was used. Extraction of the relevant attribute set from the whole set was performed by the negative elimination process [60,61].
The weakest ranked attribute is extracted in the next step of attribute ranking. After each extraction, the worst-ranked attribute was selected in WEKA software, and the estimation accuracy that is unique for all steps within these proceedings for the selected learning method was chosen.

3. Results and Findings

For determination of the relevant attributes for selection of the study program at the CPA in Belgrade, the evaluation aggregation was used, as a combination of regression analysis and subjective AHP method, respectively, in order to points gained grade and eliminate the bad side of both applied technologies. In order to calculate relevant attributes, the following procedure was used:
Step 1: In AHP analysis, the procedure has been started from the data from Table 2, Table 3 and Table 4 of the average grades per subject for each department individually, safety, criminology and police, respectively. They are compared in pairs. The result of the comparison by pairs is the formation of a reciprocal matrix in accordance with the preferences defined by the Saaty scale. Saaty scale is a liqueur type and contains divisions from 1 to 9, where 1 is the same meaning of the criteria, and 9 is the absolute meaning of one in relation to the other (even numbers are intermediate divisions on the scale). Moreover, the scale contains reciprocal values, for example if p1: p2 = 5, then P2: P1 = 1/5. What importance will be given to a certain ratio of average grades depends on the subjective opinion of the decision maker. Logically, the highest possible grade is 10, and the lowest positive grade is 6. Hence, the ratio of the significance of subjects whose grades are P1 = 10 and P2 = 6 can be determined as 10:6 = 1.66667, and, on the Saaty scale, it is the absolute significance of P1 in relation to P2 and enter 9. Conversely, the ratio P2: P1 is the lowest possible preference 6:10 = 0.6. Therefore, it is necessary to distribute preferences in the interval from 0.6 to 1.66667, where the same meaning of the criterion must be marked with 1. Precisely, because it is necessary to distinguish between similar estimates (approximate values), the divisions are distributed so that the final values occupy a larger interval. Values from 0.88 to 1.10 are divided into several smaller intervals. In that way, reciprocal matrices were formed for each direction, and further procedure of applying the AHP method involved summing each of the columns of the reciprocal matrix, so the next matrices were constituted, which were normalized values of the reciprocal matrix (each coefficient divided by the sum of the corresponding column). In the end, all normalized coefficients are added into rows and the weighting coefficient is obtained as the quotient of the sum of the rows, and the number of criteria used in the model.
AHP analysis was implemented by using VCO (multi-criteria decision-making software package of the Faculty of Organizational Sciences in Belgrade) using hierarchy and AHP analysis model [62]. Obtained results are shown in Table 5.
Reciprocal AHP numerical matrices for all departments provide CR < 0.1, so the model can be considered as a good one.
Step 2: Regression analysis was implemented by using model (1) of linear regression in the SPSS software package [63,64], and the results are shown in Table 6. One can notice that the last row contains the estimated values of the multiple coefficients of determination (R2). It represents a widely used quantitative measure of agreement of the theoretical, fitting model in relation to the given set of empirical data. The estimated values of R2 confirm that the obtained regression model can be an adequate model of dependency for the given subjects S1,…, S10.
As already pointed out in the previous section, one of the most important problems that arise here is the potential multi-correlation of the predictors in the above regression models. In order to determine the intercorrelation relationship between the predictors Si and Sj,, where i , j = 1 , , 10 , the following Table 7, Table 8 and Table 9 show the estimated values of their correlation coefficients (Rij).
As it can be seen, students’ grades at the Department of Safely satisfy the condition max Rij < 0. 7, when i j . On the other hand, in the case of the other two departments, the inequality max Rij < 0.6 holds. Thus, there is only a weak or (in the ‘worst case’) moderate correlation between students’ grades in different subjects in all three departments.
Multicollinearity can be expressed in another way, using the so-called variance inflation factor (VIF), which represents the measure of the severity of multicollinearity in regression analysis. Mathematically, for a multiple regression model given by Equation (1), the VIF can be expressed using the coefficient of determination ( R i 2 ) of the new multiple regression model with one explanatory variable (xi) as the response variable and the other variables xj ( j i ) as its explanatory variables:
V I F = 1 1 R i 2
The calculated VIF values of each of the three previously obtained regression models are shown in the following Table 10.
It can be seen that for all these three sets of data, i.e., the grades of students from different departments, the condition max ( V I F ) < 4 is valid. This result is fully consistent with both practical and theoretical interpretations of this coefficient [65,66]. In other words, the values of VIF, obtained in this way, indicate that the multi-correlation between the predictors in the previously obtained regression models is not emphasized and, therefore, can be neglected.
Step 3: Evaluation aggregation. Aggregated evaluation measures are acquired by the simple application of arithmetic middle of acquired AHP results and regression analysis in order to objectify the acquired grade and eliminate the bad sides of both applied methodologies. The results are shown in Table 11. So as to clarify, the evaluation aggregation set is given as follows:
a g g r i k = ( r e g r e s i o n _ a n a l y s i s i k + A H P i k ) / 2
Step 4: According to the proposed model of aggregation of traditional methods of regression and AHP with the method of classification from the group of machine learning algorithms which is given in previous Section 2.2.4, we apply the ReliefF algorithm with the ranker method for importance assessment of the attributes. After that, we apply accuracy evaluation for both groups of methods using decision trees classifiers J48 and, for relevant attributes, we take the number of the most important attributes, for which the deviation of accuracy for these two groups of methods is the smallest [67,68,69,70]. The results for the Department of Safety on CPA are given in Table 12, Table 13 and Table 14.
Step 5: By aggregation evaluation for the same department, we have acquired the following measurement values shown in Table 14.
The number of the important factors should be 4 because the deviation of the accuracy of the Department for Safety dataset with ranked attributes by traditional aggregation and with classification is ε = 0.854–0.876 = 0.022, and it is less than when 3 of them are taken as important factors because then it is ε = 0.883–0.819 = 0.064.

4. Discussion and Suggestions

By applying attribute rank gained by aggregation, it could be noticed that the set of 4 attributes gave the highest mid-estimation values for the Department of Safety. Attributes presented as relevant on the basis of this criterion are: (S2, S10, S8 and S6). Acquired mid-accuracy values with a set of attributes that was decreased by negative elimination of the lowest ranked by applying both of the techniques for the Department of Safety are shown in Figure 5.
Presented accuracy results for both cases show that these proceedings enable the discovery of the approximate accuracy values on the grounds, which we could determine an equal number of attributes in both sets. The comparison was performed in order to achieve the aforementioned, i.e., differences of set accuracy of ranking attributes by aggregation technique PrecA(k) and set of ranked attributes by applying ReliefF algorithm PrecR(k). In consideration of accuracy precision ε, the values of sets with sequences of 1 to 10 values were not taken into consideration. On the basis of acquired results (Figure 5), the accuracy precision is visible by applying both ranking methods, the smaller with the set of 4 attributes. For the other two departments, the closest accuracy by application of both methods was acquired by the set of 8 attributes for the Criminology Department and the set of 9 attributes for the Police Department, shown in Figure 6 and Figure 7, respectively.
Set of the appropriate number of attributes with ranking coefficient that we have acquired on the basis of previous figures for all Departments is shown in the Table 15.
Discovery of knowledge by analysis process of acquired results could be determined using the estimation accuracy on the same information by application of tree decision for all three departments, in the case of ReliefF PrecR(k) classifier application compared to the aggregation method PrecA(k) has expressed better results. The number of attributes k was less, and estimation accuracy Prec(k) is better:
  • Criminology Department: estimation accuracy is 84.5 with only 4 relevant attributes-PrecR(4) = 84.5% comparing to the accuracy of 82.5 with the set of 9 attributes-PrecA(9) = 82.5%;
  • Police Department: estimation accuracy is 93.8 with the set of 5 attributes-PrecR(5) = 93.8% comparing to the accuracy of 83.0 with 8 relevant attributes-PrecA(8) = 83.0%;
  • In the case of the Department for Safety, the situation is different. Higher precision of 88.3 was achieved by the set of 3 attributes-PrecR(3) = 88.3% extracted on the basis of their coefficients acquired by ReliefF classifier than the set of 9 relevant attributes extracted by aggregation of 83.0-PrecA(9) = 83.0%;
  • Comparing sets of extracted attributes, we can observe that both methods as the same attributes give: Police Department-8, Department for Safety-1 and Criminology Department-7.
However, if we take a look at the smallest difference of ε achieved in the mid-value of estimation accuracy by applying attributes ranking by aggregation PrecA(k) and application of ReliefF classifier PrecR(k), we have accomplished the following aberrance: Department for Safety’s value is ε = 0.022 (for the set of 4 attributes), Criminology Department’s value is ε = 0.001 (for set of 8 attributes) and Police Department’s value is ε = 0.000 (for set of 9 attributes). By comparing both sets of attributes with all three departments, we could extract the set of the same attributes that represent relevant modules when selecting the Department at the CPA. In Table 15 (marked with gray), it is clearly visible that S8-Constitutional law, S2-English language (excluding the case of application of ReliefF classifier to Safety department) and S5-Introduction to the law (excluding the case of application of aggregation to Safety department) are important for all three departments, that the modules S3-Criminal law general part, S7-Police equipment and S10-IT are important for the Police Department and Criminology Department. For the Safety Department and Police Department (excluding the case of application of aggregation to the Safety Department), module S1-Fundamentals of economics is also important.

5. Conclusions

The main hypothesis of the authors was to show the possibility of the development of one aggregated model having better characteristics than integrated techniques when they are used independently. The techniques which were considered in this paper solved the problem of determination and prediction of the importance of attributes that describes one multidimensional and multicriteria process. The evaluation was carried out on the case study to the determination of students’ future behavior regarding the choice of one of the three offered modules at the Criminal and Police University, Belgrade. The method of feature selection was used to reduce the number of attributes, thus increasing the accuracy of the model. As traditional techniques, the authors demonstrated the usage of the method of regression analysis combined with the AHP mathematical method. The application of ReliefF classificatory ranking method of machine learning was used together with decision tree classifiers J48 for aggregation purposes. The evaluation results have shown the advantage and supremacy of the proposed model. The authors have claimed that the proposed model has no significant limitations, and we will consider the inclusion of n-modular redundancy into it in our future work related to this topic.
It is important to mention that the obtained analysis of the personal contribution of the students via achieved results on the first-year subject modules can help in the process of achieving relevant attributes, which are, in this case, related to the determination of the robustness of the first-year modules that could be prearranged by the experts at the Ministry of Interior, The Republic of Serbia as well as adequate support in determination of toughness that is given to the first-year modules through the subject accreditation process with a corresponding number of European Credit Transfer and Accumulation System (ECTS) credits.

Author Contributions

Conceptualization, Investigation, Validation: M.R. and A.A.; Project administration: R.R. and V.S.; Writing—review & editing, Formal analysis, Software: M.Č.; Supervision, Methodology, Writing—original draft: D.R. All authors have read and agreed to the uploaded version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors thank the Nis Science and Technology Park for their support in the publishing of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hart, A. Machine Induction as a Form of Knowledge Acquisition in Knowledge Engineering; Machine Learning: Principles and Techniques; Chapman and Hall: London, UK, 1989. [Google Scholar]
  2. Haiyang, Z. A Short Introduction to Data Mining and Its Applications. In Proceedings of the Management and Service Science (MASS’11), International Conference IEEE, Wuhan, China, 12–14 August 2011; pp. 1–4. [Google Scholar]
  3. Saaty, T.L. Multicriteria Decision Making: The Analytic Hierarchy Process; McGraw-Hill: New York, NY, USA, 1980. [Google Scholar]
  4. Vaidya, O.S.; Kumar, S. Analytic hierarchy process: An overview of applications. Eur. J. Oper. Res. 2006, 169, 1–29. [Google Scholar] [CrossRef]
  5. Harrell, F.E. Regression Modeling Strategies: With Applications to Linear Models, Logistic Regression, and Survival Analysis; Springer: New York, NY, USA, 2001. [Google Scholar]
  6. Montgomery, D.C.; Peck, E.; Vining, G. Introduction to Linear Regression Analysis, 6th ed.; Wiley: New York, NY, USA, 2021. [Google Scholar]
  7. WEKA Software; The University of Waikato: Hamilton, New Zealand. 2015. Available online: http://www.cs.waikato.ac.nz/ml/weka (accessed on 30 May 2022).
  8. Green, B.; Johnson, C.; McCarthy, C. Predicting Academic Success in the First Year of Chiropractic College. J. Manip. Physiol. Ther. 2003, 6, 40–46. [Google Scholar] [CrossRef]
  9. Rothstein, C. College performance predictions and the SAT. J. Econom. 2004, 121, 297–317. [Google Scholar] [CrossRef] [Green Version]
  10. Montmarquette, C.; Mahseredjian, S.; Houle, R. The determinants of university dropouts: A bivariate probability model with sample selection. Econ. Educ. Rev. 2001, 20, 475–484. [Google Scholar] [CrossRef]
  11. Huang, S.; Fang, N. Regression models of predicting student academic performance in an engineering dynamics course. In ASEE Annual Conference and Exposition, Conference Proceedings; American Society for Engineering Education (ASEE): Louisville, KY, USA, 2010; pp. 151026.1–151026.17. ISBN 9781617820205. [Google Scholar] [CrossRef]
  12. Ayan, R.; Noel, M.; García, M. Prediction of University Students’ Academic Achievement by Linear and Logistic Models. Span. J. Psychol. 2008, 11, 275–288. [Google Scholar] [CrossRef] [Green Version]
  13. Saaty, L.T. Analytical Planning: The Organization of Systems; Pergamon Press: Oxford, UK, 1985. [Google Scholar]
  14. Tadisina, S.K.; Troutt, M.D.; Bhasin, V. Selecting a Doctoral Programme Using the Analytic Hierarchy Process—The Importance of Perspective. J. Oper. Res. Soc. 1991, 42, 631–638. [Google Scholar]
  15. Delibašić, B.; Suknović, M.; Stanaćev, N. Menadžment Znanja Pri Izboru Odgovarajućeg Smera na Studijama; INFO-M, 2005; Volume 4, br. 14, str. 23–27. Available online: https://scindeks.ceon.rs/article.aspx?artid=1451-43970514023D (accessed on 30 May 2022).
  16. Savić, G.; Makajić-Nikolić, D.; Ranđelović, D.; Ranđelović, M. Study program selection by aggregated DEA-AHP measure. Metal. Int. 2013, 1, 169–174. [Google Scholar]
  17. Strayhorn, T.L. An examination of the impact of first-year seminars on correlates of college student retention. J. First Year Exp. Stud. Transit. 2009, 21, 9–27. [Google Scholar]
  18. Kovačić, Z.J. Predicting student success by mining enrolment data. Res. High. Educ. J. 2012, 15, 1–20. [Google Scholar]
  19. Rajpurt, H.; Milani, A.; Labun, A. Including time dependency and ANOVA in decision making using the revised fuzzy AHP: A cese study on wafer fabrication process selection. Appl. Soft Comput. 2011, 11, 5099–5109. [Google Scholar] [CrossRef]
  20. Sandra, L.; Lumbangaol, F.; Matsuo, T. Machine Learning Algorithm to Predict Student’s Performance: A Systematic Literature Review. TEM J. 2021, 10, 1919–1927. [Google Scholar] [CrossRef]
  21. Yakubu, M.N.; Abubakar, A.M. Applying machine learning approach to predict students’ performance in higher educational institutions. Kybernetes 2021, 51, 916–934. [Google Scholar] [CrossRef]
  22. Lynn, N.D.; Emanuel, A.W.R. Using Data Mining Techniques to Predict Students’ Performance. A Review. IOP Conference Series: Materials Science and Engineering. In Proceedings of the 6th International Conference on Industrial, Mechanical, Electrical and Chemical Engineering—ICIMECE 2020, Solo, Indonesia, 20 October 2020; Volume 1096. [Google Scholar]
  23. Siddique, A.; Jan, A.; Majeed, F.; Qahmash, A.I.; Quadri, N.N.; Wahab, M.O.A. Predicting Academic Performance Using an Efficient Model Based on Fusion of Classifiers. Appl. Sci. 2021, 11, 11845. [Google Scholar] [CrossRef]
  24. Buenaño-Fernández, D.; Gil, D.; Luján-Mora, S. Application of Machine Learning in Predicting Performance for Computer Engineering Students: A Case Study. Sustainability 2019, 11, 2833. [Google Scholar] [CrossRef] [Green Version]
  25. Hwang, G.J.; Xie, H.; Wah, B.; Gasevic, D. Vision, challenges, roles and research issues of Artificial Intelligence in Education. Comput. Educ. Artif. Intell. 2020, 1, 100001. [Google Scholar] [CrossRef]
  26. Baars, G.J.A.; Stijnen, T.; Splinter, T.A.W. A Model to Predict Student Failure in the First Year of the Undergraduate Medical Curriculum. Health Prof. Educ. 2017, 3, 5–14. [Google Scholar] [CrossRef]
  27. Hashim, A.S.; Awadh, W.A.; Hamoud, A.K. Student Performance Prediction Model based on Supervised Machine Learning Algorithms. IOP Conference Series: Materials Science and Engineering. In Proceedings of the 2nd International Scientific Conference of Al-Ayen University (ISCAU-2020), Thi-Qar, Iraq, 15–16 July 2020; Volume 928. [Google Scholar]
  28. Tomasevic, N.; Gvozdenovic, N.; Vranes, S. An overview and comparison of supervised data mining techniques for student exam performance prediction. Comput. Educ. 2019, 143, 103676. [Google Scholar] [CrossRef]
  29. Kemper, L.; Vorhoff, G.; Wigger, B.U. Predicting student dropout: A machine learning approach. Eur. J. High. Educ. 2020, 10, 28–47. [Google Scholar] [CrossRef]
  30. Ahmed, A.; Rizaner, A.; Ulusoy, A.H. A Decision Tree Algorithm Combined with Linear Regression for Data Classification. In Proceedings of the 2018 International Conference on Computer, Control, Electrical, and Electronics Engineering (ICCCEEE), Khartoum, Sudan, 12–14 August 2018; pp. 1–5. [Google Scholar] [CrossRef]
  31. Milićević, M.; Župac, G. An objective approach to determining the weight of criteria. Mil. Tech. Cour. 2012, XV, 39–56. [Google Scholar] [CrossRef] [Green Version]
  32. Ranđelović, D.; Stanković, J.; Janković- Milić, V.; Stanković, J. Weight Coefficients Determination Based on Parameters in Factor Analysis. Metal. Int. 2013, 3, 128–132. [Google Scholar]
  33. Zopounidis, C.; Doumpos, M. Multi-criteria classification and sorting methods: A literature review. Eur. J. Oper. Res. 2002, 138, 229–246. [Google Scholar] [CrossRef]
  34. Zeleny, M. A concept of compromise solutions and the method of displaced ideal. Comput. Oper. Res. 1974, 1, 479–496. [Google Scholar] [CrossRef]
  35. Diakoulaki, D.; Mavrotas, G.; Papayannakis, L. Determining objective weights in multiple criteria problems: The Critic method. Comput. Oper. Res. 1995, 22, 763–770. [Google Scholar] [CrossRef]
  36. Ganjavi, A. Weakness of standard deviationof normalized scores as weight in multicriteria decision making. In Proceedings of the 30th Annual ASAC Conference, Winnipeg, MB, Canada, 25–28 May 2002; pp. 48–55. [Google Scholar]
  37. Sarker, I.H. Machine Learning: Algorithms, Real-World Applications and Research Directions. SN Comput. Sci. 2021, 2, 1–21. [Google Scholar] [CrossRef]
  38. Hastie, T.; Tibshirani, R.; Friedman, J. Overview of Supervised Learning. In The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer: New York, NY, USA, 2009. [Google Scholar] [CrossRef]
  39. Dridi, S. Supervised Learning—A Systematic Literature Review. 2021. Available online: https://www.researchgate.net/publication/354996999_Supervised_Learning_-_A_Systematic_Literature_Review?channel=doi&linkId=6157249e61a8f46670997bc9&showFulltext=true (accessed on 30 May 2022).
  40. Uddin, S.; Khan, A.; Hossain, E.; Moni, M.A. Comparing different supervised machine learning algorithms for disease prediction. BMC Med. Inform. Decis. Mak. 2019, 19, 281. [Google Scholar] [CrossRef]
  41. Altland, H.W. Regression Analysis: Statistical Modeling of a Response Variable. Technometrics 1999, 41, 367–368. [Google Scholar] [CrossRef]
  42. Kim, J.H. Multicollinearity and misleading statistical results. Korean J. Anesthesiol. 2005, 72, 558–569. [Google Scholar] [CrossRef] [Green Version]
  43. Hwang, C.L.; Yoon, K.S. Multiple Attribute Decision Making: Methods and Application; Springer: Berlin, Germany, 1981. [Google Scholar]
  44. Saaty, T. The Analytical Hierarchy Prosess; McGraw-Hill: New York, NY, USA, 1980. [Google Scholar]
  45. Guyon, I.; Gunn, S.; Nikravesh, M.; Zadeh, L. Feature Extraction: Foundations and Applications—Studies in Fuzziness and Soft Computing; Springer: New York, NY, USA, 2006. [Google Scholar]
  46. Witten, I.H.; Frank, E. Data Mining: Practical Machine Learning Tools and Techniques, 2nd ed.; Morgan Kaufmann: San Francisco, CA, USA, 2005. [Google Scholar]
  47. Kubat, M.; Bratko, I.; Michalski, R. A Review of Machine Learning Methods; John Wiley& Sons: New York, NY, USA, 1996; pp. 1–72. [Google Scholar]
  48. Cherkassky, V.; Mulier, F.M. Learning from Data: Concepts, Theory and Methods, 2nd ed.; Wiley-IEEE Press: New York, NY, USA, 2007. [Google Scholar]
  49. García-Pedrajas, N.; Fyfe, C. Immune Network based Ensembles. In Proceedings of the European Symposium on Artificial Neural Networks (ESANN’2006), Bruges, Belgium, 26–28 April 2006; pp. 437–442. [Google Scholar]
  50. Kohavi, R.; John, G.H. Wrappers for feature selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef] [Green Version]
  51. Almuallim, H.; Dietterich, T.G. Learning with many irrelevant features. In Proceedings of the Ninth National Conference on Artificial Intelligence (AAAI-91), Anaheim, CA, USA, 14–19 July 1991; AAAI Press: Anaheim, CA, USA, 1991; pp. 547–552. [Google Scholar]
  52. Abe, N.; Kudo, M. Entropy criterion for classifier-independent feature selection. Lect. Notes Comput. Sci. 2005, 3684, 689–695. [Google Scholar]
  53. Kononenko, I. Estimating attributes: Analysis and extensions of relief. In Proceedings of the Seventh European Conference on Machine Learning, Catania Italy, 6–8 April 1994; Springer: Berlin, Germany, 1994; pp. 171–182. [Google Scholar]
  54. Abraham Iorkaa, A.; Barma, M.; Muazu, H. Machine Learning Techniques, methods and Algorithms: Conceptual and Practical Insights. Int. Eng. Res. Appl. 2021, 11, 55–64. [Google Scholar]
  55. Benoit, G. Data Mining. Annu. Rev. Inf. Sci. Technol. 2002, 36, 265–310. [Google Scholar] [CrossRef]
  56. Quinlan, J.R. Induction of Decision Trees, Machine Learning 1; Kluwer Academic Publishers: Boston, MA, USA, 1986. [Google Scholar]
  57. Quinlan, J.R. Improved use of continuous attributes in c4.5. J. Artif. Intell. Res. 1996, 4, 77–90. [Google Scholar] [CrossRef] [Green Version]
  58. Quinlan, J.R. C4.5: Programs for Machine Learning; Morgan Kaufmann: San Francisco, CA, USA, 1993. [Google Scholar]
  59. Fawcett, T. ROC Graphs: Notes and Practical Considerations for Data Mining Researchers; Technical Report HP Laboratories: Palo Alto, CA, USA, 2003. [Google Scholar]
  60. Romero, C.; Ventura, S.; Espejo, P.; Hervas, C. Data mining algorithms to classify students. In Proceedings of the 1st International Conference on Educational Data Mining (EDM08), Montreal, Canada, 20–21 June 2008; pp. 20–21. [Google Scholar]
  61. Amrieh, E.A.; Hamtini, T.; Aljarah, I. Mining Educational Data to Predict Student’s academic Performance using Ensemble Methods. Int. J. Database Theory Appl. 2016, 9, 119–136. [Google Scholar] [CrossRef]
  62. VCO Software (2015), Faculty of Organizational Sciences. University of Belgrade. Available online: http://www.odlucivanje.fon.rs/images/stories/download/VKO.zip (accessed on 30 May 2022).
  63. SPSS Inc. Released 2008. SPSS Statistics for Windows, Version 17.0; SPSS Inc.: Chicago, IL, USA, 2008. [Google Scholar]
  64. Wellman, B. Doing It Ourselves: The SPSS Manual as Sociology’s Most Influential Recent Book; Required Reading: Sociology’s Most Influential Books; University of Massachusetts Press: Amherst, MA, USA, 1998; pp. 71–78. [Google Scholar]
  65. Vatcheva, K.P.; Lee, M.; McCormick, J.B.; Rahbar, M.H. Multicollinearity in Regression Analyses Conducted in Epidemiologic Studies. Epidemiology 2016, 6, 227. [Google Scholar] [CrossRef] [Green Version]
  66. Shrestha, N. Detecting Multicollinearity in Regression Analysis. Am. J. Appl. Math. Stat. 2020, 8, 39–42. [Google Scholar] [CrossRef]
  67. Santafe, G.; Inza, I.; Lozano, J.A. Dealing with the evaluation of supervised classification algorithms. Artif. Intell. Rev. 2015, 44, 467–508. [Google Scholar] [CrossRef]
  68. Dimić, G.; Prokin, D.; Kuk, K.; Micalović, M. Primena Decision Trees i Naive Bayes klasifikatora na skup podataka izdvojen iz Moodle kursa. In Proceedings of the Conference INFOTEH, Jahorina, Bosnia and Herzegovina, 21–23 March 2012; Volume 11, pp. 877–882. [Google Scholar]
  69. Yang, S.; Berdine, G. The receiver operating characteristic (ROC) curve. Southwest Respir. Crit. Care Chron. 2017, 5, 34. [Google Scholar] [CrossRef]
  70. Vuk, M.; Curk, T. ROC curve, lift chart and calibration plot. Adv. Methodol. Stat. 2006, 3, 89–108. [Google Scholar] [CrossRef]
Figure 1. ReliefF algorithm.
Figure 1. ReliefF algorithm.
Mathematics 10 02381 g001
Figure 2. Parameters for classification issues with two classes.
Figure 2. Parameters for classification issues with two classes.
Mathematics 10 02381 g002
Figure 3. ROC and AUC curves.
Figure 3. ROC and AUC curves.
Mathematics 10 02381 g003
Figure 4. Suggested model of relevant attributes selection.
Figure 4. Suggested model of relevant attributes selection.
Mathematics 10 02381 g004
Figure 5. Attributes ranked by aggregation of the Department of Safety dataset.
Figure 5. Attributes ranked by aggregation of the Department of Safety dataset.
Mathematics 10 02381 g005
Figure 6. Attributes ranked by aggregation of the dataset of Criminology Department.
Figure 6. Attributes ranked by aggregation of the dataset of Criminology Department.
Mathematics 10 02381 g006
Figure 7. Attributes ranked by aggregation of the dataset of Police Department.
Figure 7. Attributes ranked by aggregation of the dataset of Police Department.
Mathematics 10 02381 g007
Table 1. Common subjects of students at year one.
Table 1. Common subjects of students at year one.
Subjects Label
Fundamentals of economicsS1
English languageS2
Criminal law-general partS3
Criminal law-special partS4
Introduction to the LawS5
Special physical educationS6
SociologyS7
Police equipmentS8
Constitutional LawS9
ITS10
Table 2. Descriptive statistics of the success of students at Department of Safety (sample size is N = 83 ).
Table 2. Descriptive statistics of the success of students at Department of Safety (sample size is N = 83 ).
ItemsSubjects
S1S2S3S4S5S6S7S8S9S10
MIN6666666666
MAX10101010101010101010
MEAN7.298.177.127.247.417.767.027.517.417.37
MEDIAN7877787777
MODE6876776676
STDEV1.3651.0931.0291.2411.2041.2801.1071.3251.2641.428
VAR1.8621.1951.0601.5391.4491.6391.2241.7561.5992.038
KURT−0.695−0.8760.407−0.285−0.402−0.7280.633−0.916−0.405−0.908
SKEW0.7400.2480.9010.7480.5730.4080.9970.4470.7090.664
Table 3. Descriptive statistics of the success of students at Criminology Department (sample size is N = 114 ).
Table 3. Descriptive statistics of the success of students at Criminology Department (sample size is N = 114 ).
ItemsSubjects
S1S2S3S4S5S6S7S8S9S10
MIN6666666666
MAX10101010101010101010
MEAN7.518.217.407.287.657.657.657.677.617.51
MEDIAN7877878887
MODE7966878777
STDEV1.1971.1911.3341.3461.2461.1421.3291.0411.2061.255
VAR1.4331.4191.7811.8131.5531.3031.7681.0831.4561.576
KURT−0.622−0.856−0.560−0.753−0.988−0.369−0.939−0.319−0.817−0.525
SKEW0.464−0.2270.7050.6920.1340.5140.3040.4230.2910.625
Table 4. Descriptive statistics of the success of students at Police Department (sample size is N = 93 ).
Table 4. Descriptive statistics of the success of students at Police Department (sample size is N = 93 ).
ItemsSubjects
S1S2S3S4S5S6S7S8S9S10
MIN6666666666
MAX10101010101010101010
MEAN7.608.717.647.717.937.767.847.938.007.62
MEDIAN7987878887
MODE7967878877
STDEV1.2321.1801.3341.1211.2321.1111.2241.0091.2431.284
VAR1.5181.3921.7801.2561.5181.2341.4981.0181.5451.649
KURT−0.965−0.520−1.189−0.659−0.856−0.748−0.782−0.608−1.035−0.795
SKEW0.369−0.6200.2160.405−0.0210.407−0.079−0.0010.0740.423
Table 5. AHP analysis.
Table 5. AHP analysis.
Subjects
Average Grade of
Department of SAFETYCriminology DepartmentPolice Department
S10.07810.05760.0356
S20.33780.31160.3048
S30.03030.03670.0349
S40.04020.02550.0461
S50.07960.10260.1087
S60.1880.10260.0616
S70.02220.10260.0832
S80.09040.10260.1087
S90.07130.10030.1785
S100.06170.05760.0374
Table 6. Regression analysis model with standardized Beta-coefficients.
Table 6. Regression analysis model with standardized Beta-coefficients.
ModelDepartment of
Safety
Criminology DepartmentPolice
Department
(Constant)BetaBetaBeta
S10.1040.0610.084
S20.0240.1910.01
S30.1330.1190.205
S40.1440.0340.21
S50.1290.0230.111
S60.0250.0850.109
S70.0090.0660.07
S80.1260.1540.077
S90.0790.1330.087
S100.2280.1790.037
R20.8400.8950.836
Legend: Abbreviations in Table 6 correspond to the ones in Table 1.
Table 7. Correlogram of the grades of students between different subjects at the Department of Safety.
Table 7. Correlogram of the grades of students between different subjects at the Department of Safety.
SubjectsS1S2S3S4S5S6S7S8S9S10
S11.0000.2170.3650.4590.244−0.0150.0610.3850.3190.367
S20.2171.0000.3140.3930.211−0.0410.3480.3520.2910.343
S30.3650.3141.0000.5630.6040.3270.2170.5760.6900.054
S40.4590.3930.5631.0000.5670.2900.2690.6520.6710.287
S50.2440.2110.6040.5671.0000.3270.1240.4430.5790.011
S6−0.015−0.0410.3270.2900.3271.0000.2340.3110.265−0.292
S70.0610.3480.2170.2690.1240.2341.0000.2980.1180.247
S80.3850.3520.5760.6520.4430.3110.2981.0000.5860.321
S90.3190.2910.6900.6710.5790.2650.1180.5861.0000.246
S100.3670.3430.0540.2870.011−0.2920.2470.3210.2461.000
Table 8. Correlogram of the grades of students between different subjects at the Criminology Department.
Table 8. Correlogram of the grades of students between different subjects at the Criminology Department.
SubjectsS1S2S3S4S5S6S7S8S9S10
S11.0000.3120.4280.4970.4570.3940.3270.5110.4230.300
S20.3121.0000.1930.3410.3750.2920.3860.2160.2810.118
S30.4280.1931.0000.5820.4950.2590.4340.5360.3650.419
S40.4970.3410.5821.0000.4640.3440.5450.4880.4090.537
S50.4570.3750.4950.4641.0000.2630.4420.5970.3470.219
S60.3940.2920.2590.3440.2631.0000.1650.2760.2240.177
S70.3270.3860.4340.5450.4420.1651.0000.3530.4710.205
S80.5110.2160.5360.4880.5970.2760.3531.0000.4790.351
S90.4230.2810.3650.4090.3470.2240.4710.4791.0000.250
S100.3000.1180.4190.5370.2190.1770.2050.3510.2501.000
Table 9. Correlogram of the grades of students between different subjects at the Police Department.
Table 9. Correlogram of the grades of students between different subjects at the Police Department.
SubjectsS1S2S3S4S5S6S7S8S9S10
S11.000−0.1750.1740.1940.4760.0600.1240.1610.2820.319
S2−0.1751.0000.3660.0730.0020.1530.0780.1170.062−0.134
S30.1740.3661.0000.5990.4280.1390.2850.1680.4250.331
S40.1940.0730.5991.0000.3640.2890.2650.2240.2450.333
S50.4760.0020.4280.3641.0000.3530.2490.3620.2670.515
S60.0600.1530.1390.2890.3531.0000.2390.0660.000−0.050
S70.1240.0780.2850.2650.2490.2391.0000.1200.3440.265
S80.1610.1170.1680.2240.3620.0660.1201.0000.3440.296
S90.2820.0620.4250.2450.2670.0000.3440.3441.0000.242
S100.319−0.1340.3310.3330.515−0.0500.2650.2960.2421.000
Table 10. Estimated values of the coefficient of determination ( R i 2 ) and the variation inflated factor (VIF) for all modules, i.e., subjects (S1. S2. … S10).
Table 10. Estimated values of the coefficient of determination ( R i 2 ) and the variation inflated factor (VIF) for all modules, i.e., subjects (S1. S2. … S10).
DepartmentItemsS1S2S3S4S5S6S7S8S9S10
Department of Safety R i 2 0.3510.3000.6030.6550.6670.3720.2930.5510.7400.483
VIF1.5411.4292.5192.8993.0031.5921.4142.2273.8461.934
Criminology Department R i 2 0.4090.2800.4900.5950.7460.2170.4430.4930.6880.359
VIF1.6921.3891.9612.4693.9371.2771.7951.9723.2051.560
Police Department R i 2 0.2990.3090.6130.4700.5720.3300.2280.2760.3610.418
VIF1.4271.4472.5841.8872.3361.4931.2951.3811.5651.718
Table 11. Evaluation aggregation.
Table 11. Evaluation aggregation.
Firs Year ModulesDepartment for SafetyCriminology DepartmentPolice Department
Applied MethodRegr.AHPAggr.Regr.AHPAggr.Regr.AHPAggr.
S10.1050.07810.091550.0590.05760.05830.0840.03560.0598
S20.0240.33780.18090.1830.31160.24730.0100.30480.1574
S30.1340.03030.082150.1130.03670.074850.2040.03490.11945
S40.1430.04020.09160.0340.02550.029750.2100.04610.12805
S50.1290.07960.10430.0210.10260.06180.1100.10870.10935
S60.0260.1880.1070.0820.10260.09230.1110.06160.0885
S70.0080.02220.01510.0630.10260.08280.0710.08320.0771
S80.1250.09040.10770.1470.10260.12480.0770.10870.09285
S90.0780.07130.074650.1270.10030.113650.0870.17850.13275
S100.2280.06170.144850.1710.05760.11430.0370.03740.0372
Table 12. Ranked attributes by ReliefF on the Department for Safety dataset.
Table 12. Ranked attributes by ReliefF on the Department for Safety dataset.
Ranked AttributesAttributes
0.1862S8
0.1279S5
0.1268S1
0.1198S4
0.1163S9
0.0955S10
0.0766S2
0.0625S6
0.0466S3
0.0455S7
Table 13. Accuracy of the Department for Safety dataset with J48 tree.
Table 13. Accuracy of the Department for Safety dataset with J48 tree.
Weighted Avg.
No. AttributesPrecisionRecallF-MeasureAUC
100.8540.8540.8540.860
90.8540.8540.8540.795
80.8540.8540.8540.795
70.8540.8540.8540.765
60.8540.8540.8540.765
50.8540.8540.8540.755
40.8540.8540.8540.755
30.8830.8780.8800.853
20.7410.7320.7360.765
10.8670.8540.8580.718
Table 14. Accuracy of the Department for Safety dataset with ranked attributes by Aggregation.
Table 14. Accuracy of the Department for Safety dataset with ranked attributes by Aggregation.
Weighted Avg.
No. AttributePrecisionRecallF-MeasureAUC
100.8540.8540.8540.860
90.8300.9110.8690.195
80.8540.8540.8540.795
70.8540.8540.8540.795
60.7940.8050.7970.790
50.7270.7560.7330.669
40.8760.8780.8700.794
30.8190.8290.8180.735
20.5720.7560.6510.485
10.5720.7560.6510.485
Table 15. Number of attributes with ranking coefficient for all Departments.
Table 15. Number of attributes with ranking coefficient for all Departments.
No. Attr.Department of SafetyCriminology DepartmentPolice Department
ReliefFAggreg.ReliefFAggreg.ReliefFAggreg.
1S8S2S5S2S1S2
2S5S10S2S8S9S9
3S1S8S3S10S5S4
4S4S6S9S9S3S3
5 S4S6S7S5
6 S8S7S8S8
7 S7S3S10S6
8 S10S5S2S7
9 S6S1
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ranđelović, M.; Aleksić, A.; Radovanović, R.; Stojanović, V.; Čabarkapa, M.; Ranđelović, D. One Aggregated Approach in Multidisciplinary Based Modeling to Predict Further Students’ Education. Mathematics 2022, 10, 2381. https://doi.org/10.3390/math10142381

AMA Style

Ranđelović M, Aleksić A, Radovanović R, Stojanović V, Čabarkapa M, Ranđelović D. One Aggregated Approach in Multidisciplinary Based Modeling to Predict Further Students’ Education. Mathematics. 2022; 10(14):2381. https://doi.org/10.3390/math10142381

Chicago/Turabian Style

Ranđelović, Milan, Aleksandar Aleksić, Radovan Radovanović, Vladica Stojanović, Milan Čabarkapa, and Dragan Ranđelović. 2022. "One Aggregated Approach in Multidisciplinary Based Modeling to Predict Further Students’ Education" Mathematics 10, no. 14: 2381. https://doi.org/10.3390/math10142381

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop