Next Article in Journal
Extending Ramsey Numbers for Connected Graphs of Size 3
Previous Article in Journal
Hierarchical Hypervapotron Structure Integrated with Microchannels for Advancement of Thermohydraulic Performance
Previous Article in Special Issue
New Fixed Point Theorems for Generalized Meir–Keeler Type Nonlinear Mappings with Applications to Fixed Point Theory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Double Inertial Forward–Backward Splitting Algorithm for Variational Inclusion Problems: Applications in Mathematical Integrated Skill Prediction

by
Nipa Jun-On
1 and
Watcharaporn Cholamjiak
2,*
1
Department of Mathematics, Faculty of Science, Lampang Rajabhat University, Lampang 52100, Thailand
2
School of Science, University of Phayao, Phayao 56000, Thailand
*
Author to whom correspondence should be addressed.
Symmetry 2024, 16(8), 1091; https://doi.org/10.3390/sym16081091
Submission received: 17 July 2024 / Revised: 16 August 2024 / Accepted: 21 August 2024 / Published: 22 August 2024

Abstract

:
This paper introduces a new algorithm that combines the forward–backward splitting algorithms with a double inertial technique, utilizing the previous three iterations. The weak convergence theorem is established under certain mild conditions in a Hilbert space, including a relaxed inertial method in real numbers. An example of infinite dimension space is given with numerical results to support our proposed algorithm. The algorithm is applied to an asymmetrical educational dataset of students from 109 schools, utilizing asymmetric inputs as nine attributes to predict the output as students’ mathematical integrated skills. The algorithm’s performance is compared with other algorithms in the literature to demonstrate its effectiveness. The proposed algorithm demonstrates comparable precision, recall, accuracy, and F1 score but performs a relatively lower number of iterations. The contributions of each performance aspect to the mathematical integration skill of students are discussed to improve students’ mathematical learning.

1. Introduction

Over the last decade, educational data mining has been among the approaches used to analyze student performance and make related predictions of education outcomes [1]. These methods, specifically classification algorithms, enable educators to identify trends from a large dataset, supporting evidence-based decision-making processes in education [2,3]. However, with all this development in the area, the prediction of skills like a mathematical integrated skill still needs to be discovered due to the complex nature of such a skill.
Mathematical integrated skills refer to the ability of students to integrate and use mathematical knowledge from several content strands, including algebra, geometry, and data analysis, in solving complex problems [4]. These are essential skills for students in their general mathematical ability and success in high-order mathematics. Traditional assessment and prediction methods often need to include the completeness of these integrated abilities in students and, hence, are varied in our understanding and support for student learning in this area.
For that matter, this study proposes a double inertial forward–backward splitting algorithm that is improved to better predict students’ mathematical integrated skills. This algorithm is applied to an educational dataset of students’ performance across nine key mathematical content areas and predicts overall integrated skill level. This paper uses variational inclusion techniques in machine learning to obtain an effective prediction model supporting targeted educational interventions.
It would be an essential study, as it would develop the accuracy of educational predictions, allowing more effective teaching strategies to be applied at the individual level. The proposed algorithm may perform the same in terms of accuracy, precision, and recall, but it achieves those results with fewer iterations than existing methods. Hence, it has proven to be a promising tool for educational data analysis.
To investigate the predictive capacity of the variational inclusion problem (VIP), a specific area within the field of optimization, in determining students’ mathematical integrated skill, we let H be a real Hilbert space, the VIP is to find an element x * H such that
0 F x * + G x * ,
where an operator G : H 2 H is monotone and F : H H is a monotone Lipschitz continuous operator. Finding a solution for the VIP (1) encompasses several practical applications, including signal recovery, image processing, machine learning, and other related techniques [5,6,7].
One technique of solving the VIP (1) is considering designing algorithms to solve the fixed point problem of the mapping J τ G ( I τ F ) that is 0 F x * + G x * if and only if x * = J τ G ( I τ F ) x * where J τ G : H H is resolvent mapping associated with the set-value mapping G is identified by
J τ G ( x ) : = ( I + τ G ) 1 ( x ) , x H ,
for some τ > 0 , where I denotes the identity operator acting on H . A multitude of algorithms have been devised using these concepts, one of which is the well-known forward–backward splitting method [7,8], also referred to as the Picard algorithm with mapping J τ k G ( I τ k F ) ). This algorithm is characterized by the following method: x 1 H and
x k + 1 = J τ k G ( x k τ k F x k ) , k 1 ,
where τ k > 0 is the stepsize. Additionally, one of the renowned algorithms is generally known as Mann’s algorithm, which was introduced by Mann [9]. Mann’s algorithm with J τ k G ( I τ k F ) mapping is in the form of x 1 H and
x k + 1 = ( 1 α k ) x k + α k J τ k G ( I τ k F ) x k , k N ,
where α k is a nonnegative sequence in [ 0 , 1 ] . Mann’s algorithm with J τ k G ( I τ k F ) mapping has been seen in several applications, as detailed in [10,11].
The inertial forward–backward algorithm (IFB) is another well-known variational inclusion algorithm that was first proposed by Moudafi and Oliny [12]. This algorithm underwent improvement based on the inertial technique, which was proposed by [13] in 1964 in order to expedite the convergence of the algorithm. The algorithm was generated by x 0 , x 1 H and
y k = x k + θ k ( x k x k 1 ) , x k + 1 = J τ k G ( y k τ k F x k ) , k 0 ,
where { τ k } is a positive real sequence. Also, the inertial Mann forward–backward splitting algorithm (IMFB), which was proposed by Peeyada et al. [14] in a recent study, was developed by combining the inertial technique and the forward–backward algorithm. The aforementioned algorithm was applied in the field of machine learning to forecast breast cancer with favorable performance. To generate algorithm, let x 0 , x 1 H , and
y k = x k + θ k ( x k x k 1 ) , z k = α k x k + ( 1 α k ) y k , x k + 1 = J τ k G ( z k τ k F z k ) , k 0 ,
where { τ k } is a positive real sequence and { α k } [ 0 , 1 ] .
Very recently, Wang et al. [15] introduced a modified Tseng splitting method with double inertial extrapolation steps (MTSMDISs) and self-adaptive step sizes for solving monotone inclusion problems in real Hilbert spaces. The algorithm was defined by choosing x 0 , x 1 H , μ ( 0 , 1 ) and τ 1 > 0 and
w k = x k + θ k ( x k x k 1 ) , z k = x k + δ k ( x k x k 1 ) , y k = J τ k G ( I τ k F ) w k ,
where
τ k + 1 = min ( μ k + μ ) w k y k F w k F y k , τ k + p k , if F w k F y k , τ k + p k , otherwise .
If w k = y k , then stop. Otherwise,
x k + 1 = ( 1 α k ) z k + α k ( y k τ k ( F y k F w k ) ) ,
where { θ k } , { β k } are extrapolation parameters, and { μ k } , { p k } , { α k } are positive real sequences.
Inspired by the above works, we introduce a new algorithm by modifying Picard and Mann algorithms with a double inertial technique, which computes using the previous three iterative. We apply the algorithm for machine learning to a dataset of students from 109 schools using nine attributes in students’ mathematical integrated skill prediction. To show the performance of our algorithm, we compare it with the other algorithms in the literature.

2. Mathematical Preliminaries

In this section, let H be a real Hilbert space. The symbols ⇀ and → are used to denote weak and strong convergence, respectively. We proceed to gather a set of necessary definitions and lemmas that will be instrumental in establishing the validity of our main results. In the subsequent section, we will provide fundamental definitions and lemmas necessary for establishing our main theorem.
Definition 1. 
Let C be a nonempty, closed, and convex subset of a Hilbert space H . The nearest point projection of H onto C is denoted by P C , that is, a P C a a b for all a H and b C . Such P C is called the metric projection of H onto C .
Definition 2. 
A mapping F : H H is said to be
(i
Firmly nonexpansive if
F x F y 2 x y 2 ( I F ) x ( I F ) y 2 ,
or equivalently
x y , F x F y F x F y 2 ,
for all x , y H ;
(ii
α-cocoercive (also known as α-inverse strongly monotone) if α F is firmly nonexpansive, when α > 0 ;
(iii
L-Lipschitz continuous if there exists L > 0 such that
F x F y L x y ,
for all x , y H ;
(iv
Nonexpansive if F is L-Lipschitz continuous when L = 1 .
Definition 3. 
Let G : H 2 H be a multivalued mapping. Then, G is said to be
(i
Monotone if u v , x y 0 , for all ( x , u ) , ( y , v ) g r a ( G ) (the graph of mapping G );
(ii
Maximal monotone if there does not exist a proper monotone extension of g r a ( G ) .
Based on the given definition, it can be inferred that every mapping classified as α -cocoercive possesses both monotonicity and 1 α -Lipschitz continuity. It is widely recognized that if G : H 2 H is a multi-valued maximal monotone mapping and τ > 0 , then J τ G : = ( I + τ G ) 1 is a single-valued firmly nonexpansive mapping [16].
Lemma 1 
([17]). Let F : H H be a nonexpansive mapping such that F i x ( F ) . If there exists a sequence { x n } in H such that x k x H and x k F x k 0 , then x F i x ( F ) .
Lemma 2 
([8]). Let F : H H be a α-cocoercive mapping and G : H 2 H be a maximal monotone mapping. Then, we have
( i ) for τ > 0 , F i x J τ G ( I τ F ) = ( F + G ) 1 ( 0 ) ,
( i i ) for 0 < τ τ ¯ and x H , x J τ G ( I τ F ) x 2 x J τ ¯ G ( I τ ¯ F ) x .
Lemma 3 
([18], Opial). Let S be a nonempty set of H and { x k } be a sequence in H . Assume that the following conditions hold.
( i ) For every x S , x k x converges.
( i i ) Every weak sequential cluster point of { x k } belongs to S .
Then, { x k } weakly converges to an element in S .
Lemma 4 
([19]). Let { a k } and { b k } be nonnegative sequences of real numbers satisfying k = 1 b k < and a k + 1 a k + b k . Then, { a k } is a convergent sequence.

3. Analysis of Convergence

In this section, let C be a nonempty closed convex subset of a real Hilbert space H . Let F : H H be an α -inverse strongly monotone operator and G : H 2 H be a maximal monotone operator such that ( F + G ) 1 ( 0 ) C .
Remark 1. 
(i) If θ k = 0 or δ k = 0 , Algorithm 1 reduces to following two different one-inertial Picard–Mann projective forward–backward splitting algorithms:
  • (OIPMPFBS-I): δ k = 0 ,
    y k = x k + θ k ( x k x k 1 ) z k = ( 1 α k ) x k + α k J τ k G ( I τ k F ) y k x k + 1 = P C ( J τ k G ( I τ k F ) z k ) ;
  • (OIPMPFBS-II): θ k = 0 ,
    y k = x k + δ k ( x k 1 x k 2 ) z k = ( 1 α k ) x k + α k J τ k G ( I τ k F ) y k x k + 1 = P C ( J τ k G ( I τ k F ) z k ) .
  • (ii) If θ k = 0 and δ k = 0 , Algorithm 1 reduces to Picard–Mann forward–backward splitting algorithms (PMPFBS):
    z k = ( 1 α k ) x k + α k J τ k G ( I τ k F ) x k x k + 1 = P C ( J τ k G ( I τ k F ) z k ) .
  • (iii) The difference between an inertial method and a standard method with relaxing extrapolation parameters θ k and δ k is an inertial method is more relaxing than a standard method, which can be set to obtain faster convergence. The following structure shows the first step of the comparison between an inertial method and a standard method in R 2 :
From Algorithm 1, we see that y k is generated by x k + θ k ( x k x k 1 ) + δ k ( x k 1 x k 2 ) , which is relaxed by the condition of extrapolation parameters θ k and δ k , as follows:
(1) If δ k = 0 (OIPMFBS-I), θ k is impacted to generate y k ; from Figure 1, it can be observed that θ 1 ( x 1 x 0 ) can be a vector in red and blue lines when θ 1 represents positive and negative numbers, respectively, and generate y 1 in purple;
(2) If θ k = 0 (OIPMFBS-II), δ k is impacted to generate y k ; from Figure 1, it can be observed that δ 1 ( x 0 x 1 ) can be a vector in yellow and green lines when δ 1 is positive and negative numbers, respectively, and generate y 1 in dark blue;
(3) If θ k 0 and δ k 0 , y 1 , in black, is generated by θ 1 ( x 1 x 0 ) + δ 1 ( x 0 x 1 ) (light green) from L 1 , L 2 lines and x 1 ;
(4) If θ k = 0 and δ k = 0 , y 1 = x 1 ;
(5) Vector y 1 (dark blue) can be updated in the closed convex constrained set C by the projection metric P C (green square or orange circle areas).
Algorithm 1. DRIPFBS: Double Relaxed Inertial Projective Forward–Backward Splitting Algorithm
  • Initialization: Select x 1 , x 0 , x 1 H , { α k } ( a , b ) ( 0 , 1 ] , { τ k } ( c , d ) ( 0 , 2 α ) , { θ k } , { δ k } ( , ) satisfies the condition such that
    k = 1 | θ k | x k x k 1 < a n d k = 1 | δ k | x k 1 x k 2 < .
  • Iterative Steps: Construct { x k } by using the following steps:
  • Step 1. Define
    y k = x k + θ k ( x k x k 1 ) + δ k ( x k 1 x k 2 ) ,
  • Step 2. Compute
    z k = ( 1 α k ) x k + α k J τ k G ( I τ k F ) y k .
  • Step 3. Compute
    x k + 1 = P C ( J τ k G ( I τ k F ) z k ) .
  • Replace n with n + 1 and later repeat Step 1.
Theorem 2. 
The sequence { x k } generated by the DRIFBS algorithm converges weakly to an element in ( F + G ) 1 ( 0 ) .
Proof. 
Let x * ( F + G ) 1 ( 0 ) and J k = J τ k G ( I τ k F ) . Since { τ k } ( 0 , 2 α ) , J k is nonexpansive mapping by [20], and we deduce the following:
x k + 1 x * = P C ( J k ( 1 α k ) x k + α k J k y k ) x * ( 1 α k ) x k x * + α k J k y k x * ( 1 α k ) x k x * + α k y k x * x k x * + α k | θ k | x k x k 1 + α k | δ k | x k 1 x k 2 .
By the conditions of { θ k } , { δ k } , it follows from Lemma 4 that we attain that lim k x k x * exists. Since J τ k G is firmly nonexpansive, we have
x k + 1 x * 2 = P C ( J k ( 1 α k ) x k + α k J k y k ) x * 2 ( 1 α k ) ( x k x * ) + α k ( J k y k x * ) 2 ( 1 α k ) x k x * 2 + α k J k y k x * 2 ( 1 α k ) x k x * 2 + α k y k x * τ k ( F y k F x * ) 2 y k J k y k τ k ( F y k F x * ) 2 = ( 1 α k ) x k x * 2 + α k ( y k x * 2 ( τ k ) 2 F y k F x * 2 2 τ k y k x * , F y k F x * y k J k y k τ k ( F y k F x * ) 2 ) x k x * 2 + 2 α k θ k ( x k x k 1 ) + δ k ( x k 1 x k 2 ) , y k x * α k τ k ( 2 α τ k ) F y k F x * 2 α k y k J k y k τ k ( F y k F x * ) 2 .
Repeatedly, by the conditions of the sequences { θ k } , { δ k } , and (12), we have
lim k F y k F x * = lim n y k J k y k τ k ( F y k F x * ) = 0 .
This then implies that
lim k y k J k y k = 0 .
Since lim inf k τ k > 0 , there exists τ > 0 such that 0 < τ τ k . By Lemma 2(ii), we have
lim k y k J τ G ( I τ F ) y k lim k y k J k y k = 0 .
By the definition of { y k } , we obtain
lim k y k x k lim k | θ k | x k x k 1 + | δ k | x k 1 x k 2 = 0 .
Since { x k } is bounded, there exists a subsequence { x k n } of { x k } that weakly converges to x * H . Since { x k } is generated into C and C is closed, x * C . According to (15), { y k } also converges weakly to x * . Moreover, by Lemma 2(i) and Lemma 1, it follows from (14), we obtain x * F i x J τ G ( I τ F ) = ( F + G ) 1 ( 0 ) . Finally, by Opial’s lemma (Lemma 3), we achieve that { x k } converges weakly to an element in ( F + G ) 1 ( 0 ) . □
Next, we give an example in the infinite dimension space L 2 [ 0 , 1 ] = { x ( s ) : 0 1 x ( s ) d s < } , where . is L 2 -norm, defined by x = 0 1 | x ( s ) | 2 d s , supporting our main theorem (Algorithm 1).
Example 1. 
Let H = L 2 [ 0 , 1 ] , F x ( s ) = 4 x ( s ) and G x ( s ) = 2 x ( s ) , where x ( s ) L 2 [ 0 , 1 ] .
In our experiments, we consider two cases of closed convex sets C :
(i) C = H ;
(ii) C = { x ( s ) : x ( s ) s 3 1 } ; then
P C ( x ( s ) ) = s 3 + 4 x ( s ) s 3 x ( s ) s 3 , if x ( s ) C x ( s ) , otherwise .
We use Cauchy errors x k x k + 1 2 < 10 6 to stop the iteration. In order to optimize the performance of our algorithm, we will examine the necessary parameters of Algorithm 1 in four distinct cases when
θ k = 1 x k x k 1 k 3 , if x k x k 1 and k > M θ , otherwise
and
δ k = 1 x k 1 x k 2 k 3 , if x k 1 x k 2 and k > M δ , otherwise
where M is the iteration number that we want to stop, and we choose the initialization x 1 ( s ) = cos s , x 0 ( s ) = 1 3 e s + 3 s , and x 1 ( s ) = 2 s 2 3 s + 5 .
Case1. We set θ = 0 , δ = 0 , and α k = k 2 k + 1 for (i) and (ii). The different parameters τ k are considered in Table 1.
Case2. We set θ = 0 , δ = 0 , and τ k = 1 5 for (i) and (ii). The different parameters α k are considered in Table 2.
Case3. We set δ = 0 , τ k = 1 5 , and α k = k k + 1 for (i), and δ = 0 , τ k = 1 5 , and α k = k 2 k + 1 for (ii). The different parameters θ are considered in Table 3.
Case4. We set θ = 0.5 , τ k = 1 5 , and α k = k k + 1 for case (i) and θ = 0.5 , τ k = 1 5 , and α k = k 2 k + 1 for case (ii). The different parameters δ are considered in Table 4.
From Table 1, Table 2, Table 3 and Table 4 and Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9, we see that the parameters α k = k k + 1 , τ k = 0.2 , θ = 0.5 , and δ = 10 of case (i) and α k = k 2 k + 1 , τ k = 0.2 , θ = 0.5 , and δ = 0 of case (ii) provide the most favorable results for the initialization x 1 ( s ) = cos s , x 0 ( s ) = 1 3 e s + 3 s , and x 1 ( s ) = 2 s 2 3 s + 5 .

4. Application to Educational Data Classification

To demonstrate the performance of our algorithm when applied to a data classification problem, we classify the empirical data obtained from the Thailand Ordinary National Educational Test (O-NET) regarding students’ mathematical integrated skill scores. We also collected the data from the assessment of nine learning standards following mathematics content strands designated by the Thailand primary core curriculum [4].
The learning standards were designated, serving as the development objectives for each strand. These standards specify the knowledge and skills that learners should possess. In addition, the learning standards serve as crucial mechanisms for advancing mathematics education, as they inform us of the contents and methods of teaching and evaluation. It is essential because it reveals the degree of success in attaining the quality specified by the applicable standards [4]. Therefore, by designated learning standards, we employ students’ performances in number representation, number operation, measurement, geometric figures, spatial reasoning, pattern and function, algebraic understanding, data analysis, and valid estimation as nine attributes in this classification problem.
We proceed to apply Algorithm 1 for machine learning to solve data classification problems using an extreme learning machine (ELM), which was introduced by Huang et al. [21] for classifying students’ mathematical integrated skills, categorizing them into five levels, defined as the following.
The training dataset is defined by U : = { ( x k , t k ) : x k R n , t k R m , k = 1 , 2 , , N } , N is a number of distinct samples, x k is input training data, and t k is the target of x k . We focus on using single-hidden layer feed-forward neural networks (SLFNs) such that the output function is computed by
O k = j = 1 M o j A ( w j x k + b j ) ,
where M represents hidden nodes, A is the activation function, o j is the optimal output weight at the j-th hidden node, w j is the parameter weight at the j-th hidden node, and b j is the bias. The hidden layer output matrix P is generated as follows:
P = A ( w 1 x 1 + b 1 ) A ( w M x 1 + b M ) A ( w 1 x N + b 1 ) A ( w M x N + b M ) .
Finding optimal output weight o = [ o 1 T , , o M T ] T such that P o = T is the goal in the ELM method, where T = [ t 1 T , , t N T ] T represents the training target data. We cannot write the solution o in the form o = P T , where P is the Moore–Penrose generalized inverse of P , not existing. In this particular example, the least squares problem was taken into consideration. The subsequent regularization of the least squares problem is considered for obtaining a good fit model:
min o R M { 1 2 P o T 2 2 + λ o 1 } ,
where λ is a regularization parameter. This problem is well recognized as the least absolute shrinkage and selection operator (LASSO) [22].
The performance of the algorithm was evaluated using classification efficient matrices, including accuracy, precision, recall, and F1 score, as shown in (20)–(23).
Precision ( % ) = tp tp + fp × 100 ,
Recall ( % ) = tp tp + fn × 100 ,
F 1 score ( % ) = 2 × Precision × Recall Precision + Recall × 100 ,
Accuracy ( % ) = tp + tn tp + fp + tn + fn × 100 ,
where these matrices gave True Positive ( tp ), True Negative ( tn ), False Positive ( fp ), and False Negative ( fn ).
The multi-class cross-entropy loss function is used in multi-class classification. The following average was computed:
Loss = j = 1 N o j log o ^ j ,
where o ^ j is the j-th scalar value in the model output, o j is the corresponding target value, and N is the number of scalar values in the model output.
These data contain a 109-instance educational dataset containing nine attributes of students’ performances in number representation, number operation, measurement, geometric figures, spatial reasoning, pattern and function, algebraic understanding, data analysis, and valid estimation. Before starting data training, a description of each attribute and an overview of the data are shown in Table 5 and Table 6, respectively.
We next solve the ELM model (19) by setting F ( o ) = ( 1 2 P o T 2 2 ) and G ( o ) = ( λ o 1 ) with λ = 0.5 and C = R M . The necessary parameters that are used in Algorithm 1, IFB (5), and IMFB (6) algorithms can be seen in Table 6, where
θ k = θ k 2 x k x k 1 i f k > K a n d x k x k 1 , θ o t h e r w i s e ;
δ k = δ ¯ k k 2 x k 1 x k 2 i f k > K a n d x k 1 x k 2 , δ ¯ k o t h e r w i s e ;
α k = α ¯ k i f k K , 0.5 o t h e r w i s e ,
where K is the number of iterations that we aim to stop.
In this experiment, we let A be a sigmoid function and let M = 100 . The following Table 7 shows the numerical training results when the model stops at the highest stable accuracy.
Remark 2. 
According to the findings presented in Table 8, it can be observed that
(i) IFB (5) and IMFB (6) required 747 iterations, both taking slightly longer training times (0.0208 and 0.0206, respectively) but achieving the same accuracy, precision, recall, and F1 score as Algorithm 1;
(ii) MTSMDIS (7) required 514 iterations with a training time of 0.0214, also yielding the same performance metrics;
(iii) OIPMPFBS-I and DRIPFBS were significantly more efficient, requiring only 51 and 46 iterations, respectively, with training times of 0.0045 and 0.0057, while still achieving the same performance metrics;
(iv) OIPMPFBS-II and PMPFBS required 550 and 371 iterations, respectively, with training times of 0.0173 and 0.0139, and achieved the same accuracy, precision, recall, and F1 score.
Algorithm 1 (iii) showcases the highest efficiency in precision, recall, accuracy, and F1 score, while requiring the lowest number of iterations. This ensures not only faster convergence but also reduces computational costs, making it a more efficient and time-saving option for accurately classifying tumors compared to other examination methods.
Next, we proceed to exhibit the accuracy and loss graphs for both the training and testing datasets to assess the models’ good fit adequacy.
Based on the data presented in Figure 10, it is evident that Algorithm 1 (DRIPFBS), for θ = 0.999 and δ ¯ k = 1 k + 1 , provides good accuracy, and the loss plots during training and validation consistently exhibit close proximity, particularly after 150 iterations, surpassing Algorithm 1; OIPMPFBS-I, OIPMPFBS-II, and PMPFBS are presented in Figure 11, Figure 12 and Figure 13. Additionally, it also outperforms the IFB and IMFB algorithms, as shown in Figure 14 and Figure 15.

5. Discussion on Mathematical Integrated Skills Prediction Results

According to the mathematical result, every aspect of performance in number representation, number operation, measurement, geometric figures, spatial reasoning, pattern and function, algebraic understanding, data analysis, and valid estimation contributes to the development of students’ mathematical integrated skills. The results from educational data can be discussed as follows.
Firstly, students proficient in number representation and number operations possess the capacity to precisely manipulate numerical values, execute mathematical operations, and establish correlations among diverse mathematical principles [23,24]. This skillset empowers individuals to utilize their comprehension of numerical concepts in diverse mathematical content strands, including algebra, measurement, and data analysis [25]. Moreover, students can execute mathematical procedures with precision and expediency, employ suitable methodologies and computational techniques, and logically analyze numerical interconnections. This proficiency enables individuals to resolve problems and establish correlations among diverse mathematical concepts [24,25].
Secondly, mastering measurement skills can also augment the spatial reasoning capabilities of students, as it necessitates comprehension and visualization of geometric concepts such as magnitude, configuration, and orientation [26]. Moreover, students who demonstrate proficiency in measurement possess the ability to establish correlations between geometric figures and measurements, employ formulas and techniques to compute areas and volumes, and scrutinize patterns and associations in spatial data [25]. Moreover, students who exhibit proficiency in measurement can utilize their comprehension of measurement units and conversions to resolve problems that entail proportions, scaling, and dimensional analysis, addressing mathematical challenges in algebra [26]. Also, they can scrutinize and construe data obtained via measurement, reinforcing their comprehension of statistical principles.
Additionally, acquiring expertise in geometric figures and spatial reasoning can augment students’ problem-solving aptitude and logical reasoning capabilities [26]. The utilization of geometric principles allows individuals to engage in the analysis and resolution of geometric problems, the construction and substantiation of geometric arguments, and the application of geometric principles to practical, real-world scenarios [25]. Competence in geometric figures and spatial reasoning also facilitates students’ comprehension of measurement, given that geometric shapes frequently entail the quantification of lengths, areas, volumes, and angles [25].
Moreover, proficiency in pattern recognition, function analysis, and algebraic understanding is crucial to the overall mathematical proficiency of students [25]. Students who demonstrate proficiency in these domains possess the ability to recognize and articulate mathematical relationships, represent them symbolically and algebraically, and utilize them to resolve problems spanning diverse mathematical fields [27]. Through acquiring expertise in these domains, students can augment their capacity to engage in critical thinking, establish correlations between diverse mathematical concepts, and employ mathematical knowledge in various contexts [25].
Finally, the ability to perform data analysis and make valid estimations empowers students to effectively manipulate empirical data, exercise critical thinking, and utilize mathematical principles in pragmatic situations. Students with a robust grounding in data analysis can interpret and explore data depictions, recognize recurring themes and tendencies, and formulate significant deductions. Individuals can utilize data to substantiate their mathematical assertions and authenticate their rationale [28]. Furthermore, proficient estimation skills enable students to employ techniques across diverse mathematical fields, including measurement, quantity evaluation, and problem-solving. Estimation can be a valuable problem-solving tool when precise values are impractical.
By developing competence in each of these areas, students improve their mathematical integrated skills, which entails the ability to connect and apply knowledge from different mathematical domains, recognize relationships between concepts, and solve complex problems using a variety of mathematical techniques and strategies.
In conclusion, the implementation of educational data classification can yield various advantages. The utilization of data analytics in education can assist in the identification of concealed patterns and trends in student performance, the anticipation of academic achievement or deficiency, the customization of learning experiences according to individual requirements, the assessment of the efficacy of teaching methodologies and interventions, and the provision of data-driven insights for educational policy and planning.

6. Conclusions

This paper is devoted to an improved variant of the double inertial forward–backward splitting algorithm for predicting students’ mathematical integrated skills based on performance variability across different mathematical content strands. The results show that this method maintains accuracy, precision, recall rate, and F1 score but requires fewer iterations.
Our numerical experiments have shown that this algorithm realizes the classification of students’ skills with an accuracy of 81.40, a precision and recall rate of 72.09, and an F1 score of 72.09 using just 46 iterations. Results of such kind testify that the algorithm reliably predicts integrated mathematical skills necessary for the all-rounded mathematical ability of students. The results imply that successful learners gain access to many complex mathematics abilities, of which those with a deep-rooted conception of numbers, geometric reasoning, and algebraic thinking are the most common courses.
Our study results in the advancement of algorithms that can be used for skill prediction and the definition of cognitive strategies for individuals. This work will serve as a basis for more research shortly, as it will provide for a revision of the methodology so that it can be tested in datasets with other types of education data, making the whole education area even more stimulating and precise.
Consequently, the proposed algorithm is effective. It may be an excellent support tool for teaching professionals and researchers involved in enhancing students’ mathematics learning outcomes. Applying this technique in teaching can bring about a leap in educational intervention, making it more effective and, thus, personalized, thereby helping more students succeed in mathematics at the school level.

Author Contributions

Conceptualization, N.J.-O. and W.C.; methodology, N.J.-O.; software, W.C.; validation, N.J.-O. and W.C.; formal analysis, W.C.; investigation, N.J.-O.; resources, N.J.-O.; data curation, N.J.-O.; writing—original draft preparation, N.J.-O.; writing—review and editing, W.C.; visualization, W.C.; project administration, W.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Research Council of Thailand and the University of Phayao (grant number N42A650334); Thailand Science Research and Innovation, University of Phayao (Fundamental Fund 2024); and Research Council of Thailand (NRCT): N42A670856. The APC was funded by Lampang Rajabhat University.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors upon request.

Acknowledgments

This research work was supported by the National Research Council of Thailand and the University of Phayao (N42A650334); Thailand Science Research and Innovation, University of Phayao (Fundamental Fund 2024); Research Council of Thailand (NRCT): N42A670856; and the revenue budget in 2023–2024 of Lampang Rajabhat University. N.J.-o would like to acknowledge Suthep Suantai as a mentor for initiation of this project.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Romero, C.; Ventura, S. Educational data mining: A review of the state of the art. IEEE Trans. Syst. Man Cybern. Syst. 2010, 40, 601–618. [Google Scholar] [CrossRef]
  2. Jalota, C.; Agrawal, R. Analysis of educational data mining using classification. In Proceedings of the 2019 International Conference on Machine Learning, Big Data, Cloud and Parallel Computing, Faridabad, India, 14 February 2019; pp. 243–247. [Google Scholar]
  3. Agarwal, S.; Pandey, G.N.; Tiwari, M.D. Data mining in education: Data classification and decision tree approach. Int. J. E-Educ. E-Bus. E-Manag. E-Learn 2012, 2, 140–144. [Google Scholar] [CrossRef]
  4. Ministry of Education. Basic Education Core Curriculum B.E. 2551 (A.D. 2008), 1st ed.; Agricultural Cooperative Federation of Thailand Press: Bangkok, Thailand, 2008.
  5. Bauschke, H.H.; Borwein, J.M. On projection algorithms for solving convex feasibility problems. SIAM Rev. 1996, 38, 367–426. [Google Scholar] [CrossRef]
  6. Chen, P.; Huang, J.; Zhang, X. A primal-dual fixed point algorithm for convex separable minimization with applications to image restoration. Inverse Probl. 2013, 29, 025011. [Google Scholar] [CrossRef]
  7. Combettes, P.L.; Wajs, V.R. Signal recovery by proximal forward-backward splitting. Multiscale Model. Simul. 2005, 4, 1168–1200. [Google Scholar] [CrossRef]
  8. López, G.; Martín-Márquez, V.; Wang, F.; Xu, H.K. Forward-backward splitting methods for accretive operators in Banach spaces. Abstr. Appl. Anal. 2012, 2012, 109236. [Google Scholar] [CrossRef]
  9. Mann, W.R. Mean value methods in iteration. Proc. Am. Math. Soc. 1953, 4, 506–510. [Google Scholar] [CrossRef]
  10. Peeyada, P.; Cholamjiak, W. A new projection algorithm for variational inclusion problems and its application to cervical cancer disease prediction. J. Comput. Appl. Math. 2024, 441, 115702. [Google Scholar] [CrossRef]
  11. Peeyada, P.; Dutta, H.; Shiangjen, K.; Cholamjiak, W. A modified forward-backward splitting methods for the sum of two monotone operators with applications to breast cancer prediction. Math. Methods Appl. Sci. 2023, 46, 1251–1265. [Google Scholar] [CrossRef]
  12. Moudafi, A.; Oliny, M. Convergence of a splitting inertial proximal method for monotone operators. J. Comput. Appl. Math. 2003, 155, 447–454. [Google Scholar] [CrossRef]
  13. Polyak, B.T. Some methods of speeding up the convergence of iteration methods. USSR Comput. Math. Math. Phys. 1964, 4, 1–17. [Google Scholar] [CrossRef]
  14. Peeyada, P.; Suparatulatorn, R.; Cholamjiak, W. An inertial Mann forward-backward splitting algorithm of variational inclusion problems and its applications. Chaos Solitons Fractals 2022, 158, 112048. [Google Scholar] [CrossRef]
  15. Wang, Z.B.; Lei, Z.Y.; Long, X.; Chen, Z.Y. A modified Tseng splitting method with double inertial steps for solving monotone inclusion problems. J. Sci. Comput. 2023, 96, 92. [Google Scholar] [CrossRef]
  16. Marino, G.; Xu, H.K. Convergence of generalized proximal point algorithms. Commun. Pure Appl. Anal. 2004, 3, 791. [Google Scholar] [CrossRef]
  17. Goebel, K.; Kirk, W.A. Topics in Metric Fixed Point Theory, 1st ed.; Cambridge University Press: New York, NY, USA, 1990. [Google Scholar]
  18. Bauschke, H.H.; Combettes, P.L. Convex Analysis and Monotone Operator Theory in Hilbert Spaces, 2nd ed.; Springer: New York, NY, USA, 2011. [Google Scholar]
  19. Auslender, A.; Teboulle, M.; Ben-Tiba, S. A logarithmic-quadratic proximal method for variational inequalities. In Computational Optimization: A Tribute to Olvi Mangasarian; Mangasarian, O.L., Pang, J., Eds.; Kluwer Academic Publishers: New York, NY, USA, 1999; pp. 31–40. [Google Scholar]
  20. Zhang, S.S.; Lee, J.H.; Chan, C.K. Algorithms of common solutions to quasi variational inclusion and fixed point problems. Appl. Math. Mech. 2008, 29, 571–581. [Google Scholar] [CrossRef]
  21. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks, Budapest, Hungary, 25–29 July 2004; p. 04CH37541. [Google Scholar]
  22. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  23. Sterner, G.; Wolff, U.; Helenius, O. Reasoning about representations: Effects of an early math intervention. Scand. J. Educ. Res. 2020, 64, 782–800. [Google Scholar] [CrossRef]
  24. Yuhasriati, Y.; Johar, R.; Khairunnisak, C.; Rohaizati, U.; Jupri, A.; Zubaidah, T. Students’ mathematical representation ability in learning algebraic expression using realistic mathematics education. J. Didaktik. Mat. 2022, 9, 151–169. [Google Scholar] [CrossRef]
  25. Mulligan, J.; Mitchelmore, M. Awareness of pattern and structure in early mathematical development. Math. Educ. Res. J. 2009, 21, 33–49. [Google Scholar] [CrossRef]
  26. Xie, F.; Zhang, L.; Chen, X.; Xin, Z. Is spatial ability related to mathematical ability: A meta-analysis. Educ. Psychol. Rev. 2020, 32, 113–155. [Google Scholar] [CrossRef]
  27. Zazkis, R.; Liljedahl, P. Generalization of patterns: The tension between algebraic thinking and algebraic notation. Educ. Stud. Math. 2002, 49, 379–402. [Google Scholar] [CrossRef]
  28. Garfield, J.; Ahlgren, A. Difficulties in learning basic concepts in probability and statistics: Implications for research. J. Res. Math. Educ. 1988, 19, 44–63. [Google Scholar] [CrossRef]
Figure 1. A structure comparison of relaxing conditions of double inertial in R 2 .
Figure 1. A structure comparison of relaxing conditions of double inertial in R 2 .
Symmetry 16 01091 g001
Figure 2. Cauchy error plots of the different parameter τ k used in Algorithm 1 for Example 1 of case (i).
Figure 2. Cauchy error plots of the different parameter τ k used in Algorithm 1 for Example 1 of case (i).
Symmetry 16 01091 g002
Figure 3. Cauchy error plots of the different parameter τ k used in Algorithm 1 for Example 1 of case (ii).
Figure 3. Cauchy error plots of the different parameter τ k used in Algorithm 1 for Example 1 of case (ii).
Symmetry 16 01091 g003
Figure 4. Cauchy error plots of the different parameter α k used in Algorithm 1 for Example 1 of (i).
Figure 4. Cauchy error plots of the different parameter α k used in Algorithm 1 for Example 1 of (i).
Symmetry 16 01091 g004
Figure 5. Cauchy error plots of the different parameter α k used in Algorithm 1 for Example 1 of (ii).
Figure 5. Cauchy error plots of the different parameter α k used in Algorithm 1 for Example 1 of (ii).
Symmetry 16 01091 g005
Figure 6. Cauchy error plots of the different parameter θ used in Algorithm 1 for Example 1 for (i).
Figure 6. Cauchy error plots of the different parameter θ used in Algorithm 1 for Example 1 for (i).
Symmetry 16 01091 g006
Figure 7. Cauchy error plots of the different parameter θ used in Algorithm 1 for Example 1 for (ii).
Figure 7. Cauchy error plots of the different parameter θ used in Algorithm 1 for Example 1 for (ii).
Symmetry 16 01091 g007
Figure 8. Cauchy error plots of the different parameter δ used in Algorithm 1 for Example 1 of (i).
Figure 8. Cauchy error plots of the different parameter δ used in Algorithm 1 for Example 1 of (i).
Symmetry 16 01091 g008
Figure 9. Cauchy error plots of the different parameter δ used in Algorithm 1 for Example 1 of (ii).
Figure 9. Cauchy error plots of the different parameter δ used in Algorithm 1 for Example 1 of (ii).
Symmetry 16 01091 g009
Figure 10. Accuracy and loss plots of Algorithm 1 (DRIPFBS).
Figure 10. Accuracy and loss plots of Algorithm 1 (DRIPFBS).
Symmetry 16 01091 g010
Figure 11. Accuracy and loss plots of Algorithm 1 (OIPMPFBS-I).
Figure 11. Accuracy and loss plots of Algorithm 1 (OIPMPFBS-I).
Symmetry 16 01091 g011
Figure 12. Accuracy and loss plots of Algorithm 1 (OIPMPFBS-II).
Figure 12. Accuracy and loss plots of Algorithm 1 (OIPMPFBS-II).
Symmetry 16 01091 g012
Figure 13. Accuracy and loss plots of Algorithm 1 (PMPFBS).
Figure 13. Accuracy and loss plots of Algorithm 1 (PMPFBS).
Symmetry 16 01091 g013
Figure 14. Accuracy and loss plots of IFB (5) algorithm.
Figure 14. Accuracy and loss plots of IFB (5) algorithm.
Symmetry 16 01091 g014
Figure 15. Accuracy and loss plots of IMFB (6) algorithm.
Figure 15. Accuracy and loss plots of IMFB (6) algorithm.
Symmetry 16 01091 g015
Table 1. Numerical results of different parameters τ k .
Table 1. Numerical results of different parameters τ k .
τ k 0.45 1 3 0.20.10.01
(i)CPU time0.53570.50880.50750.72002.7371
Number of iterations6551050
(ii)CPU time0.98870.92630.90501.28174.8820
Number of iterations8771150
Table 2. Numerical results of different parameters α k .
Table 2. Numerical results of different parameters α k .
α k k k + 1 k 2 k + 1 k 5 k + 1 k 10 k + 1 k 100 k + 1
(i)CPU time0.50090.50750.50380.55140.5496
Number of iterations55566
(ii)CPU time0.94980.92500.96181.03511.0495
Number of iterations77788
Table 3. Numerical results of different parameters θ .
Table 3. Numerical results of different parameters θ .
θ 0.10.5110100
(i)CPU time0.53130.50050.50260.50690.5385
Number of iterations55566
(ii)CPU time0.96260.82610.94911.22164.9595
Number of iterations7671050
Table 4. Numerical results of different parameters δ .
Table 4. Numerical results of different parameters δ .
δ −100−10−1110
(i)CPU time0.50690.43090.58070.51230.5853
Number of iterations54566
(ii)CPU time2.08461.82721.07561.11611.6258
Number of iterations19148914
Table 5. A description of each attribute.
Table 5. A description of each attribute.
Attribute: PerformanceDescription
Number representationUnderstanding diverse methods of number representation and applying them to real-world problems
Number operationUnderstanding number operations and their relationships and applying them to problem-solving
MeasurementUnderstanding the basics of measurement; measuring and estimating objects; and solving measurement problems
Geometric figuresExplaining and analyzing two-dimensional and three-dimensional geometric figures
Spatial reasoningUnderstanding visualization and spatial reasoning, and applying geometric models for problem-solving
Pattern and functionUnderstanding and analyzing patterns, relations, and functions
Algebraic understandingApplying and interpreting algebraic expressions, equations, inequalities, graphs, and other mathematical models to represent various situations and applying them for problem-solving
Data analysisUnderstanding and applying statistical methodology for data analysis
Valid estimationApplying statistical methodology and knowledge of probability for valid estimation
Table 6. An overview of students’ performance dataset.
Table 6. An overview of students’ performance dataset.
PerformanceAverageMaxMinSDCV
Number representation2.0275510.985648.61
Number operation1.3945310.544639.05
Measurement1.5046510.777351.66
Geometric figures3.2294510.987330.57
Spatial reasoning1.4312310.583240.75
Pattern and function1.8807310.662835.24
Algebraic understanding1.9358310.436222.53
Data analysis1.7615310.607034.46
Valid estimation1.6697310.609436.50
Table 7. The necessary parameters for Algorithm 1, IFB (5), IMFB (6), and MTSMDIS (7).
Table 7. The necessary parameters for Algorithm 1, IFB (5), IMFB (6), and MTSMDIS (7).
Algorithm τ k , τ 1 θ , θ k δ ¯ k , δ k α ¯ k , α k μ , μ k , p k
IFB (5) 1.999 2 ( m a x ( e i g e n v a l u e ( P T × P ) ) ) 0.999---
IMFB (6) 1.999 2 ( m a x ( e i g e n v a l u e ( P T × P ) ) ) 0.999- 1 2 k + 1 -
MTSMDIS (7) 1 2 ( m a x ( e i g e n v a l u e ( P T × P ) ) ) 1 5 2 6 + k 1 6 1 6 + k 1 2 1 6 + k μ = 0.5 , μ k = p k = 1 k 2
Algorithm 1OIPMPFBS-I 1.999 ( m a x ( e i g e n v a l u e ( P T × P ) ) ) 0.9990 1 k + 1 -
OIPMPFBS-II 1.999 ( m a x ( e i g e n v a l u e ( P T × P ) ) ) 0 1 k + 1 1 k + 1 -
PMPFBS 1.999 ( m a x ( e i g e n v a l u e ( P T × P ) ) ) 00 1 k + 1 -
DRIPFBS 1.999 ( m a x ( e i g e n v a l u e ( P T × P ) ) ) 0.999 1 k + 1 1 k + 1 -
Table 8. The numerical results of data training of each algorithm.
Table 8. The numerical results of data training of each algorithm.
AlgorithmNo. Iter.Training TimeAccuracyPrecisionRecallF1 Score
IFB (5)7470.020881.4072.0972.0972.09
IMFB (6)7470.020681.4072.0972.0972.09
MTSMDIS (7)5140.021481.4072.0972.0972.09
Algorithm 1OIPMPFBS-I510.004581.4072.0972.0972.09
OIPMPFBS-II5500.017381.4072.0972.0972.09
PMPFBS3710.013981.4072.0972.0972.09
DRIPFBS460.005781.4072.0972.0972.09
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jun-On, N.; Cholamjiak, W. Enhanced Double Inertial Forward–Backward Splitting Algorithm for Variational Inclusion Problems: Applications in Mathematical Integrated Skill Prediction. Symmetry 2024, 16, 1091. https://doi.org/10.3390/sym16081091

AMA Style

Jun-On N, Cholamjiak W. Enhanced Double Inertial Forward–Backward Splitting Algorithm for Variational Inclusion Problems: Applications in Mathematical Integrated Skill Prediction. Symmetry. 2024; 16(8):1091. https://doi.org/10.3390/sym16081091

Chicago/Turabian Style

Jun-On, Nipa, and Watcharaporn Cholamjiak. 2024. "Enhanced Double Inertial Forward–Backward Splitting Algorithm for Variational Inclusion Problems: Applications in Mathematical Integrated Skill Prediction" Symmetry 16, no. 8: 1091. https://doi.org/10.3390/sym16081091

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop