Next Article in Journal
R-DDQN: Optimizing Algorithmic Trading Strategies Using a Reward Network in a Double DQN
Next Article in Special Issue
Remote-Sensing Satellite Mission Scheduling Optimisation Method under Dynamic Mission Priorities
Previous Article in Journal
Enhanced Unmanned Aerial Vehicle Localization in Dynamic Environments Using Monocular Simultaneous Localization and Mapping and Object Tracking
Previous Article in Special Issue
Adaptive Neural Control for an Uncertain 2-DOF Helicopter System with Unknown Control Direction and Actuator Faults
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Expeditious Meta-Heuristic Clustering Method for Classifying Student Psychological Issues with Homogeneous Characteristics

School of Physics and Electronic Engineering, Hanshan Normal University, Chaozhou 521000, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(11), 1620; https://doi.org/10.3390/math12111620
Submission received: 23 April 2024 / Revised: 12 May 2024 / Accepted: 18 May 2024 / Published: 22 May 2024
(This article belongs to the Special Issue Deep Learning and Adaptive Control, 3rd Edition)

Abstract

:
Nowadays, cluster analyses are widely used in mental health research to categorize student stress levels. However, conventional clustering methods experience challenges with large datasets and complex issues, such as converging to local optima and sensitivity to initial random states. To address these limitations, this research work introduces an Improved Grey Wolf Clustering Algorithm (iGWCA). This improved approach aims to adjust the convergence rate and mitigate the risk of being trapped in local optima. The iGWCA algorithm provides a balanced technique for exploration and exploitation phases, alongside a local search mechanism around the optimal solution. To assess its efficiency, the proposed algorithm is verified on two different datasets. The dataset-I comprises 1100 individuals obtained from the Kaggle database, while dataset-II is based on 824 individuals obtained from the Mendeley database. The results demonstrate the competence of iGWCA in classifying student stress levels. The algorithm outperforms other methods in terms of lower intra-cluster distances, obtaining a reduction rate of 1.48% compared to Grey Wolf Optimization (GWO), 8.69% compared to Mayfly Optimization (MOA), 8.45% compared to the Firefly Algorithm (FFO), 2.45% Particle Swarm Optimization (PSO), 3.65%, Hybrid Sine Cosine with Cuckoo search (HSCCS), 8.20%, Hybrid Firefly and Genetic Algorithm (FAGA) and 8.68% Gravitational Search Algorithm (GSA). This demonstrates the effectiveness of the proposed algorithm in minimizing intra-cluster distances, making it a better choice for student stress classification. This research contributes to the advancement of understanding and managing student well-being within academic communities by providing a robust tool for stress level classification.

1. Introduction

Nowadays, student mental health is a significant problem in higher education institutes that requires intense attention and contribution [1]. Mental health disorders are common among students, with different challenges such as anxiety, depression, and stress significantly impacting academic performance [2,3]. Psychological factors like increasing workloads, disruptions in eating and sleeping patterns, poor time management, as well as changes in social and academic environments contribute to increased mental health problems among students [4,5]. According to a survey of 126,000 students, it is found that 20.3% of students experience depression and abnormal psychological problems [6]. In 2023, students faced 76% psychological problems, with 36% suffering anxiety, 28% depression, and 99% facing academic challenges [7]. Students faced considerable variations in mental health, influencing their perspectives on daily life and academic success [8]. Avoiding mental health can further lead to decreased academic performance, increased suspension rates, and social isolation, as well as contributing to long-term consequences on individuals’ lives. Efforts to address mental health problems can be significantly enhanced through the collection and analysis of data from university mental health databases. Early and precise detection of data is important for effective decision-making and optimizing psychological therapy services in higher education institutes [9,10].
Student mental health includes various emotional, cognitive, and behavioral factors affecting academic performance and well-being. These factors involve emotional distress, behavioral changes, cognitive difficulties, physical symptoms, social challenges, and academic effects. However, accurately classifying student mental health problems is challenging because these problems are subjective, overlap across different disorders, and some of them are homogeneous. Moreover, the homogeneity of certain mental health issues further increases the classification complexity [11,12,13]. Efforts to improve classification consider these complexities and adopt a comprehensive approach that effectively addresses them.
Recently, machine learning and computational techniques have been analyzed for classifying mental stress and the evaluation of anxiety and depression among students [14,15]. The deep learning methods are analyzed, particularly for analyzing expression demonstrations of mental illness and integrating diverse data sources for developing more precise evaluation methods for mental health conditions. Moreover, these methods are based on mental health [16] monitoring schemes customized for students, utilizing convolutional neural networks to accurately classify students’ mental health problems and address various mental health indicators, such as sleeping disorders, depression, suicidal ideation, personality development, and self-esteem. Different computational techniques, including the AdaBoost algorithm and the Mamdani fuzzy rule-based method, are being explored to determine students’ mental health issues. The aim is to identify an effective algorithm for classifying students and understand various factors that influence mental health among students [17,18].
Machine learning and computational techniques have made significant developments in various fields, including healthcare and psychology. However, when it comes to categorizing the mental health of students, they face certain challenges. These challenges arise due to the homogeneity of symptoms and the complexity of disorders [19,20,21]. Additionally, iteration in specific datasets [22], limited data availability, and accuracy issues in assessing psychological problems contribute to these challenges [23,24,25]. By addressing these challenges in the existing methods, the proposed research work aims to classify the mental health of students.
In response to these challenges in existing research work, researchers have introduced metaheuristic clustering algorithms (MCA) in recent years [26,27]. Traditional MCAs often struggle with a large amount of data and the complexity of the problems, encouraging the adoption of meta-framework algorithms. Nature-inspired algorithms are now commonly employed across different fields to address these issues effectively [28,29,30]. Clustering techniques, along with other data mining methods, have advanced significantly by using collective computational algorithms. The use of metaheuristic algorithms for clustering is gaining attraction as a choice to traditional methods, signifying significant progress in data analysis techniques [31,32]. However different MCAs have been suggested for addressing clustering issues [33] but their effectiveness remains challenging in accurately classifying student psychological problems. Therefore, there is still a need for further advancements in metaheuristic algorithms to effectively address clustering problems.
However, due to some challenges in the existing methods, this research developed a classification system for student psychological problems and proposed an iGWCA clustering algorithm. The iGWCA is a derivative of the original GWO algorithm, which is inspired by the social behavior and hunting strategies of grey wolves.
The original GWO can rapidly achieve better solutions due to its hierarchical structure. However, the GWO algorithm struggles with effectively classifying mental health problems characterized by homogeneous symptoms due to certain limitations. For example, GWO may ignore small fluctuations in data, resulting in misclassification, and GWO may not completely explore the solution space, due to prematurely converging to local optima, limiting its ability to find accurate classification solutions when symptoms are similar across different disorders. Due to these challenges, a modification in the GWO using parameter “a” is adjusted dynamically, which means its value changes during the optimization process. Dynamic adjustment of “a” is important because it allows the algorithm to adapt more effectively to different stages of optimization. This modification is important as it improved the algorithm’s adaptability, enabling it to make a better balance between exploration and exploitation. By dynamically adjusting “a” the iGWCA can potentially improve convergence speed, solution quality, and robustness across student classification problems. The main contribution of this proposed work is as follows.
  • Improvement of GWO. This research analyzes an iGWCA to classify student psychological problems. The improvement focuses on two key issues such as slow convergence and the tendency to local optima.
  • Performance Analysis. The effectiveness of the iGWCA algorithm has been comprehensively evaluated and confirmed to have better performance than other competing algorithms, including GWO, MOA [34], FFO [35], PSO [36], HSCCS [37], FAGA [38], and GSA [39,40].
  • Validation of iGWCA. Statistical analysis and comparison of obtained results from iGWCA with various established algorithms consistently demonstrate superior results in both accuracy and efficiency.

2. Related Work

In recent years, there has been a growing interest in utilizing various methods to assess students’ mental health. Data mining techniques are also explored to predict psychological issues in students. This aims to improve psychological management systems through early warning systems and continuous monitoring of at-risk students and increase problem alert efficiency [41]. To analyze burnout measures in students, highlighting the direct correlation between emotional tiredness and the onset of illnesses among participants is explored using partial least squares, and structural equation modeling is presented in [42]. The author classifies identical mental health classification in adolescents using latent profile analysis [43], however, it is bound to limited data size.
The authors proposed a neuro-fuzzy-based guidance and counseling system customized for college students, aiming to address the lack of guidance and counseling services in career path prediction [44]. An aspect-oriented convolutional neural network is introduced in [45], along with long short-term memory equipped with an attention mechanism, for academic emotion classification and recognition. These approaches offer valuable insights into students’ emotional experiences within academic environments. In [46], a clustering algorithm for intuitionistic fuzzy graphs is analyzed, focusing on edge density and practical applications, and introducing algorithms addressing clustering and minimum spanning tree problems, emphasizing parameterization for optimal data processing. There are different data sources, including social media activity, smartphone sensor data, electronic health records, and self-reported surveys, to gain insights into students’ psychological well-being. By applying machine learning algorithms to analyze these diverse data streams, researchers and mental health professionals aim to identify patterns, predict mental health results, and provide timely interventions to support students’ mental health [47,48,49]. The machine learning methods offer different advantages in evaluating the mental health of students, including early detection, objectivity, personalization, scalability, and resource efficiency, however, they have several challenges such as quality, interpretability, slow performance with a large number of datasets [50], limitations, such as their need for a large amount of data [51], reproducing model solution [52], and stochastic graph problems [53]. Additionally, the heuristic fuzzy c-means clustering algorithm struggles with limited dataset sizes [54]. Addressing these challenges involves careful data collection, interdisciplinary collaboration, robust evaluation techniques, and validation to maximize the benefits and minimize the risks of using machine learning in student mental health assessment.
Cluster analysis, an unsupervised data mining technique, is gaining attention for categorizing student psychological levels. With proper focus, these methods can offer valuable insights for guiding policy and implementation [55]. Clustering techniques are categorized into hierarchical clustering and partitional clustering. In hierarchical clustering, data members are initially assigned to individual clusters and then merged gradually, as one cluster. Partitional clustering, which is the focus of this paper, aims to divide the dataset into non-overlapping clusters without a nested structure [56,57,58]. Each cluster in this method is represented by its centroid, and initially, each student data object is assigned to the nearest centroid. The centroids are then updated based on the current problem and by optimizing criteria [59]. Traditional partitional clustering algorithms, such as the k-means algorithm are generally used for their speed and simplicity, aiming to minimize the average squared distance [60]. However, k-means’ efficiency heavily depends on the initial selection of cluster centers and is inclined to converge toward local optima. To address these challenges, the k-means++ algorithm is introduced to enhance the initial centroid selection [61]. Despite this improvement, k-means++ still faces difficulties in converging to local optima [62].
Therefore, in recent years, MCAs have gained attention for solving NP-hard problems like data clustering, aiming to minimize convergence issues and reduce the risk of being trapped in local optima. The literature review shows that there are multiple challenges associated with existing techniques to achieving better student classification, prediction accuracy, and student performance. Therefore, in this research work, iGWCA is proposed for effectively classifying student stress levels. There are the following main reasons for selecting iGWCA as a metaheuristic technique for clustering.
  • iGWCA provides simplicity and ease of implementation.
  • Only a few control parameters for tuning.
  • The iGWCA offers better convergence speed and an improved balance between exploration and exploitation processes.
  • It shows robustness and applicability in classifying psychological problems.

3. Problem Model Construction

Davies–Bouldin Index (DBI) [63] is a measure used in cluster analysis to evaluate the quality of clustering results. The objective function for optimizing DBI in the area of student clustering problems using iGWCA can be formulated as follows.
Assume that
  • X is the set of the student data set;
  • C is the set of clusters;
  • dij is the distance between data point i and data point j;
  • μk is the centroid of cluster k;
  • σi is the average distance from the centroid of cluster k to the student data point within the cluster.
f x = 1 C k = 1 C m a x i k σ i + σ k d μ i μ k
where
  • C is the set of clusters;
  • μi and μk are the centroids of clusters i and k, respectively;
  • dμiμk is the distance between μi and μk;
  • σi and σk are the average distances from centroids μi and μk to the data points within clusters i and k, respectively.
The constraints for this problem are that each student data point is assigned to only one cluster and possibly setting boundaries on the minimum and maximum number of student data points assigned to each cluster. These constraints can be formulated mathematically as follows.
Each student data point must belong to exactly one cluster
k = 1 C x i j = 1             j X
Number of student data points allocated to each cluster
j = 1 X x i j minimum   cluster   size             j C
j = 1 X x i j maximum   cluster   size             j C
where
  • Xij is the binary variable indicating whether data point i belongs to cluster j;
  • min_cluster_size and max_cluster size are the minimum and maximum allowed number of data points in a cluster, respectively.
The number of clusters k must be predefined
k = predefine   value
The DBI should be minimized to achieve more homogeneous clusters
Minimized   f ( x )

4. Proposed Method

4.1. Introduction to the Principle and Basic Process of GWO Algorithm

The GWO algorithm is initially suggested by Mirjalili in [64] and it is based on the social structure and hunting activities of grey wolves. GWO has gained recognition as a dependable and efficient optimization method in different applications. Its efficiency and ability to make a balance between exploration and exploitation have established it as a preferred approach for resolving complex optimization challenges [65]. The original GWO can rapidly achieve better solutions due to its structure. Furthermore, its low time complexity enables it to fulfill a significant role in addressing optimization problems.
Additionally, the innovative features and working principles of the GWO are explained as follows.
  • Initially, the algorithm randomly places a pack of grey wolves in the search space, with each wolf representing a potential solution corresponding to a set of cluster centroids.
  • Secondly, a predefined DBI objective function computes the current positions of the wolves (i.e., cluster centroids) and the student psychological problem data.
  • Then, during each iteration, the algorithm evaluates the fitness of each wolf using the objective function and updates their positions accordingly, aiming to minimize the DBI and improve clustering quality.
  • Finally, once the optimization reaches the maximum number of iterations, the final solution, consisting of optimized cluster centroids, is obtained. These clusters can be used to classify student psychological problems based on their similarities, with each student assigned to a specific cluster.
However, there are certain limitations, such as the GWO being susceptible to becoming trapped in local optima during the classification of psychological problems.

4.2. The Proposed iGWCA Algorithm

In this research, iGWCA is proposed, focusing on optimizing the balance between exploration and exploitation, these two processes are the fundamental aspects of optimization algorithms. Exploration involves extensively searching the solution space to discover better solutions and avoid local optima. As iterations increase, the position shifts towards exploitation, where solutions are developed to converge towards the optimal solution. To keep a balance between these processes is important for optimizing algorithm performance and motivating the exploration of new approaches. In the original GWO, the parameter “a” serves as a regulator for these elements, emphasizing exploration to entirely explore the solution space [26]. However, as iterations increase, “a” gradually transitions towards exploitation, concentrating on promising regions identified during the exploration phase. This transition is important for the algorithm to converge to an optimal solution. The proposed modification introduces variations in the linear behavior of parameter “a” throughout the optimization process. By adjusting “a” dynamically, the algorithm can effectively adapt its exploration and exploitation levels. This dynamic adjustment enables a more balanced approach, finally improving the convergence speed and solution quality.
a = 2 2 t t m a x k
where
  • t m a x is the total number of iterations, t is the current iteration, k is a constant
  • Step 1: Initialization of the student datasets
    • Initialize student stress datasets for dataset-I and dataset-II, then set the search agents and maximum number of iterations. The grey wolf at the location moves and updates its location according to the prey location and the algorithm defines the position of the best agent based on the positions of the grey wolf and prey by controlling parameters A and C using Equations (4) and (5). The value of the parameter A decreases during the algorithm’s simulation, and the fluctuation rate also decreases to maintain a balanced process.
  • Step 2: DBI Objective Function
    • Next, initialize the population of grey wolves within the search space and assess the quality of each grey wolf’s solution using the DBI objective function f(x).
  • Step 3: Hierarchy Formation
    • Grey wolves in the population are categorized into alpha (α), beta (β), and delta (δ) types of wolves. The α wolves are the best solutions, followed by β and δ wolves. Within the algorithm loop, continuously update the positions of the α, β, and δ wolves within the search space, ensuring all wolf positions remain feasible.
  • Step 4: Update α, β, and δ Positions:
    • The algorithm selects the α, β, and δ wolves based on their performance. These wolves estimate the prey’s location and update the positions of wolves around the prey according to the following Equations.
D = C     X p X t
X ( i t e r + 1 ) = X p A   D
A = 2 a · r 1 a
C = 2 r 2
D α = C 1 X α ( t ) X D β = C 2 X β ( t ) X D δ = C 3 X δ ( t ) X
X 1 = X α t A 1 D α X 2 = X β t A 2 D β X 3 = X δ t A 3 D δ
where
  • T is the current iteration, A and C are the co-efficient, X p is the position of prey, X is the position of a wolf.
  • D α ,   D β ,   D δ are the distances, r 1 , r 2 are random values and a is the value from 2 to 0.
  • Step 5: Update the Position of the prey
Update the positions of the prey using the following Equation (14).
X ( i t e r + 1 ) = X 1 + X 2 + X 3 3
In Equation (1), the parameter “a” is adjusted in a non-linear manner, gradually decreasing from 2 to 0. When “k” varied from 0 to 1, the algorithm focused more on exploitation, potentially leading to a compromise in search quality. However, for “k” values > 1, the algorithm thoroughly explores the search space before transitioning to exploitation. To improve the performance of GWO, it is important to experiment with various trials to determine the most suitable value for “k”.
To address the reduced number of iterations during exploitation phases, a mapping method is proposed to conduct local searches around the best solution. This method requires mapping the position of the best wolf to a new position and evaluating whether it results in improved fitness. If the fitness improves, the best wolf is moved to this new position using Equation (15)
X n = X α t + r ( U b L b ) m 0.5
where
  •   U b and   L b are the upper and lower boundaries, r is the center, and m is the mapping parameter, which is updated at each iteration
m ( t + 1 ) = 4 × m × ( 1 m t )
  • Step 6: Boundary Constraints
    • The DBI function evaluates student clustering quality based on cluster separation and constraints such as ensuring each student data point belongs to exactly one cluster and allocating the correct number of student data points to each cluster.
  • Step 7: Update Fitness Values
Analyze the fitness of each wolf’s updated position using the DBI objective function.
  • Step 8: Update Hierarchy of α, β, and δ wolves
    • Update the hierarchy of wolves according to their fitness values.
  • Step 9: Termination Criteria
    • Update the convergence curve to monitor optimization progress until a stopping criterion is satisfied. When the stopping criterion is satisfied, return the best solution obtained from the optimization process otherwise, start the process.
The iGWCA iteratively searches the solution space, with the α, β, and δ wolves guiding the exploration and exploitation process. The flowchart and pseudo-code of the proposed algorithm are presented in Algorithm 1 and Figure 1, respectively.
Algorithm 1: Pseudo Code of the iGWCA. Algorithm 1. PSEUDO code of proposed method.
1Input number students stress dataset
2Initialize the population of grey wolves randomly within the search space using iGWCA
3Initialize a , A and C
4Evaluate the fitness value of each grey wolf solution using the Davies-Bouldin Index f(x)
5While the stopping criterion is not met
6for each search agent
7Update the positions of X α , X β   a n d   X δ wolves
8end for
9Update a , A and C
10Evaluate the fitness value of each grey wolf Davies-Bouldin Index f(x)
11Update the new position of X α , X β   a n d   X δ wolves
12t = t + 1
13End while
14Return  X α as the best obtained solution for iGWCA

4.3. Time Complexity Analysis of iGWCA Algorithm

In the iGWCA, if the number of population size is N, D is the dimension of the student dataset, and Max: Iter is the number of maximum iterations. The time complexity of the iGWCA is presented in Table 1 using big O notation. The algorithm’s time complexity plays a significant role in evaluating the abilities and limitations of the algorithm.
The actual time achieved on the computer for the best-case scenario is 8.162354 s, for the worst-case scenario is 20.817287 s, and for the average-case scenario is 15.979641 s.

5. Experimental Results

This section presents the results of the proposed iGWCA algorithm, with a detailed description provided below. The results are then discussed, analyzed, and compared to those of other algorithms.

5.1. Experimental Method Description

The proposed algorithm is implemented in MATLAB R2019a on a computer equipped with an Intel(R) Core(TM) i5-7200U CPU @ 2.50 GHz, 16 GB of memory, and running Windows 10 with a 64-bit operating system. The iGWCA algorithm performance is evaluated with competent algorithms such as GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA. The best optimal solution for our datasets based on the average and standard deviation (SD) are determined. Each algorithm simulates 30 independent runs for dataset-I and dataset-II, with each consisting of 100 maximum number of iterations. Table 2 demonstrates the parameter configurations for all algorithms. In iGWCA the search agents are 30, however, in PSO, the acceleration coefficients c1 and c2 are set to 1.5 and 2, respectively. Meanwhile, the inertia weight (w) dynamically changes from 0.9 to 0.5 during the optimization process [66].
The statistical metrics, including mean (avg), standard deviation (SD), and efficiency, are mentioned in Equations (17) and (18). These metrics provide insights into the overall performance and variability of the proposed method. Moreover, to ensure robustness, each algorithm is run in more >30 trials, and the results are rounded to five decimal places to minimize statistical errors and ensure statistically significant outputs. The mean computes the average of the best-obtained values by the algorithm across a predetermined number of runs as expressed in Equation (17).
A v g = 1 n = i = 1 n F F V i
where
  • n is the number of runs, and the FFVi represents the best value obtained in the ith iteration.
The SD measure is explored to analyze whether the algorithm under evaluation consistently achieves the same best value across multiple trials, therefore it serves as an accuracy test for the algorithm results. It is expressed as follows.
S D = 1 n 1 i = 1 n F F V i A v g 2
E f f i c i e n c y = F F V m i n F F V i × 100 %
where
  • FFVmin is the minimum value of the fitness function.

5.2. Discussion on Statistical, Convergence Performance

Nowadays, statistical analysis has gained more attention as an important tool for evaluating the performance of computational methods. These analyses are commonly employed to compare the efficiency of different algorithms in the proposed work as mentioned in Table 3 using dataset-I. In this research, the iGWCA demonstrated better accuracy compared to GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA. Moreover, the iGWCA algorithm’s ability to avoid becoming stuck in local optima, unlike traditional methods, is beneficial because it integrates social behaviors and hunting strategies observed in grey wolves.

5.2.1. Efficiency Experiments

The effectiveness of the iGWCA method is further validated through statistical analysis, confirming it shows better accuracy. This comparison is important for determining whether a proposed method offers a significant improvement over existing methods for a given problem. Researchers often prefer this analysis due to their simplicity and efficiency in computational processing. The statistical analysis includes the best, worst, and average value, SD, number of hits, and computational time, analyzing dataset-I with parameters as mentioned in Table 3.
Table 3 compares the performance of various optimization algorithms. The average best values obtained for iGWCA, GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA are 0.52461, 0.53248, 0.57456, 0.57301, 0.53781, 0.54452, 0.57150, and 0.57452, respectively, with iGWCA demonstrating the best performance. Considering computational time, iGWCA achieves the lowest computational time at 8.162354 s, followed by GWO at 12.4542199 s, MOA at 18.791092 s, FFO at 26.140449 s, PSO at 29.559220 s, HSCCS at 15.777735 s, FAGA at 14.877140 s and GSA at11.791202 s. Additionally, FFO shows lower computational time than PSO, while GSA ranks second in this analysis. However, iGWCA has the lowest computational time among competent algorithms, emphasizing its efficiency. On average, the proposed algorithm shows a 5.175% reduction in optimal value compared to other techniques. These results indicate that the iGWCA algorithm effectively obtains the best optimal solution, making it a better choice for student stress classification tasks due to its superior performance. Therefore, it is a better choice than other methods.
The iGWCA convergence is marked by a better choice, aiming to make a balance between exploration and exploitation while improving solutions toward the global optimum solution obtained using dataset-I. It consistently exhibits smooth convergence curves and faster rates compared to GWO, particularly for optimization problems. However, GWO may show slower convergence rates as compared to iGWCA as mentioned in Figure 2. In WOA, convergence involves iteratively improving solutions through exploration and exploitation phases, resembling whale hunting behavior. PSO convergence curve typically shows a gradual decrease in the objective function with fluctuations due to particle movement. HSCCS combines elements of the sine cosine and cuckoo search algorithms to enhance optimization, measured by reducing the objective function. In FAGA, solutions are refined through a combination of evolutionary operators from the Firefly and Genetic Algorithms. GSA attracts solutions towards better ones, mimicking gravitational attraction.
In addition to this, convergence in deep learning models is assessed by monitoring loss function decrease over epochs, while data mining technology convergence depends on specific algorithm strategies. AdaBoost convergence is followed by observing classification accuracy improvements, and the gradient boosting machine algorithm minimizes a differentiable loss function similar to AdaBoost.
Table 4 presents a comparative analysis of the efficiency of various algorithms. The machine learning algorithm outperforms the AdaBoost algorithm with a performance ratio of 0.8571. Furthermore, the CatBoost model achieves a performance ratio of 0.826, exceeding the gradient boosting machine algorithm. However, the data mining method’s performance ratio is slightly lower compared to the proposed method. These results demonstrate the efficiency and capability of the proposed method for further applications.

5.2.2. Exploring Data Credibility

The iGWCA shows multiple advantages over GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA, such as smooth convergence, lowest optimal value, a balance between exploration and exploitation as mentioned in Figure 3, ease of implementation, and a smaller number of parameters. These attributes make iGWCA a better choice for addressing the student clustering problem analyzing dataset-I. The proposed method is further analyzed as presented in Table 5 using common evaluation metrics such as False Negative Ratio (FNR), False Positive Ratio (FPR), True Negative Ratio (TNR), and True Positive Ratio (TPR) conducted [69,70].
Table 5 demonstrates the confusion matrix of GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA, indicating that iGWCA has the best values for TPR, TNR, FPR, and FNR, with 0.9666, 0.35, 0.75, and 0.20687, respectively. To prove the ability of iGWCA in the classification of psychological problems, it is compared with different competent algorithms. The obtained result shows that iGWCA achieved better performance than others. In addition to this, it is observed that the results of GWO and PSO show better performance than MOA, FFO, HSCCS, FAGA, and GSA.
Figure 4 shows the dispersion of student data points across different trials, serving as a measure of consistency compared to other methods. Analyzing convergence patterns and box plots allows for a detailed analysis of the proposed method’s performance in student psychological problems. The highest and lowest values are represented by cross symbols positioned at the top and bottom of the box, respectively. The rectangular box shows the interquartile range, including half of the data (50%). This visualization effectively shows the distribution of optimal values achieved across multiple runs, providing insights into strategies to reduce the risk of local optima. Furthermore, the probability of attaining the minimum fitness function value is particularly higher, given that the median for the proposed method is closer to the lower quartile.
Figure 3 compares the box plots of the proposed method with GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA. The minimum and maximum data points are presented as points at the ends of the whiskers extending from the box. Box plots show the distribution of quantitative data, facilitating comparisons among fitness functions across iGWCA and other algorithms. The averages of iGWCA outperform GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA. Specifically, PSO outperforms FFO, HSCS, FAGA, and GSA, while GWO achieves a lower average value than GWO. These box plots indicate better chances of obtaining the minimum fitness function value, as the median from iGWCA is closer to the lower quartile.

5.3. Comparative Analysis of Classification Results Based on Dataset-I

5.3.1. Dataset Description

The study involved analyzing student stress data from dataset-I, which consisted of 1100 students obtained from the Kaggle database, each characterized by three attributes. In this work, these attributes are selected as academic performance as test case 1, future career concerns as test case 2, and headaches as test case 3. Below is an explanation of these attributes.
(1)
Test Case 1: Academic Performance Stress
Academic performance stress among students is a universal problem characterized by psychological and emotional stress arising from various academic pressures. These stressors include the pressure to succeed, fear of failure, heavy workload, time pressure, performance expectations, competition, perfectionism, uncertainty about the future, and challenges with time management and study skills. Such stress can lead to physical symptoms like headaches and fatigue, as well as mental health concerns such as anxiety and depression. Identifying and addressing academic stress is significant for promoting student well-being and academic success. Time management techniques, seeking support from teachers or counselors, practicing self-care, and developing coping skills can help students manage academic stress more effectively.
(2)
Test Case 2: Future Career Concern Stress
Future career concern stress among students during their studies is a significant issue characterized by anxiety and fear about their professional prospects after graduation. This stress arises from various factors, including uncertainty about job opportunities, fear of career expectations, pressure to secure employment in a competitive market, and concerns about academic choices with future career goals. Additionally, students may experience stress related to the perceived importance of obtaining internships, networking, and building relevant skills for their desired career paths. Future career concern stress can be determined by physical symptoms such as tension and fatigue, as well as psychological symptoms like worry and self-doubt. Addressing these stress problems is important for supporting students’ well-being and career development. Strategies such as career counseling, networking opportunities, skill-building workshops, and fostering resilience can help students navigate future career concerns and stress more effectively, enabling them to make informed decisions about their professional trajectories.
(3)
Test Case 3: Headache Stress
Headache stress among students during their studies is a significant issue characterized by physical discomfort and tension in the head region. This stress is often triggered by academic pressures and demands. Factors contributing to headache stress include excessive studying, continuous screen time, insufficient rest, dehydration, and high levels of tension related to academic performance. Headache stress can significantly impact students’ ability to concentrate, focus, and retain information, thus affecting both their academic performance and overall well-being. Addressing headache stress involves implementing strategies such as taking regular breaks, staying hydrated, practicing relaxation techniques, managing time effectively, and seeking medical advice if headaches exist. By effectively managing headache stress, students can alleviate discomfort and optimize their study environment to improve their academic success. The iGWCA classifies the student data into two clusters, and statistical analysis is conducted using SPSS to evaluate the results obtained from the clustering process.

5.3.2. Classification Result and Analysis

Figure 4 shows the classification of student psychological problems on the dataset-I with three attributes, the number of clusters k = 2 is selected in this study. The concept of similarity is measured via Euclidean distance between student data points and centroids. Each small circle represents a student data point in cluster 1 and cluster 2, and they are connected with each data point. However, the limitation is set as each student data point must belong to exactly one cluster, as explained in Section 3. Then, the boundaries do not cross each other, and the data points become properly grouped.
According to Figure 4a, and in view of the obtained results, it is observed that the iGWCA classifies the student data points having similar characteristics into clusters 1 and 2 with the lowest average value. The graphical representation obtained by iGWCA facilitates effective exploration of the results for cluster 1 and cluster 2. The first one is a cluster with 824 population distribution where the clusters should be well separated. The second one is a 276 population distribution with separate clusters. According to Figure 4b, which shows the results obtained from GWO, the student dataset is grouped into cluster 1 and cluster 2 (n = 709, 391). Figure 4c shows the results obtained from MOA, showing a similar clustering pattern with cluster 1 containing n = 730 data points and cluster 2 containing n = 370 data points. Additionally, FFO resulted in a dataset of points of n = 748 and 352 for clusters 1 and 2 clusters, as mentioned in Figure 4d, respectively. According to Figure 4e, PSO obtained n = 746 and 354 dataset sizes for both clusters. FAGA obtained (n = 730, 370) sample size according to Figure 4f. According to Figure 4g,h HSCCS and GSA are obtained (n = 711, 389) for both clusters. However, iGWCA consistently demonstrates superior efficiency in classifying more accurate and compact student clusters across different cluster configurations. Moreover, from the graphical representation of the cluster, it is confirmed that iGWCA consistently exhibits the best performance and maintains a balanced distribution among all student data points inside the cluster.
The performance of iGWCA utilized in this study is presented in Table 6, demonstrating its superiority over other methods employed. The clustering results are evaluated using six statistical indicators, including cluster size, average, median, rank, and SD, with mean rank serving as the primary metric for assessing performance. These results are tabulated in Table 6, which illustrates the distribution of the dataset from the simulation experiment on students’ psychological problems. The obtained result shows that iGWCA consistently obtained the lowest average value across test cases 1 and 2, indicating its effectiveness, particularly in test case 3, where iGWCA outperforms the competing algorithms, securing the top rank among them. Finally, to further validate the obtained results, comparisons are made with other competent algorithms. The iGWCA demonstrates not only lower computational time but also higher accuracy compared to others. These findings show the robustness of the iGWCA algorithm in the classification of psychological problems based on attributes, indicating its superiority over established techniques. The comprehensive evaluation, along with other algorithms, underscores the credibility of iGWCA as a better choice for student psychological problems. In Table 6 there is a variability in SD observed, and it can be attributed to several factors. Firstly, it is important to note that each algorithm undergoes >30 trials. During these trials, significant fluctuations in both minimum and maximum values are observed. Then, the SD values are obtained from these trials. This variability further shows the importance of robust statistical analysis to monitor the range of obtained results.
According to Table 6, the average values obtained using iGWCA, GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA for test case 1, considering cluster 1, are 5.89%, 7.70%, 7.91%, 8.11%, and 8.08%, respectively. For cluster 2, the average value for iGWCA is 20.60%, GWO is 21.36%, MOA is 21.72%, FFO is 22.01%, PSO is 21.98%, HSCCS, GSA is 7.72%, and FAGA is 7.91%. In this case, HSCCS outperforms PSO, however, iGWCA demonstrates the best performance overall. Therefore, it is suggested that the proposed method shows better performance compared to the other methods. In test case 2, considering cluster 1, the iGWCA achieves an average value of 1.89%, GWO 1.86%, MOA 1.94%, FFO 1.97%, PSO 1.98%, HSCCS 1.86%, FAGA is 1.94% and GSA is 1.87%. Particularly, iGWCA has better performance than PSO and FFO by achieving the lowest average value, indicating its efficacy in generating more precise and concise clusters. For cluster 2, the average values obtained for iGWCA are 3.50%, GWO is 3.39%, MOA is 4.06%, FFO is 4.09%, PSO is 4.07%, HSCCS, GSA is 4.08%, and FAGA is 4.06%.
In test case 3, considering cluster 1, the average values obtained are 1.84% for iGWCA, 1.86% for GWO, 1.93% for MOA, 1.96% for FFO, 1.97% for PSO, 1.86% for HSCCS, 1.93% for FAGA, and GSA is 1.87%. From these results, iGWCA, GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA show better performance. However, iGWCA consistently demonstrates superior efficiency in classifying more accurate and compact student clusters across different test cases and cluster configurations. Although iGWCA and GWO demonstrate comparable performance, PSO lags slightly in achieving lower average values, indicating its lesser effectiveness in this specific context. For cluster 2 in test case 3, the average values obtained are 3.31% for iGWCA, 3.68% for GWO, 3.65% for MOA, 3.67% for FFO, and 3.65% for PSO, however, HSCCS is 3.69%, FAGA is 3.65%, and GSA is 3.68%. The iGWCA significantly outperforms both FFO and MOA, demonstrating its superior ability to address the dataset-I more effectively. While the differences between FFO and PSO are relatively minor in this case, the proposed method maintains a slight edge by achieving a marginally lower average value.
Based on graphical representations of these two clusters, as mentioned in Figure 4, using iGWCA and other algorithms can help in the classification and monitoring of psychological problems more accurately. The graphical representation makes it easier to explore the results of cluster 1 and cluster 2, such as how these students in dataset-I are grouped and how they relate to one another within each cluster. Researchers can assess the consistency and separation of clusters visually, checking whether the two clusters are well-defined and different from one another. This can ensure the chosen number of clusters and the overall quality of the clustering solution. According to the obtained result mentioned in Table 6, iGWCA demonstrates the lowest average value between two clusters across different test cases, compared to GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA. The analysis confirms iGWCA’s better performance in classifying student psychological characteristics, as mentioned in the DBI function.

5.4. Comparative Analysis of Classification Results Based on Dataset-II

5.4.1. Dataset Description

To explore the efficiency of the proposed algorithm, it is further verified using dataset-II comprising 824 individuals, including students, housewives, professionals, and others obtained from [71]. In this research work, the number of clusters k = 2, and two attributes are selected. Test Case 4 represents growing stress, and Test Case 5, represents social weakness. The original values in this dataset are “No/Maybe/Yes”, but for this work, they are mapped to 10/15/20, respectively. The following is the explanation of these attributes.
(1)
Test Case 5: Growing Stress
Growing stress is an escalating level of psychological pressure experienced by individuals over time. It refers to the increasing burden of stressors that every individual perceives in their lives, which may stem from various sources such as work, school, relationships, financial concerns, or other life events. Growing stress can manifest in physical, emotional, and behavioral symptoms, including but not limited to fatigue, irritability, difficulty concentrating, changes in sleep patterns, and heightened anxiety. It is a significant aspect of mental health research and is often studied to understand its impact on individuals’ well-being and to develop strategies for stress management and prevention.
(2)
Test Case 6: Social Weakness
Social weakness refers to a condition where individuals experience difficulties in their social interactions, relationships, or support networks. It encompasses various challenges such as feelings of isolation, loneliness, inadequate social skills, a lack of social support, or difficulties in forming and maintaining meaningful connections with others. Social weakness can have negative impacts on individuals’ mental and emotional well-being, contributing to stress, depression, anxiety, and a sense of disconnection from society.

5.4.2. Classification Result and Analysis

Table 7 shows the comparative analysis of iGWCA, GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA. In test case 4, for cluster 1, the minimum values achieved are iGWCA 11.0%, GWO 16.9%, MOA 16.1%, FFO 18.6%, PSO 15.9%, HSCCS 18.6%, FAGA 18.4%, and GSA 15.9%. However, for cluster 2, the proposed method achieves the lowest average value of 11.5%. Therefore, it is suggested that the proposed method exhibits superior performance compared to the other methods. Similarly, in test case 5, for cluster 1, the average values obtained are iGWCA 16.0%, GWO 13.4%, MOA 17.9%, FFO 17.8%, PSO 14.3%, HSCCS 17.7%, FAGA 17.6%, and GSA 14.3%. Particularly, GWO outperforms others by achieving the lowest average value, however, iGWCA’s average value is slightly high in this case. For cluster 2, the average values are iGWCA 13.7%, GWO 18.2%, MOA 14.0%, FFO 14.4%, PSO 17.8%, HSCCS and GSA 14.4%, and FAGA 18.2%. The iGWCA demonstrates efficacy in generating more precise and concise clusters, while MOA and FFO yield better results than GWO and GSA.
The comparative analysis of iGWCA, GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA, based on the clustering experiments shows that iGWCA consistently demonstrates better efficiency in classifying more accurate and compact student clusters across test case 4, and test case 5 with different clusters configuration.
The combination of customized improvements in iGWCA and the utilization of the DBI function enables it to outperform GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA in classifying dataset-I and dataset-II student psychological problems. By optimizing cluster quality according to the DBI objective function, iGWCA ensures more accurate and meaningful classifications, making it a better choice for psychological classification problems in higher education institutes.
The iGWCA combines improvements to the original GWO algorithm with the help of dynamic parameter “a” making it more attractive at addressing the classification of student psychological problems. These improvements include improved exploration and exploitation and better handling of constraints (each student data point is assigned exclusively to one cluster, with the number of student data points allocated to each cluster specified) customized to the classification of dataset-I and dataset-II. The DBI objective function is well-suited for evaluating the quality of clusters formed during optimization. It measures both the inter-cluster separation and intra-cluster integration, providing a comprehensive assessment of clustering quality. By optimizing this objective function, iGWCA ensures the formation of clusters that are both different from each other and internally unified, leading to more meaningful and accurate classifications of student psychological problems. The iGWCA likely shows better convergence properties and exploration–exploitation balance compared to GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA. Through its improved optimization mechanisms, iGWCA can efficiently navigate the solution space, effectively balancing exploration to discover different cluster configurations and exploitation to refine better solutions. These improvements enable iGWCA to allow it to achieve more optimal solutions with improved convergence speed and greater consistency compared to other methods.

6. Conclusions

In this proposed work, the classifications of student stress levels are analyzed using iGWCA. The classification of student stress levels using the proposed method presents a capable approach to understanding and addressing the complexities of student psychological dataset-I and dataset-II. The experimental analysis demonstrates that the iGWCA achieves the lowest average values for test case 1 at 5.89% and 20.60% for test case 2. For test case 3, the iGWCA method obtained 1.62% and 3.5% for cluster 1 and cluster 2, respectively, compared to other methods. However, in test case 4 and test case 5, iGWCA achieved 11.0% and 11.5%, 16.0%, and 13.7% average values, respectively. The comparison between iGWCA, GWO, MOA, FFO, PSO, HSCCS, FAGA, and GSA indicates the significance of selecting an appropriate optimization technique for effectively classifying student stress levels. However, iGWCA shows better performance in this research work, offering more accurate and efficient classification results. The importance of classifying stress levels among students cannot be ignored. Understanding student stress levels is important for universities, administrators, and policymakers to provide effective help and support. By accurately classifying stress levels, educational institutions can modify resources and programs to meet students’ specific needs, ultimately fostering a healthier and more conducive learning environment. Moreover, the classification of stress levels enables early detection of potential issues and facilitates measures to minimize the adverse effects of stress on academic performance, mental health, and overall well-being. The classification of student stress levels using optimization techniques like iGWCA not only offers valuable insights into student well-being but also covers the way for evidence-based interventions that promote resilience and academic achievement. The experimental result shows that iGWCA achieves the lowest best and average values, at 0.52461 and 0.54532, respectively.
The iGWCA for classifying student psychological issues with homogeneous characteristics presents few limitations. Initially, student psychological issues are complex, sometimes influenced by various factors such as genetics, environment, and individual experiences. iGWCA, like many optimization algorithms, may struggle to capture the full complexity of these issues and adequately classify them based on limited input data. Secondly, the effectiveness of iGWCA relies heavily on the quality and availability of data. If the dataset used for classification is incomplete or contains errors, it may lead to inaccurate classifications of the results. Then the performance of the algorithm could be impacted by parameter configurations, necessitating careful tuning, which might present difficulties for users lacking extensive optimization skills. Addressing these limitations and improving the algorithm’s scalability, ease of implementation, and flexibility are important areas for future research and development, particularly for extensive applications in classification systems. This study establishes a robust foundation for future advancements in the classification of psychological problems. The success of the iGWCA method suggests following a few promising directions for further exploration.
(1)
Evaluation of other optimization algorithms, such as whale optimization or moth flame optimization, for classifying student psychological issues is a promising area for future research. By refining the algorithm parameters and exploring adaptive mechanisms, its efficiency will further improve in accurately classifying different psychological issues.
(2)
Additionally, integrating multimodal data sources, such as physiological measurements, behavioral patterns, and academic performance metrics, could further enhance the algorithm classification accuracy.
(3)
Furthermore, investigating the customization of the algorithm to address the unique challenges modeled by student psychological data, such as variability in symptom expression and the dynamic nature of mental health conditions, is significant.
Overall, the effectiveness of the iGWCA approach acts as an attractive tool for future research efforts aimed at further improving the accuracy and applicability of psychological problems.

Author Contributions

Methodology, M.S.S. and X.D. software, M.S.S.; writing—original draft preparation, M.S.S.; writing—review and editing, X.D., C.W. and G.Z.; visualization, C.W. and Y.L.; supervision, X.D.; project administration, X.D. and G.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the innovation teams of ordinary Universities in Guangdong Province (2021KCXTD038, 2023KCXTD022), Key Laboratory of Ordinary Universities in Guangdong Province (2022KSYS003), China University Industry, University, and Research Innovation Fund Project (2022XF058), Key Discipline Research Ability Improvement Project of Guangdong Province (2021ZDJS043, 2022ZDJS068), Hanshan Normal University Research platform project(PNB2104), and the Quality Engineering Project of Hanshan Normal University (HSJYS-KC22719).

Data Availability Statement

Not applicable.

Acknowledgments

The authors are grateful for the support of the work by the Research Grant Scheme, Hanshan Normal University, Guangdong, China. The authors also thank the anonymous reviewers for their constructive advice.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Antonopoulou, M.; Mantzorou, M.; Serdari, A.; Bonotis, K.; Vasios, G.; Pavlidou, E.; Trifonos, C.; Vadikolias, K.; Petridis, D.; Giaginis, C. Evaluating Mediterranean diet adherence in university student populations: Does this dietary pattern affect students’ academic performance and mental health? Int. J. Health Plan. Manag. 2020, 35, 5–21. [Google Scholar] [CrossRef] [PubMed]
  2. Auerbach, R.P.; Mortier, P.; Bruffaerts, R.; Alonso, J.; Benjet, C.; Cuijpers, P.; Demyttenaere, K.; Ebert, D.D.; Green, J.G.; Hasking, P. WHO world mental health surveys international college student project: Prevalence and distribution of mental disorders. J. Abnorm. Psychol. 2018, 127, 623. [Google Scholar] [CrossRef]
  3. Eisenberg, D.; Hunt, J.; Speer, N. Mental health in American colleges and universities: Variation across student subgroups and across campuses. J. Nerv. Ment. Dis. 2013, 201, 60–67. [Google Scholar] [CrossRef] [PubMed]
  4. Herbert, C.; Meixner, F.; Wiebking, C.; Gilg, V. Regular physical activity, short-term exercise, mental health, and well-being among university students: The results of an online and a laboratory study. Front. Psychol. 2020, 11, 491804. [Google Scholar] [CrossRef]
  5. Idris, F.; Zulkipli, I.N.; Abdul-Mumin, K.H.; Ahmad, S.R.; Mitha, S.; Rahman, H.A.; Rajabalaya, R.; David, S.R.; Naing, L. Academic experiences, physical and mental health impact of COVID-19 pandemic on students and lecturers in health care education. BMC Med. Educ. 2021, 21, 542. [Google Scholar] [CrossRef]
  6. Tang, Q.; Zhao, Y.; Wei, Y.; Jiang, L. Research on the mental health of college students based on fuzzy clustering algorithm. Secur. Commun. Netw. 2021, 2021, 3960559. [Google Scholar] [CrossRef]
  7. American College Health Association. American College Health Association-National College Health Assessment III: Undergraduate Student Reference Group Data Report Spring 2023; American College Health Association: Silver Spring, MD, USA, 2023; pp. 1–117. [Google Scholar]
  8. Parker, P.C.; Perry, R.P.; Hamm, J.M.; Chipperfield, J.G.; Pekrun, R.; Dryden, R.P.; Daniels, L.M.; Tze, V.M. A motivation perspective on achievement appraisals, emotions, and performance in an online learning environment. Int. J. Educ. Res. 2021, 108, 101772. [Google Scholar] [CrossRef]
  9. Naser, A.Y.; Dahmash, E.Z.; Al-Rousan, R.; Alwafi, H.; Alrawashdeh, H.M.; Ghoul, I.; Abidine, A.; Bokhary, M.A.; AL-Hadithi, H.T.; Ali, D. Mental health status of the general population, healthcare professionals, and university students during 2019 coronavirus disease outbreak in Jordan: A cross-sectional study. Brain Behav. 2020, 10, e01730. [Google Scholar] [CrossRef]
  10. Viskovich, S.; Pakenham, K.I. Randomized controlled trial of a web? Based Acceptance and Commitment Therapy (ACT) program to promote mental health in university students. J. Clin. Psychol. 2020, 76, 929–951. [Google Scholar] [CrossRef]
  11. Cuijpers, P.; Auerbach, R.P.; Benjet, C.; Bruffaerts, R.; Ebert, D.; Karyotaki, E.; Kessler, R.C. The world health organization world mental health international college student initiative: An overview. Int. J. Methods Psychiatr. Res. 2019, 28, e1761. [Google Scholar] [CrossRef]
  12. Lipson, S.K.; Eisenberg, D. Mental health and academic attitudes and expectations in university populations: Results from the healthy minds study. J. Ment. Health 2018, 27, 205–213. [Google Scholar] [CrossRef]
  13. Patel, V.; Flisher, A.J.; Hetrick, S.; McGorry, P. Mental health of young people: A global public-health challenge. Lancet 2007, 369, 1302–1313. [Google Scholar] [CrossRef] [PubMed]
  14. Sau, A.; Bhakta, I. Screening of anxiety and depression among seafarers using machine learning technology. Inform. Med. Unlocked 2019, 16, 100228. [Google Scholar] [CrossRef]
  15. Ahuja, R.; Banga, A. Mental stress detection in university students using machine learning algorithms. Procedia Comput. Sci. 2019, 152, 349–353. [Google Scholar] [CrossRef]
  16. Du, C.; Liu, C.; Balamurugan, P.; Selvaraj, P. Deep learning-based mental health monitoring scheme for college students using convolutional neural network. Int. J. Artif. Intell. Tools 2021, 30, 2140014. [Google Scholar] [CrossRef]
  17. Pandey, D.C.; Kushwaha, G.S.; Kumar, S. Mamdani fuzzy rule-based models for psychological research. SN Appl. Sci. 2020, 2, 913. [Google Scholar] [CrossRef]
  18. Ogunseye, E.O.; Adenusi, C.A.; Nwanakwaugwu, A.C.; Ajagbe, S.A.; Akinola, S.O. Predictive analysis of mental health conditions using AdaBoost algorithm. ParadigmPlus 2022, 3, 11–26. [Google Scholar] [CrossRef]
  19. Chung, J.; Teo, J.; Minutolo, A. Mental Health Prediction Using Machine Learning: Taxonomy, Applications, and Challenges. Appl. Comput. Intell. Soft Comput. 2022, 2022, 9970363. [Google Scholar] [CrossRef]
  20. Ku, W.L.; Min, H. Evaluating Machine Learning Stability in Predicting Depression and Anxiety Amidst Subjective Response Errors. Healthcare 2024, 12, 625. [Google Scholar] [CrossRef] [PubMed]
  21. Katarya, R.; Maan, S. Predicting mental health disorders using machine learning for employees in technical and non-technical companies. In Proceedings of the 2020 IEEE International Conference on Advances and Developments in Electrical and Electronics Engineering (ICADEE), Coimbatore, India, 10–11 December 2020; pp. 1–5. [Google Scholar]
  22. Tuan, T.M.; Lan, L.T.H.; Chou, S.-Y.; Ngan, T.T.; Son, L.H.; Giang, N.L.; Ali, M. M-CFIS-R: Mamdani complex fuzzy inference system with rule reduction using complex fuzzy measures in granular computing. Mathematics 2020, 8, 707. [Google Scholar] [CrossRef]
  23. Singh, S.; Hooda, S. A Study of Challenges and Limitations to Applying Machine Learning to Highly Unstructured Data. In Proceedings of the 2023 7th International Conference on Computing, Communication, Control and Automation (ICCUBEA), Pune, India, 18–19 August 2023; pp. 1–6. [Google Scholar]
  24. Krichen, M. Convolutional neural networks: A survey. Computers 2023, 12, 151. [Google Scholar] [CrossRef]
  25. Hornyák, O.; Iantovics, L.B. AdaBoost Algorithm Could Lead to Weak Results for Data with Certain Characteristics. Mathematics 2023, 11, 1801. [Google Scholar] [CrossRef]
  26. Ahmadi, R.; Ekbatanifard, G.; Bayat, P. A modified grey wolf optimizer based data clustering algorithm. Appl. Artif. Intell. 2021, 35, 63–79. [Google Scholar] [CrossRef]
  27. Nasiri, J.; Khiyabani, F.M. A whale optimization algorithm (WOA) approach for clustering. Cogent Math. Stat. 2018, 5, 1483565. [Google Scholar] [CrossRef]
  28. Faieghi, M.R.; Delavari, H.; Baleanu, D. A novel adaptive controller for two-degree of freedom polar robot with unknown perturbations. Commun. Nonlinear Sci. Numer. Simul. 2012, 17, 1021–1030. [Google Scholar] [CrossRef]
  29. Farnad, B.; Jafarian, A.; Baleanu, D. A new hybrid algorithm for continuous optimization problem. Appl. Math. Model. 2018, 55, 652–673. [Google Scholar] [CrossRef]
  30. Sharma, S.; Buddhiraju, K.M. Spatial-spectral ant colony optimization for hyperspectral image classification. Int. J. Remote Sens. 2018, 39, 2702–2717. [Google Scholar] [CrossRef]
  31. Dongwen, C. Application of Whale Optimization Algorithm in Reservoir Optimal Dispatching. Adv. Sci. Technol. Water Resour. 2017, 37, 72–76. [Google Scholar]
  32. Omran, M.G.; Salman, A.; Engelbrecht, A.P. Dynamic clustering using particle swarm optimization with application in image segmentation. Pattern Anal. Appl. 2006, 8, 332–344. [Google Scholar] [CrossRef]
  33. Kumar, V.; Chhabra, J.K.; Kumar, D. Grey wolf algorithm-based clustering technique. J. Intell. Syst. 2017, 26, 153–168. [Google Scholar] [CrossRef]
  34. Zervoudakis, K.; Tsafarakis, S. A mayfly optimization algorithm. Comput. Ind. Eng. 2020, 145, 106559. [Google Scholar] [CrossRef]
  35. Zervoudakis, K.; Tsafarakis, S. A global optimizer inspired from the survival strategies of flying foxes. Eng. Comput. 2023, 39, 1583–1616. [Google Scholar] [CrossRef]
  36. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995. [Google Scholar]
  37. Yaqoob, A.; Verma, N.K.; Aziz, R.M. Optimizing gene selection and cancer classification with hybrid sine cosine and cuckoo search algorithm. J. Med. Syst. 2024, 48, 10. [Google Scholar] [CrossRef] [PubMed]
  38. Zervoudakis, K.; Tsafarakis, S.; Paraskevi-Panagiota, S. A new hybrid firefly-genetic algorithm for the optimal product line design problem. In Learning and Intelligent Optimization: Proceedings of the 13th International Conference, LION 13, Chania, Crete, Greece, 27–31 May 2019; Revised Selected Papers 13; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  39. Rashedi, E.; Nezamabadi-Pour, H.; Saryazdi, S. GSA: A gravitational search algorithm. Inf. Sci. 2009, 179, 2232–2248. [Google Scholar] [CrossRef]
  40. Hatamlou, A.; Abdullah, S.; Othman, Z. Gravitational search algorithm with heuristic search for clustering problems. In Proceedings of the 2011 3rd Conference on Data Mining and Optimization (DMO), Putrajaya, Malaysia, 28–29 June 2011. [Google Scholar]
  41. Liu, J.; Shi, G.; Zhou, J.; Yao, Q. Prediction of college students-psychological crisis based on data mining. Mob. Inf. Syst. 2021, 2021, 9979770. [Google Scholar] [CrossRef]
  42. Reyna-Castillo, M.; Pulgarín-Rodríguez, M.A.; Ríos-Serna, A.H.; Santiago, A. PLS-SEM validation for burnout measures in latino college students: A socially sustainable educational return. Sustainability 2022, 14, 14635. [Google Scholar] [CrossRef]
  43. Moore, S.A.; Dowdy, E.; Nylund-Gibson, K.; Furlong, M.J. An Empirical Approach to Complete Mental Health Classification in Adolescents. Sch. Ment Health 2019, 11, 438–453. [Google Scholar] [CrossRef] [PubMed]
  44. Ademiluyi, O.; Ukaoha, K.; Otokiti, K. A neuro fuzzy-based guidance and counselling system for students. Afr. J. MIS 2020, 2, 39–54. [Google Scholar]
  45. Feng, X.; Wei, Y.; Pan, X.; Qiu, L.; Ma, Y. Academic emotion classification and recognition method for large-scale online learning environment-Based on A-CNN and LSTM-ATT deep learning pipeline method. Int. J. Environ. Res. Public Health 2020, 17, 1941. [Google Scholar] [CrossRef]
  46. Karunambigai, M.; Akram, M.; Sivasankar, S.; Palanivel, K. Clustering algorithm for intuitionistic fuzzy graphs. Int. J. Uncertain. Fuzziness Knowl.-Based Syst. 2017, 25, 367–383. [Google Scholar] [CrossRef]
  47. Saeb, S.; Zhang, M.; Karr, C.J.; Schueller, S.M.; Corden, M.E.; Kording, K.P.; Mohr, D.C. Mobile phone sensor correlates of depressive symptom severity in daily-life behavior: An exploratory study. J. Med. Internet Res. 2015, 17, e4273. [Google Scholar] [CrossRef] [PubMed]
  48. Iyortsuun, N.K.; Kim, S.-H.; Jhon, M.; Yang, H.-J.; Pant, S. A review of machine learning and deep learning approaches on mental health diagnosis. Healthcare 2023, 11, 285. [Google Scholar] [CrossRef]
  49. Hatcher, W.G.; Yu, W. A survey of deep learning: Platforms, applications and emerging research trends. IEEE Access 2018, 6, 24411–24432. [Google Scholar] [CrossRef]
  50. Talpur, N.; Abdulkadir, S.J.; Alhussian, H.; Hasan, M.H.; Aziz, N.; Bamhdi, A. Deep Neuro-Fuzzy System application trends, challenges, and future perspectives: A systematic survey. Artif. Intell. Rev. 2023, 56, 865–913. [Google Scholar] [CrossRef]
  51. Thiele, F.; Windebank, A.J.; Siddiqui, A.M. Motivation for using data-driven algorithms in research: A review of machine learning solutions for image analysis of micrographs in neuroscience. J. Neuropathol. Exp. Neurol. 2023, 82, 595–610. [Google Scholar] [CrossRef] [PubMed]
  52. Spurk, D.; Hirschi, A.; Wang, M.; Valero, D.; Kauffeld, S. Latent profile analysis: A review and “how to” guide of its application within vocational behavior research. J. Vocat. Behav. 2020, 120, 103445. [Google Scholar] [CrossRef]
  53. Peng, Y.; Zhu, X.; Nie, F.; Kong, W.; Ge, Y. Fuzzy graph clustering. Inf. Sci. 2021, 571, 38–49. [Google Scholar] [CrossRef]
  54. Han, H. Fuzzy clustering algorithm for university students’ psychological fitness and performance detection. Heliyon 2023, 9, e18550. [Google Scholar] [CrossRef]
  55. Fogaca, J.L. Combining mental health and performance interventions: Coping and social support for student-athletes. J. Appl. Sport Psychol. 2021, 33, 4–19. [Google Scholar] [CrossRef]
  56. Celebi, M.E. Partitional Clustering Algorithms; Springer: Berlin/Heidelberg, Germany, 2014. [Google Scholar]
  57. Murtagh, F.; Contreras, P. Algorithms for hierarchical clustering: An overview. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2012, 2, 86–97. [Google Scholar] [CrossRef]
  58. Nanda, S.J.; Panda, G. A survey on nature inspired metaheuristic algorithms for partitional clustering. Swarm Evol. Comput. 2014, 16, 1–18. [Google Scholar] [CrossRef]
  59. Reynolds, A.P.; Richards, G.; de la Iglesia, B.; Rayward-Smith, V.J. Clustering rules: A comparison of partitioning and hierarchical clustering algorithms. J. Math. Model. Algorithms 2006, 5, 475–504. [Google Scholar] [CrossRef]
  60. Hartigan, J.A.; Wong, M.A. Algorithm AS 136: A k-means clustering algorithm. J. R. Stat. Society. Ser. C (Appl. Stat.) 1979, 28, 100–108. [Google Scholar] [CrossRef]
  61. Arthur, D. K-means++: The advantages if careful seeding. In Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms, New Orleans, LA, USA, 7–9 January 2007. [Google Scholar]
  62. Jain, A.K.; Murty, M.N.; Flynn, P.J. Data clustering: A review. ACM Comput. Surv. (CSUR) 1999, 31, 264–323. [Google Scholar] [CrossRef]
  63. Davies, D.L.; Bouldin, D.W. A cluster separation measure. IEEE Trans. Pattern Anal. Mach. Intell. 1979, PAMI-1, 224–227. [Google Scholar] [CrossRef]
  64. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  65. Shaikh, M.S.; Hua, C.; Jatoi, M.A.; Ansari, M.M.; Qader, A.A. Application of grey wolf optimisation algorithm in parameter calculation of overhead transmission line system. IET Sci. Meas. Technol. 2021, 15, 218–231. [Google Scholar] [CrossRef]
  66. Zervoudakis, K.; Mastrothanasis, K.; Tsafarakis, S. Forming automatic groups of learners using particle swarm optimization for applications of differentiated instruction. Comput. Appl. Eng. Educ. 2020, 28, 282–292. [Google Scholar] [CrossRef]
  67. Sang, H.; Hong, D. Analysis and Research of Psychological Education Based on Data Mining Technology. Secur. Commun. Netw. 2021, 2021, 8979507. [Google Scholar] [CrossRef]
  68. Alzahrani, N.A.; Abdullah, M.A. Student Engagement Effectiveness In E-Learning System. Biosci. Biotechnol. Res. Commun. 2019, 12, 208–218. [Google Scholar] [CrossRef]
  69. Jawad, K.; Mahto, R.; Das, A.; Ahmed, S.U.; Aziz, R.M.; Kumar, P. Novel Cuckoo Search-Based Metaheuristic Approach for Deep Learning Prediction of Depression. Appl. Sci. 2023, 13, 5322. [Google Scholar] [CrossRef]
  70. Joshi, A.A.; Aziz, R.M. A two-phase cuckoo search based approach for gene selection and deep learning classification of cancer disease using gene expression data with a novel fitness function. Multimed. Tools Appl. 2024. [Google Scholar] [CrossRef]
  71. Amin, N.; Salehin, I.; Baten, M.A.; Noman, R.A. RHMCD-20 dataset: Identify rapid human mental health depression during quarantine life using machine learning. Data Brief 2024, 54, 110376. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Flowchart of the proposed method.
Figure 1. Flowchart of the proposed method.
Mathematics 12 01620 g001
Figure 2. Convergence characteristic curves of optimization techniques.
Figure 2. Convergence characteristic curves of optimization techniques.
Mathematics 12 01620 g002
Figure 3. Box and whisker plot of optimization techniques.
Figure 3. Box and whisker plot of optimization techniques.
Mathematics 12 01620 g003
Figure 4. Graphical representation of clusters. (a) iGWCA; (b) GWO; (c) MOA; (d) FFO; (e) PSO; (f) HSCCS; (g) FAGA; and (h) GSA.
Figure 4. Graphical representation of clusters. (a) iGWCA; (b) GWO; (c) MOA; (d) FFO; (e) PSO; (f) HSCCS; (g) FAGA; and (h) GSA.
Mathematics 12 01620 g004
Table 1. Time complexity of iGWCA.
Table 1. Time complexity of iGWCA.
Phase Generation of the Initial PhaseControl Parameters CalculationUpdate the Search Agent Position Evaluate the Fitness Value of Each Search AgentiGWCA
Time ComplexityO(N  ×  D)O(N  ×  D)O(N  ×  D)O(N  ×  D)O(N  ×  D  × Max: Iter )
Table 2. Parameter setting.
Table 2. Parameter setting.
Types Parameter
iGWCASearch agents = 30, m = 0.15, k = 2
GWOSearch agents = 30, r1, r2 = [0, 1]
MOAParticle number = 40, α1 = 1, α2 = 1.5, β = 2, d = 0.1, fl = 0.1
FFOpopulation size = 20, a1, pa = 0.92
PSOParticle number = 30; c1 = 1.5, c2 = 2; ω = 0.9
HSCCSAlpha = 0.7, A and a = 2
FAGAPopulation size = 30, mutation rate = 0.05
GSAAlpha = 20, gravitational constants = 100
Table 3. Statistical analysis of optimization techniques.
Table 3. Statistical analysis of optimization techniques.
Statistical AnalysisiGWCA [Proposed]GWO [Studied]MOA [Studied]FFO
[Studied]
PSO
[Studied]
HSCCS
[Studied]
FAGA
[Studied]
GSA
[Studied]
Best Value0.524610.532480.574560.573010.537810.544520.571500.57452
Worst Value0.584520.582310.653220.647990.609920.584430.671410.66444
Avg0.54530.54970.60170.59890.56270.58440.60600.6055
SD0.02900.02410.03800.03630.03490.01930.04830.0435
Number of Hits3132283230333033
Computational Time (s)8.16235412.454219 18.79109226.14044929.55922015.77773514.87714011.791202
% Decrease1.48%8.69%8.45%2.45%3.65%8.20%8.68%
Table 4. Performance analysis of the proposed method.
Table 4. Performance analysis of the proposed method.
S.NoMethodPerformance (%)
1Machine Learning Algorithms 0.8571 [15]
2Data Mining Technology0.90 [67]
3AdaBoost Algorithm0.8175 [18]
4Aspect-Oriented Convolutional Neural Network (A-CNN)0.89 [45]
5Machine Learning Technology (CatBoost)0.826 [14]
6Gradient Boosting Machine Algorithm (Dataset-I, Class H) 0.78 [68]
7iGWCA0.96
Table 5. Confusion matrix for different optimization techniques.
Table 5. Confusion matrix for different optimization techniques.
Types iGWCA [Proposed]GWO [Studied]MOA [Studied]FFO
[Studied]
PSO
[Studied]
HSCCS
[Studied]
FAGA
[Studied]
GSA
[Studied]
TPR0.96660.93330.80.83330.90.8660.80.7666
TNR0.350.50.650.60.40.130.23330.2666
FPR0.750.60.20.50.60.20.060.8
FNR0.206870.2142860.240.2151560.24130.26500.2130.2234
Table 6. The result of clustering for dataset-I (N = 1100, cluster size = 2).
Table 6. The result of clustering for dataset-I (N = 1100, cluster size = 2).
AlgorithmsTest Case ClusterSizeAvgMedianRankSD
iGWCA [Proposed]Test Case 118245.895.013.90
GWO [Studied]Test Case 117097.708.024.36
MOA [Studied]Test Case 117307.918.034.47
FFO [Studied]Test Case 117488.119.074.59
PSO [Studied]Test Case 117468.089.064.57
HSCCS [Studied]Test Case 117117.728.044.37
FAGA [Studied]Test Case 117307.918.054.47
GSA [Studied]Test Case 117117.728.044.37
iGWCA [Proposed]Test Case 218241.891.011.26
GWO [Studied]Test Case 217091.862.021.03
MOA [Studied]Test Case 217301.948.034.47
FFO [Studied]Test Case 217481.972.041.15
PSO [Studied]Test Case 217461.982.051.15
HSCCS [Studied]Test Case 217111.862.021.03
FAGA [Studied]Test Case 217301.942.031.11
GSA [Studied]Test Case 217111.872.031.04
iGWCA [Proposed]Test Case 318241.841.011.14
GWO [Studied]Test Case 317091.862.021.04
MOA [Studied]Test Case 317301.932.031.11
FFO [Studied]Test Case 317481.962.051.15
PSO [Studied]Test Case 317461.972.061.14
HSCCS [Studied]Test Case 317111.862.021.04
FAGA [Studied]Test Case 317301.932.041.11
GSA [Studied]Test Case 317111.872.031.05
iGWCA [Proposed]Test Case 1227620.6021.015.04
GWO [Studied]Test Case 1239121.3621.023.62
MOA [Studied]Test Case 1237021.7222.033.38
FFO [Studied]Test Case 1235222.0122.053.21
PSO [Studied]Test Case 1235421.9822.043.22
HSCCS [Studied]Test Case 1238921.3921.023.60
FAGA [Studied]Test Case 1237021.7222.033.38
GSA [Studied]Test Case 1238921.3921.023.60
iGWCA [Proposed]Test Case 222763.394.011.70
GWO [Studied]Test Case 223914.084.041.22
MOA [Studied]Test Case 223704.064.021.24
FFO [Studied]Test Case 223524.094.051.20
PSO [Studied]Test Case 223544.074.031.23
HSCCS [Studied]Test Case 223894.08441.22
FAGA [Studied]Test Case 223704.064.021.24
GSA [Studied]Test Case 223894.08441.22
iGWCA [Proposed]Test Case 322763.654.011.28
GWO [Studied]Test Case 323913.684.031.21
MOA [Studied]Test Case 323703.654.021.22
FFO [Studied]Test Case 323523.674.041.19
PSO [Studied]Test Case 323543.654.021.22
HSCCS [Studied]Test Case 323893.694.051.21
FAGA [Studied]Test Case 323703.654.021.22
GSA [Studied]Test Case 323893.684.041.22
Table 7. The result of clustering for dataset-II (N = 824, cluster size = 2).
Table 7. The result of clustering for dataset-II (N = 824, cluster size = 2).
AlgorithmsTest CaseClusterSizeAvgMedianRankSD
iGWCA [Proposed]Test Case 4122911.01514.62
GWO [Studied]Test Case 4157216.92043.67
MOA [Studied]Test Case 4119516.11534.25
FFO [Studied]Test Case 4112918.62062.25
PSO [Studied]Test Case 4169215.91523.99
HSCCS [Studied]Test Case 4113118.62062.24
FAGA [Studied]Test Case 4113518.42052.65
GSA [Studied]Test Case 4170415.91523.99
iGWCA [Proposed]Test Case 5122916.01532.06
GWO [Studied]Test Case 5157213.41512.41
MOA [Studied]Test Case 5119517.92072.91
FFO [Studied]Test Case 5112917.82062.79
PSO [Studied]Test Case 5169214.31533.98
HSCCS [Studied]Test Case 5113117.72054.02
FAGA [Studied]Test Case 5113517.62042.86
GSA [Studied]Test Case 5170414.31523.97
iGWCA [Proposed]Test Case 4259511.51515.44
GWO [Studied]Test Case 4225211.61022.35
MOA [Studied]Test Case 4262915.01564.03
FFO [Studied]Test Case 4269514.71544.07
PSO [Studied]Test Case 4213211.81052.78
HSCCS [Studied]Test Case 4269314.61534.07
FAGA [Studied]Test Case 4268914.71544.07
GSA [Studied]Test Case 4212011.51012.28
iGWCA [Proposed]Test Case 5259513.71512.56
GWO [Studied]Test Case 5225218.22052.41
MOA [Studied]Test Case 5262914.01523.88
FFO [Studied]Test Case 5269514.41534.01
PSO [Studied]Test Case 5213217.82042.91
HSCCS [Studied]Test Case 5269314.41534.02
FAGA [Studied]Test Case 5268914.41534.02
GSA [Studied]Test Case 5212018.22052.67
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shaikh, M.S.; Dong, X.; Zheng, G.; Wang, C.; Lin, Y. An Improved Expeditious Meta-Heuristic Clustering Method for Classifying Student Psychological Issues with Homogeneous Characteristics. Mathematics 2024, 12, 1620. https://doi.org/10.3390/math12111620

AMA Style

Shaikh MS, Dong X, Zheng G, Wang C, Lin Y. An Improved Expeditious Meta-Heuristic Clustering Method for Classifying Student Psychological Issues with Homogeneous Characteristics. Mathematics. 2024; 12(11):1620. https://doi.org/10.3390/math12111620

Chicago/Turabian Style

Shaikh, Muhammad Suhail, Xiaoqing Dong, Gengzhong Zheng, Chang Wang, and Yifan Lin. 2024. "An Improved Expeditious Meta-Heuristic Clustering Method for Classifying Student Psychological Issues with Homogeneous Characteristics" Mathematics 12, no. 11: 1620. https://doi.org/10.3390/math12111620

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop