Skip Content
You are currently on the new version of our website. Access the old version .
SystemsSystems
  • Article
  • Open Access

12 February 2025

Intelligent Assessment of Personal Credit Risk Based on Machine Learning

and
School of Management and Engineering, Capital University of Economics and Business, Beijing 100070, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue AI-Empowered Modeling and Simulation for Complex Systems

Abstract

In the 21st-century global economy, the rapid growth of the finance industry, particularly in personal credit, fuels economic growth and market prosperity. However, the rapid expansion of personal credit business has brought explosive growth in the amount of data, which puts forward higher requirements for the risk management of financial institutions. To solve this problem, this paper constructs an intelligent evaluation model of personal credit risk under the background of big data. Firstly, based on the forest optimization feature selection algorithm, combined with initialization based on chi-square check, adaptive global seeding, and greedy search strategies, key risk factors are accurately identified from high-dimensional data. Then, the XGBoost algorithm is used to evaluate the credit risk level of customers, and the traditional Sparrow Search Algorithm is improved by using Tent chaotic mapping, sine and cosine search, reverse learning, and Cauchy mutation strategy to improve the optimization performance of algorithm parameters. Finally, using the Lending Club dataset for empirical analysis, the experiment shows that the model improves the accuracy of personal credit risk assessment and enhances the ability of risk control.

1. Introduction

In recent years, with the rapid development of the economy and the improvement of people’s living standards, personal credit business has experienced rapid growth globally. As an important form of financial service, personal credit not only provides consumers with flexible and diverse channels for obtaining funds but also greatly promotes the prosperity of the consumer market and economic growth. However, with the continuous expansion of the personal credit market, its potential risk factors have become increasingly prominent. In particular, credit risk, as a core issue, has become one of the key factors restricting the healthy development of the market.
Credit risk refers to the risk that borrowers are unable to repay the principal and interest of loans in full and on time due to various reasons. Its origins are complex and diverse, including but not limited to changes in the economic environment, deterioration of personal financial conditions, and decreased willingness to repay. In the traditional credit evaluation system, due to limited information sources and outdated processing methods, financial institutions often find it difficult to comprehensively and accurately grasp the credit status of borrowers, leading to deviations in risk assessment results, which in turn affect the effectiveness and safety of credit decisions. In the event of large-scale defaults, it will bring huge economic losses to banks and may even lead to instability in the entire financial system. Therefore, how to effectively identify and prevent credit risks has become an important issue that financial institutions urgently need to address.
With the rapid development of big data technology and the gradual maturity of intelligent management methods, personal credit risk management is steadily moving towards intelligence and refinement. Using algorithms and big data technology to quickly process massive amounts of information can improve evaluation speed. Using credit risk assessment models to quantify default probability can provide more objective risk assessment results. Therefore, this paper aims to construct an intelligent identification and risk assessment model for personal credit risk factors in the context of big data. It precisely identifies key factors affecting personal credit risk among high-dimensional features, enhancing the comprehensiveness and intelligence of risk factor identification. Based on the specific circumstances of each applicant, such as credit history, income level, consumption behavior, etc., personalized risk assessment results are provided, improving the accuracy of risk assessment. At the same time, this research provides new approaches and methods for promoting the development of big data technology and intelligent management algorithms in the financial field.
The structure of this paper is arranged as follows: the second part introduces relevant research work, including the current application status and existing problems of personal credit risk factor identification and personal credit risk assessment methods; the third part elaborates on the construction process of intelligent identification models for personal credit risk factors and personal credit risk assessment models in the context of big data; the fourth part conducts empirical analysis, including dataset construction, data preprocessing, parameter setting, evaluation indicators, and result analysis, and compares with other commonly used models; and the fifth part summarizes the research results, points out the limitations of the research, and indicates future research directions.

3. Method

3.1. Credit Risk Factor Identification Model

3.1.1. Overview of Basic Algorithms

The feature selection using forest optimization algorithm (FSFOA) is an enhanced version of the forest optimization algorithm (FOA), which can be regarded as a forest optimization algorithm for binary discrete vectors. FOA is a bionic evolutionary algorithm proposed by Ghaemi in 2014, inspired by the growth and competition phenomena of trees in nature. In forests, towering trees tend to grow in areas with abundant water and sunlight. The process of trees’ seeds seeking the best habitat is precisely a search and optimization process. By modeling this process, Ghaemi proposed FOA to solve nonlinear optimization problems. In 2016, Ghaemi M and Feizi-Derakhshi M R adapted FOA to discrete vectors by modifying two operators, local seeding and global seeding, and applied it to feature selection, validating the feasibility of FSFOA in the field of feature selection.
FSFOA is a method that utilizes a tree-like structure to represent and select feature subsets. In this approach, each tree explicitly identifies which features are selected for the learning process through a specific 0/1 sequence—where 0 represents a feature that is not selected, and 1 signifies that the feature will be utilized. The entire FSFOA process has been meticulously designed with five core steps to ensure the effectiveness and efficiency of feature selection. The first step involves initializing the forest. During this phase, the system randomly generates N trees, collectively forming the initial forest. Each tree comprises three key components: feature selection status, age, and fitness. The feature selection status is determined by a randomly generated 0/1 sequence, while the age is initialized to 0. Subsequently, the system calculates the fitness of each tree based on the selected features, serving as the basis for subsequent selection and optimization. The second step involves local seeding, which focuses on performing seeding operations on trees with an age of 0. For each tree participating in seeding, the system will replicate LSC trees and reset the ages of these replicated trees to 0. Then, the system will randomly select a feature state to perform a negation operation, in order to explore different feature combinations. After completing these operations, the new tree will be added to the forest, and the ages of all trees in the forest will be incremented by 1. The third step involves size limitation, which aims to control the size of the forest and ensure the efficiency of the algorithm. The system achieves this by setting an age limit (life time) and an area limit. Initially, trees exceeding the age limit are moved to the candidate forest. If, even after this, the number of trees in the forest still exceeds the area limit, the system sorts the trees in descending order of fitness and retains only the area limit trees with the highest fitness in the forest, while the rest are also moved to the candidate forest. The fourth step involves global seeding to further enhance the exploration capability of FSFOA. Based on a preset transfer rate, the system randomly selects a certain proportion of trees for replication. For each replicated tree, the system randomly selects GSO feature states for negation operations, thereby generating new trees, which are then added to the forest. This step aids in exploring a wider range of feature combinations globally. The fifth step involves updating the optimal tree to ensure the continuous optimization of the algorithm. The system identifies the tree with the highest fitness within the current forest, considers it as the optimal tree, and resets its age to 0. Subsequently, this optimal tree is reintroduced into the forest to participate in the next round of local seeding and subsequent optimization processes. By iteratively performing these steps, FSFOA can gradually approach the optimal feature subset selection. Although FSFOA has achieved promising results in feature selection, it still has room for improvement.
Although FSFOA has achieved good results in feature selection, there are still the following issues: (1) When initializing the forest, FSFOA adopts a random selection strategy for feature selection, which introduces high randomness and blindness. This strategy fails to fully cover high-quality feature combinations in the feature space, thereby affecting the algorithm’s subsequent search efficiency and global optimization capabilities. Furthermore, poor initial quality of the forest trees can lead to slow forest convergence and poor adaptability to high-dimensional datasets. (2) During the local seeding stage, FSFOA only randomly selects one feature state for each tree to perform a negation operation. This strategy limits the exploration range of the search space, especially when the LSC (local seeding copy tree count) parameter is small. This limitation reduces the algorithm’s search efficiency, making it difficult to discover potential global optimal feature combinations and being prone to falling into local optima. (3) In the size limitation stage, FSFOA adds all trees with ages exceeding the upper limit to the candidate forest, and retains some trees based on fitness ranking when the forest size exceeds the regional upper limit. However, this approach has computational redundancy issues because the candidate forest may contain a large number of inferior trees with low fitness. These trees may still be selected in the subsequent global seeding stage, increasing the algorithm’s computational cost and potentially leading the algorithm to fall into local solution space. Furthermore, repeatedly using inferior trees may also reduce the algorithm’s global search capability and the quality of the final solution. (4) Experiments have shown that when only considering accuracy as the fitness metric, multiple optimal trees with the same fitness may appear. Moreover, solely using classification accuracy as the optimization objective is too simplistic and cannot effectively improve the feature dimension reduction rate of the feature subset.
Therefore, in order to enhance the overall convergence speed of the forest and strengthen its spatial search and optimization capabilities, this paper proposes a feature selection using the forest optimization algorithm based on hybrid strategies (HSFSFOA).

3.1.2. Initialization Strategy Based on Chi-Squared Test

When initializing the forest, the FSFOA generates the initial solution space in a random manner, resulting in some trees of lower quality. These low-quality trees may fail to discover high-quality populations during subsequent search processes, affecting the quality of the final results. To address this issue, the HSFSFOA proposes a new initialization strategy that selects high-quality features based on the relationship between features and labels, thereby improving the overall quality of the initial solution space.
The HSFSFOA selects features with strong correlation to the label during the initialization stage. This is because such features contribute significantly to classification tasks, while features with weaker correlation to the label are not selected into the optimal feature subset. To achieve this, this paper proposes an initialization method based on the chi-squared test.
The chi-square test is a statistical testing method used to evaluate whether two events are independent or whether there is a significant difference between the actual observed values and the expected values. Its calculation formula is as follows:
x 2 = i = 1 k A i E i 2 / E i
Here, k represents the number of observations, A i denotes the i-th observation, and E i signifies the corresponding theoretical value. The larger the value of x 2 , the greater the deviation between the actual observation and the expected value, indicating a closer correlation between the feature and the prediction target.
The initialization strategy based on the chi-squared test is executed in two steps. Firstly, calculate the chi-squared values for all features and labels in the dataset, and arrange the features in descending order based on the obtained chi-squared values. Then, during the initialization process of each tree, select n features from the sorted feature list as the feature set used by the current tree. Here, n is a randomly generated non-zero integer less than or equal to the total number of features. This method not only ensures that highly relevant features to the labels are selected during initialization, but also promotes exploration diversity by introducing a certain degree of randomness. This helps avoid falling into local optima, thereby improving the overall algorithm performance.

3.1.3. Adaptive Global Seeding Strategy

During the global seeding stage, the GSC (global search control) parameter controls the selection of features, thereby influencing the adjustment of tree morphology. Fixed GSC parameters limit the algorithm’s flexibility in adjusting tree morphology, making it difficult to achieve effective global search and escape from local optima. This paper proposes an improved GSC parameter generation strategy, aiming to fully utilize the global search mechanism to avoid falling into local optima and more effectively find the optimal feature subset.
There are differences in the quality of trees in the candidate forest: some trees perform poorly due to their feature sets being far from the optimal feature subset; others perform better because they have feature sets close to the optimal solution. For the former, more adjustments need to be made to their feature sets; for the latter, only minor adjustments are required.
Inspired by automatic control theory, this paper introduces a method based on a feedback adjustment mechanism to adaptively generate GSC parameter values. This method dynamically sets the GSC value based on the fitness value of each tree (i.e., the classification accuracy of the feature subset determined by the current tree). Specifically, inferior trees will be assigned larger GSC values to allow for more adjustments in feature selection, while superior trees will be assigned smaller GSC values to maintain the stability of their feature sets. Additionally, to distinguish between global and local seeding strategies, upper and lower limits for the GSC parameter are set. In the adaptive global seeding strategy, the GSC parameter value is a function of the tree’s fitness value.
mum G S C = f (   Fitness   )
When evaluating the quality of a tree, not only the performance of the current tree is considered, but also the performance of the ideal optimal tree is referenced to ensure the accuracy of the evaluation. When assessing the quality of a tree, it is necessary to consider not only the quality of the current tree but also the quality of the ideal tree, in order to make an appropriate judgment on the tree’s quality. This article compares the current tree quality F i t n e s s n o w with the ideal optimal tree quality F i t n e s s b e s t to determine the quality of the current tree. When determining the quality of the tree, if there is a large gap between F i t n e s s n o w and F i t n e s s b e s t , a conservative evaluation of the tree quality is made; if the gap is small, a more aggressive evaluation of the tree is made. Since the fitness value of the ideal optimal tree is difficult to determine, this article uses the current optimal tree to approximate the ideal optimal tree. The tree quality function is as follows:
  Fitness   = 1 1 + e Fithess   best     Fithess   now   1 2
After obtaining the quality of the current tree, it needs to be mapped onto the actual number of features that need to be modified. Assuming that the quality of the current tree is directly proportional to the number of features that need to be modified,
  num   change     Fitness  
When finalizing the GSC parameter value, calculate the number of features that should be modified based on the total number of features in the dataset. To ensure that the global seeding process differs from local seeding, set the upper limit of the GSC parameter to 0.5n (where n is the total number of features) and the lower limit to 2. Therefore, the formula for determining the GSC parameter value is as follows:
  num   G S C = 1 + 0.5 n ×   Fitness
By adaptively adjusting the GSC value, it can effectively expand the search range of global seeding, helping the algorithm avoid falling into local optima. At the same time, it can more easily find a better search space and promote convergence towards the optimal solution.

3.1.4. Greedy Search Strategy

During the local seeding process of the FSFOA, all trees with a tree age of “0” undergo local search, but this approach limits the overall optimization performance of the algorithm. This paper proposes an improvement to this process, aiming to fully leverage the effectiveness of local search and more efficiently identify the optimal feature subset.
During the local seeding stage, trees with an age of “0” exhibit variations in their morphology. The characteristics of inferior trees differ significantly from those of ideal optimal trees, making it difficult to transform them into high-quality trees by altering individual features. However, the feature subset of high-quality trees is relatively close to the ideal optimal feature subset. Therefore, when conducting multiple local searches, there is a greater likelihood of finding a better feature subset. Based on this, this paper adjusts the local seeding strategy by increasing the selection restrictions on local search trees and expanding the local search range.
Specifically, when selecting a local search tree, we first choose all trees with a tree age of “0” from the forest, and then further filter them based on their quality. Inspired by the greedy strategy, this paper prioritizes selecting trees with higher fitness values for subsequent local search, thereby reducing the search for inferior trees. If only the tree with the highest fitness value is selected, although it can improve the quality of local search, it can lead to a lack of diversity in the forest, potentially causing the algorithm to fall into local optima. Therefore, this paper selects half of the trees in the forest as the objects of local search to maintain the diversity of the forest. For the judgment of tree quality, trees with a tree age of “0” are sorted from high to low according to their fitness values, with the first half marked as high-quality trees for subsequent local search, while the second half are considered inferior trees and do not participate in subsequent local search.
In the FSFOA, the range of local search is controlled by the LSC (local search control) parameter, which is typically set as follows:
L S C = n 5
To enhance the effectiveness of local search, set the LSC value to twice its original value:
L S C = 2 × n 5
This not only enables the search for better feature subsets but also ensures the algorithm’s operational efficiency. The local seeding process is essentially a “selecting the best from the good” local search process. With the aforementioned adjustments, the new local seeding strategy fully leverages the “selecting the best from the good” characteristic of local search, eliminates ineffective searches for inferior trees, expands the search range for high-quality trees, and thereby facilitates convergence to the globally optimal or near-optimal feature subset.
The entire flowchart of the risk factor identification algorithm is shown in Figure 1.
Figure 1. Flowchart of risk factor identification algorithm.

3.2. Credit Risk Assessment Model

3.2.1. Overview of Basic Algorithms

The Sparrow Search Algorithm (SSA), proposed in 2020, is a novel swarm intelligence optimization algorithm that demonstrates strong capabilities in finding optimal solutions. However, compared to most intelligent optimization algorithms, SSA tends to get stuck in local optima during the later stages of iteration, and still faces the issue of low convergence accuracy.
To enhance the global optimization capability of the Sparrow Search Algorithm (SSA), researchers have proposed various improvement strategies. Zhou et al. introduced the TC chaotic mapping to initialize the sparrow population, resulting in a more uniform and diverse population distribution [39]. Liu et al. refined the discoverer position update rule in the SSA by utilizing the Golden Sinusoidal Algorithm, addressing the issue of reduced search dimensions [40]. Xiao et al. introduced a hybrid sine–cosine algorithm to update the discoverer position and incorporated the Levy flight strategy to update the follower positions, thereby improving the SSA and enhancing the scientific nature of parameter selection [41]. Wang et al. proposed a sand cat swarm optimization algorithm that integrates the SSA with the Cauchy mutation strategy [42]. By integrating the search mechanism strategy of the SSA, the algorithm’s search capability is enhanced in the later stages. By adopting the Cauchy mutation strategy, the algorithm changes the position of individuals, improving population diversity and enabling the algorithm to escape local optima. Wang et al. introduced Cauchy mutation and variable spiral search strategies into the follower position update to enhance the algorithm’s search efficiency and global search performance [43]. Xue et al. proposed a multi-strategy improved SSA to address issues such as decreased population diversity and susceptibility to local optima when the algorithm approaches the global optimum [44].
Based on the current research status, this paper improves the traditional Sparrow Search Algorithm (SSA) and proposes an improved Sparrow Search Algorithm to optimize the XGBoost model for credit risk prediction. The improved Sparrow Search Algorithm uses Tent chaotic mapping for initialization; utilizes the sine–cosine search strategy to effectively balance global and local search capabilities; introduces the reverse learning strategy and Cauchy mutation for perturbation to expand the search domain and improve the ability to escape local optima; and adopts greedy rules to determine the optimal solution. By optimizing the XGBoost parameters using the improved Sparrow Search Algorithm, the accuracy of the model in identifying credit risks is improved.

3.2.2. XGBoost Algorithm

XGBoost (extreme gradient boosting) belongs to the Boosting technique in ensemble learning methods. The basic idea of Boosting is to construct a strong learner by combining multiple weak learners. In XGBoost, these weak learners are usually decision trees. Each tree is trained to correct the errors of the previous tree. Each new model is trained based on the difference (residual) between the prediction result of the previous model and the true value, in order to reduce the prediction error of the overall model.
(1) Define the objective function
The core of XGBoost lies in the design of its objective function, which is used to evaluate the quality of the model and guide the learning process of model parameters. The objective function typically consists of two parts: the training loss (loss function) and the regularization term. The training loss measures the difference between the model’s predicted values and the actual values; the regularization term is used to control model complexity and prevent overfitting. For the t-th iteration, the objective function can be expressed as
obj ( t ) = i = 1 n L y i , y ^ i ( t ) + k = 1 t Ω f k
where L y i , y ^ i ( t ) is the loss function, Ω f k is the complexity of the k-th tree, n is the number of samples, y i is the true label of the i-th sample, and y ^ i ( t ) is the predicted value of the i-th sample by the model after the t-th iteration.
(2) Taylor’s second-order expansion
To optimize the objective function, XGBoost performs a second-order Taylor expansion of the loss function. In the t-th iteration, y ^ i ( t ) is represented as y ^ i ( t 1 ) + f t x i , where f t x i is the predicted value of the t-th tree. Then, the loss function is expanded using Taylor’s formula:
L y i , y ^ i ( t ) L y i , y ^ i ( t 1 ) + g i f t x i + 1 2 h i f t 2 x i
where g i = L y i , y ^ i ( t 1 ) y ^ i ( t 1 ) is the first derivative and h i = 2 L y i , y ^ i ( t 1 ) y ^ i ( t 1 ) 2 is the second derivative.
(3) Simplify the objective function
Substituting the Taylor series expansion into the objective function and removing the constant terms unrelated to f t , we obtain
obj ( t ) i = 1 n g i f t x i + 1 2 h i f t 2 x i + Ω f t
(4) Define the complexity of a tree
In XGBoost, the complexity of a tree is composed of the number of leaf nodes and the L2 norm of the leaf node weights:
Ω f t = γ T + 1 2 λ j = 1 T w j 2
where T represents the number of leaf nodes, w j denotes the weight of the j-th leaf node, and γ and λ are regularization parameters.
(5) Optimize the objective function
Substituting the complexity of the tree into the objective function, we obtain
obj ( t ) i = 1 n g i f t x i + 1 2 h i f t 2 x i + γ T + 1 2 λ j = 1 T w j 2
Since each sample will eventually fall into a leaf node, the objective function can be rewritten as a function of the weights of the leaf nodes. Let I j be the set of sample indexes contained in the j-th leaf node. Then, the objective function can be expressed as
obj ( t ) = j = 1 T i I j g i w j + 1 2 i I j h i + λ w j 2 + γ T
To minimize the objective function, we take the derivative of w j and set it to 0, thereby obtaining the optimal weight of the leaf node:
w j * = i I j g i i I j h i + λ
Substituting w j * into the objective function, we obtain the optimal objective function value (also known as the gain):
  Gain   = 1 2 j = 1 T i I j g i 2 i I j h i + λ + γ T
(6) Construct the optimal tree
In XGBoost, the optimal tree is constructed through a greedy algorithm. The specific approach is as follows: starting from the root node, traverse all possible split points for all features, calculate the gain after splitting, and select the split point with the largest gain for splitting. Then, repeat this process for the split child nodes until reaching a preset stopping condition (such as the maximum depth of the tree, the number of samples in the node is below a certain threshold, the gain is less than a certain threshold, etc.).
(7) Model updating and iteration
After completing the construction of a tree, update the model’s predicted values:
y ^ i ( t ) = y ^ i ( t 1 ) + f t x i
Then, proceed to the next iteration and repeat the aforementioned process until the preset number of iterations is reached or the model’s performance no longer shows significant improvement.
XGBoost exhibits high accuracy in classification and regression problems. In this paper, we utilize XGBoost to construct a model for credit risk prediction and employ an enhanced sparrow algorithm to optimize its parameters.

3.2.3. Sparrow Search Algorithm

The Sparrow Search Algorithm is a swarm intelligence optimization algorithm inspired by the foraging behavior of sparrows. In SSA, the sparrow population is divided into three roles: discoverers, joiners, and awareness scouts. These roles exhibit different behavioral patterns during the search process and collaborate to find the optimal solution to the problem.
During the population initialization process, a certain number of sparrow individuals are randomly generated, and each individual is assigned an initial position. These positions are randomly selected within the search space. Parameters such as the maximum iteration count, population size, proportion of discoverers, proportion of vigilantes, vigilance value ( R 2 ), and safety threshold (ST) are set.
Then, role division is carried out. Based on the fitness value, a portion of sparrows are selected as discoverers, responsible for exploring new search areas. The remaining sparrows act as joiners, following the discoverers to forage. In each iteration, a random subset of sparrows are chosen as guardians, tasked with monitoring the surrounding environment and avoiding local optima.
The discoverer determines the search direction and range based on its energy reserve (fitness value). Discoverers with high energy explore a wider area, while those with low energy search near their current location. The formula for updating the discoverer’s position is as follows:
X i , j k + 1 = X i , j k exp i α   iter   max , R 2 < S T X i , j k + Q L , R 2 S T
where X i , j k represents the position of the i-th sparrow in the j-th dimension at the k-th iteration; α is a random number belonging to (0, 1];   iter   max denotes the maximum iteration count; R 2 is the warning value, a random number belonging to [0, 1]; ST is the safety threshold, with a value range of [0.5, 1]; Q is a random number satisfying a normal distribution; and L is an l × d matrix with an average element of 1.
The participant updates their location based on the discoverer’s location and their own location, using the following formula:
X i , j k + 1 = Q exp X worst   k X i , j k i 2 , i > N / 2 X best   k + X i , j k X best   k A + L , i N / 2
Among them, X worst   k represents the current global worst position; X best k represents the current global best position; and A + is a matrix calculated by A T A A T 1 , where A is an l × d matrix with randomly assigned elements of 1 or −1.
The guard decides whether to flee the current location based on the current environment (i.e., the fitness value). The update formula is as follows:
X i , j k + 1 = X best   k + β X i , j k X best   k , f i > f best   X i , j k + K X i , j k X wost   k f i f worst   k + ε , f i = f best  
where X best k represents the current global optimal position; β is the step size control parameter, satisfying a normal distribution with mean 0 and variance 1; K is a random number with a value range of [−1, 1]; f i is the current fitness value of the sparrow; f b e s t is the current global optimal fitness value; f w o r s t is the current global worst fitness value; and ε is a minimum constant to prevent the denominator from being zero.

3.2.4. Improved Sparrow Search Algorithm

The traditional Sparrow Search Algorithm (SSA) tends to fall into local optima and exhibits low convergence accuracy in the later stages of iteration. To enhance the global optimization capabilities of SSA, researchers have proposed various improvement strategies. This paper combines the Tent chaotic mapping, sine–cosine search strategy, reverse learning strategy, and Cauchy mutation to improve the traditional Sparrow Search Algorithm. The aim is to balance global and local search capabilities, avoid falling into local optima, and thereby enhance the algorithm’s optimization performance.
During the optimization process of the Sparrow Search Algorithm (SSA), randomly initializing the population can lead to a large number of sparrow individuals clustering together, resulting in uneven distribution. To enhance the diversity of the sparrow population and improve the algorithm’s global search capability, this paper introduces the Tent chaotic mapping function into the SSA. The sine–cosine search strategy is employed to update the position of the discoverer. By combining the periodicity and fluctuation of the sine and cosine functions, it strikes a good balance between global and local search, effectively improving the algorithm’s exploration capability and convergence speed. This strategy effectively balances global and local search capabilities, enhancing the algorithm’s optimization efficiency. The reverse learning strategy expands the search domain by generating the reverse solution of the current optimal solution. In each iteration, the current optimal solution is replaced with its reverse solution with a certain probability, helping the algorithm escape local optima and further explore the solution space. Cauchy mutation perturbs the current optimal solution to generate new candidate solutions. The Cauchy distribution has a long tail, allowing for a wide range of perturbation values, which facilitates a broader search within the solution space, further enhancing the algorithm’s global search capability.
Firstly, the positions of the initial population, including discoverers, joiners, and vigilantes, are generated using the Tent map. Parameters such as population size, maximum iteration count, proportion of discoverers, proportion of vigilantes, vigilance value, and safety threshold are set. The Tent map is used to generate the positions of the initial population, ensuring diversity and uniformity in population distribution. The formula for the Tent map is as follows:
X n + 1 = u × X n , 0 X n < 0.5 u × 1 X n , 0.5 X n 1
where X n is the value of the n-th iteration, and u is a constant between 0 and 1, usually set to 0.5 to ensure the uniformity and chaotic characteristics of the mapping.
Then, generate the initial population. Choose an initial value X 0 , iterate through the Tent mapping formula to generate N random numbers, and linearly map these random numbers between the upper and lower bounds of the search space to generate the positions of the initial population. Assuming that the upper and lower bounds of the search space are lb and ub, the initial position X i is calculated as follows:
X i = lb + X n × ( ub lb )
Next, calculate the fitness value of each sparrow. Sort the positions of the sparrows based on their fitness values, and determine the individuals with the current optimal and worst fitness.
Update the position of the discoverer using the sine–cosine search strategy. The position update formula for the sine–cosine search strategy is as follows:
X i k + 1 = X i k + A 1 × cos B 1 × X best   k X i k + A 2 × sin B 2 × X best   k X i k
where X i k represents the position of the i-th explorer in the k-th iteration; X b e s t k represents the position of the i-th explorer in the k-th iteration; A 1 and A 2 are two random numbers, each drawn from the interval [0, 1]; and B 1 and B 2 are angle values randomly selected from the interval [0, 2π].
The positions of the joiner and the guarder are updated according to the traditional SSA formula.
To enable individuals to expand their search domain and find the optimal solution, a reverse learning strategy is introduced into the algorithm. Additionally, Cauchy mutation is introduced to perturb and update the optimal solution position, improving the ability to avoid local optima and enhance the algorithm’s ability to obtain the global optimal solution. To enhance the optimization performance of the algorithm, this paper adopts a dynamic selection strategy, alternating the execution of the reverse learning strategy and Cauchy mutation perturbation strategy at a certain probability. The selection probability q determines whether to use the reverse learning strategy or Cauchy mutation to perturb the current optimal solution. The calculation formula for the selection probability q is
q = v 1 v 2 ×   iter   max k   iter   max
Among them, v 1 is set to 0.5, and v 2 is set to 0.1.
When the randomly generated number is less than the selection probability q, the reverse learning strategy is used for position updating, with the formula being
X best   ( k ) = ub + r × lb X best   ( k )
Among them, X best   ( k ) represents the inverse solution of the optimal solution in the k-th generation; ub and lb denote the upper and lower bounds; r is an l × d random number matrix following a uniform distribution between (0, 1); and r is the information exchange control parameter, with the formula being
r =   iter   max k   iter   max k
where i t e r max represents the maximum iteration count, and k denotes the iteration count.
Otherwise, perform Cauchy mutation perturbation, with the expression as follows:
X i , j k + 1 = X best   k + cauchy ( 0 , 1 ) × X best   k
Among them, c a u c h y 0 , 1 represents the standard Cauchy distribution function.
Use a greedy selection mechanism to update the optimal solution; that is, compare the fitness values of the old and new optimal solutions, and select the one with a better fitness value as the new optimal solution. The update rule is as follows:
X best   = X best   , f X i , j t + 1 f X best   X i , j t + 1 , f X i , j t + 1 < f X best  
Finally, determine whether the maximum iteration count has been reached. If the maximum iteration count is reached, output the global optimal solution; otherwise, return to the previous step to continue iterating.
Through the aforementioned improvement strategies, the algorithm achieves a better balance between global and local search capabilities, effectively avoiding the trap of local optima, thereby enhancing the algorithm’s optimization performance. When optimizing the parameters of the XGBoost model, it can search for more optimal parameter combinations, further improving the accuracy of credit risk prediction.

3.2.5. XGBoost Model Combined with Improved Sparrow Search Algorithm

The performance and computational cost of the XGBoost model are influenced by multiple parameters. Among them, setting the learning rate too low can slow down the model’s operation, while setting it too high may affect the model’s prediction accuracy. The maximum depth of the tree (max_depth) serves as a tuning parameter to control overfitting in the model. Optimizing the number of iterations (n_estimators) ensures that the model achieves the best iterative effect. The parameter gamma determines the minimum decrease in the loss function required for node splitting, enabling reasonable node splitting.
To enhance model performance, this algorithm adjusts these four key parameters. Firstly, the improved Sparrow Search Algorithm is utilized to set the initial parameters of XGBoost, and the position parameters of the sparrows are passed to the model. Subsequently, the fitness value of each individual is calculated, and sorted according to fitness; the position of the sparrows is updated; and the global optimal solution is recorded. When the predetermined number of iterations is reached, the algorithm ends the traversal and outputs the optimal solution, namely the position of the optimal sparrow, which is passed to the XGBoost model as the optimal parameters. Finally, these optimal parameters are used to retrain the XGBoost prediction model. The entire algorithm flow is shown in Figure 2.
Figure 2. Risk assessment algorithm flowchart.

4. Experiments

4.1. Identification of Credit Risk Factors

4.1.1. Dataset

Lending Club is one of the largest platforms in the United States in terms of total loan volume. The dataset used in the empirical analysis section of this article is a publicly available real-user loan dataset from Lending Club. The dataset is 421 MB in size and contains 887,379 loan records issued from 2007 to 2015. Each loan record has 74 dimensional features, and the Excel file downloaded along with the dataset contains detailed descriptions of the meanings of each field. The data download source is the Kaggle data competition platform.

4.1.2. Data Preprocessing

This dataset encompasses user loan information, economic data, and personal basic information. This section involves data processing, which includes handling missing values, removing irrelevant variables, processing continuous features, handling discrete features, processing time-type features, and handling label features.
(1) Handling missing values
There are a large number of missing values in this dataset. This article deals with variables with missing values in two steps. Firstly, to avoid variables with too many missing values affecting model performance, variables with more than 50% missing values are eliminated, resulting in the deletion of 21 variables in total. Then, for continuous features, missing values are filled using the median. For discrete features, missing values are filled using the mode.
(2) Remove irrelevant variables
In the data preprocessing stage, this paper removed variables that were significantly less correlated with the research content, such as loan ID, member ID, URL, zip code, employee title, zip code, address state, payment plan, title, etc. These variables are either unique identifiers or have significantly lower associations with credit risk, and may even introduce noise or increase the computational complexity of the model. For example, although the zip code may indirectly reflect certain hidden variables such as income level or education level, this paper chooses to directly use individual-level economic and social characteristics to avoid potential biases introduced by proxy variables. By removing variables with significantly lower correlations, the efficiency and interpretability of the model can be improved.
(3) Processing continuous features
For continuous feature values, this paper adopts normalization processing, using the interval scaling method to ensure that the input features have the same magnitude, thereby enabling subsequent algorithms to learn effectively.
(4) Handling discrete features
Due to the presence of text categorical variables such as gender and loan purpose in this dataset, it is necessary to digitize their features by mapping discrete categorical variables to numerical variables. The mapping relationship is shown in Table 1:
Table 1. Map discrete variables to numerical variables.
(5) Processing of time-type features
Use the to_datetime function in the pandas package to convert the features ‘issue_d’, ‘earliest_cr_line’, ‘last_pymnt_d’, ‘next_pymnt_d’, and ‘last_credit_pull_d’ into datetime types.
(6) Label feature processing
The field used to define good and bad users as labels in the original feature data is loan_status, which contains seven possible values representing seven different loan statuses. For specific explanations, please refer to Table 2.
Table 2. Explanation of loan status fields in Lending Club dataset.
In this paper, we define users with loan statuses of ‘Current’ and ‘Fully Paid’ as good users, labeled as 0. Users with loan statuses of ‘In Grace Period’, ‘Late (31–120 days)’, ‘Late (16–30 days)’, ‘Charged Off’, and ‘Default’ are defined as bad users, labeled as 1.

4.1.3. Experimental Environment and Parameter Settings

In terms of the experimental environment, the experimental platform in this paper is a computer with an Intel i9-9880H CPU, 32 GB of memory, and Windows 10 as the operating system. The code runs in Python 3.8, compiled using Jupyter Notebook, and utilizes the scikit-learn machine learning library.
The intelligent recognition algorithm for personal credit risk factors in this article involves parameters such as the upper limit of tree age (life time), the upper limit of forest size (area limit), the global seeding rate (transfer rate), local search control (LSC), and global search control (GSC). Among them, the three parameters of life time, area limit, and transfer rate are used to control the search range and iteration cycle, and are less affected by the dataset. Therefore, the value of life time is fixed at 15, the value of the area limit is fixed at 50, and the value of the transfer rate is fixed at 5%. The remaining parameter setting methods are slightly adjusted, and the parameter values are reset based on the methods in this section.

4.1.4. Evaluation Indicators

To comprehensively evaluate the performance of the improved forest optimization feature selection algorithm, this paper employs three primary evaluation metrics: classification accuracy (CA), dimensionality reduction (DR), and AUC value (Area Under the Curve). Among them, classification accuracy and AUC value are used to assess the algorithm’s classification capabilities, while dimensionality reduction is used to evaluate its dimensionality reduction capabilities.
(1) Classification accuracy: Classification accuracy is a key metric for evaluating the performance of the optimal feature subset on a classifier. It reflects the accuracy of the feature subset selected by the algorithm for classification tasks. The specific definition of classification accuracy is as follows:
C A = CC AC
Among them, CC (Correct Classification) represents the number of instances that can be correctly classified, while AC (All Classification) represents the total number of instances involved in classification. A higher classification accuracy indicates that the feature subset selected by the algorithm performs better in the classification task and can more accurately identify instances of different categories.
(2) Dimensionality reduction rate: The dimensionality reduction rate refers to the proportion of reduction in the number of original features during feature selection or dimensionality reduction. It reflects the algorithm’s ability to effectively reduce the number of features while maintaining classification performance. The specific definition of the dimensionality reduction rate is as follows:
D R = 1 SF AF
Among them, SF (selected features) represents the number of selected features in the feature set, while AF (all features) denotes the total number of features in the dataset, which is n. A higher dimensionality reduction rate indicates that the algorithm can select fewer features while maintaining high classification accuracy, thereby achieving more effective feature dimensionality reduction.
(3) AUC value (Area Under Curve): The True Positive Rate (TPR) represents the proportion of true positives that are correctly predicted as positive. The False Positive Rate (FPR) represents the proportion of false positives that are incorrectly predicted as positive. The calculation formula is as follows:
T P R = T P T P + F N
F P R = F P F P + T N = 1 T P R
The ROC curve can be plotted with FPR as the horizontal axis and TPR as the vertical axis of the coordinate system. The area enclosed by the ROC curve and the coordinate axes is the AUC value. The AUC value represents the strength of the classification performance of the current machine learning algorithm. A higher AUC value indicates that the algorithm’s classification performance is more stable at different thresholds, and its overall classification ability is stronger.
(4) Comprehensive weighted evaluation: In practical business applications, people will comprehensively weigh the benefits brought by classification ability and dimensionality reduction ability. Therefore, in order to unify the evaluation criteria and further verify the effectiveness of the proposed algorithm, this paper adopts comprehensive weighted evaluation to evaluate the effectiveness of the algorithm in practical applications. The definition of comprehensive weighted evaluation is as follows:
  Fitness   = C A + 0.01 × D R

4.1.5. Result Analysis

By utilizing an intelligent identification model for personal credit risk factors, this paper has screened out 19 key features from the original dataset. These features play a crucial role in enhancing the predictive power of the model. For example, if there is a longer employment period (emp_ length), it may mean that the borrower has a more stable source of income to repay the debt. The feature of the last payment amount (last_pymnt_mamnt) reveals the total amount of the last payment received, which helps analyze the borrower’s repayment behavior and habits, and understand whether they have repaid on time and in full, and is crucial for evaluating their credit status. The debt-to-income ratio (DTI) is calculated by dividing the borrower’s monthly total debt payments by their self-reported monthly income. A high DTI indicates that the borrower has a heavy debt burden, which may increase their risk of default. Considering these key factors can provide financial practitioners with a more comprehensive perspective to evaluate the overall credit risk of borrowers and enhance their understanding of borrowers’ debt repayment ability. Compared to the full feature set, the filtered feature set retains only the most predictive features, significantly reducing the dimensionality of the data (with a dimensionality reduction rate of approximately 57.78%). This not only simplifies the model’s complexity but also enhances computational efficiency.
Using key feature data screened out by the intelligent identification model for personal credit risk factors, combined with the XGBoost model for training and testing, this paper obtained the following results: the model’s accuracy reached 97.71%, and the AUC value was 0.93. These data indicate that the model exhibits excellent performance in identifying credit risks, accurately distinguishing between loan default and non-default situations. Furthermore, the high AUC value suggests that the model possesses strong discriminative ability, maintaining good predictive stability even under different threshold settings.
The fitness value, calculated according to the formula, is 0.983. This result indicates that the proposed algorithm exhibits excellent comprehensive performance in practical applications, maintaining high classification capability while achieving certain effectiveness in dimensionality reduction. A fitness value close to 1 signifies that the algorithm has reached a good level in balancing classification accuracy and dimensionality reduction effect, further verifying the effectiveness and practicality of the proposed algorithm.
Finally, by plotting the ROC curve, we visually demonstrate the performance of the model at different thresholds, as shown in Figure 3. The ROC curve closely approaches the upper left corner, and the AUC value is as high as 0.93, proving the superior performance of the model in credit risk identification tasks and the necessity of using an intelligent identification model for personal credit risk factors for feature selection. This result not only provides strong support for credit institutions in identifying personal credit risk factors but also lays a solid foundation for our subsequent further evaluation of personal credit risk.
Figure 3. ROC curve.
In addition, this paper selects Recursive Feature Elimination (RFE), random forest, PCA, and Isomap as comparative algorithms for comparative experiments with the HSFSFOA model. The key feature data selected by each feature selection model are utilized, and combined with the XGBoost model for training and testing. The experiments compare the classification accuracy and AUC values of HSFSFOA with those of four other feature selection algorithms on the Lending Club dataset, highlighting the effectiveness of feature selection methods. Detailed experimental data are shown in Table 3.
Table 3. Comparison of evaluation indexes of each model.
According to the evaluation metrics for each model listed in the table, the HSFSFOA model demonstrates strong performance in both classification accuracy and AUC value. The classification accuracy of the HSFSFOA model reaches as high as 97.71%, surpassing other comparative algorithms, highlighting the model’s robust ability to identify key features in datasets and perform accurate classification based on them. Simultaneously, its AUC value attains 0.93, further validating the model’s high robustness and generalization capability when dealing with complex datasets. This establishes a solid foundation for its widespread application in similar datasets and feature selection tasks.

4.2. Credit Risk Assessment

4.2.1. Dataset

This paper establishes an intelligent identification model for personal credit risk factors in the context of big data, conducts in-depth analysis and screening of 74 dimensional features, and identifies a series of features that have a significant impact on credit risk prediction. Below, we will use these screened important feature data to train an intelligent evaluation model for personal credit risk in the context of big data, and predict and evaluate personal credit risk.

4.2.2. Parameter Settings

The parameter settings for the personal credit risk assessment model are as follows: set the population size (pop_size) to 30; set the maximum iteration count (max_iter) to 100; set the discoverer ratio (discoverer_ratio) to 0.2; set the watcher ratio (watcher_ratio) to 0.1; set the safety threshold (ST) to 0.8; and set the weights (w1, w2) to 0.5 and 0.1. The parameters of the XGBoost model are optimized through an improved Sparrow Search Algorithm, including the learning rate (learning_rate), maximum depth (max_depth), number of estimators (n_estimators), and gamma value (gamma).

4.2.3. Evaluation Indicators

In addition to the evaluation metrics mentioned above, this section will also evaluate the model using precision and recall. In the case of an imbalanced number of sample categories, it is unreasonable to solely use accuracy as a measurement metric. Suppose in a credit dataset, the number of credit defaulters accounts for 1% of the total number of credit borrowers. During the prediction process, if the algorithm simply predicts all predicted data as non-defaulting, its credit risk prediction accuracy will be as high as 99%, which is obviously unreasonable. Therefore, recall and precision are introduced to evaluate the performance of the algorithm. Recall represents the proportion of samples correctly predicted as positive to the total number of positive samples, while precision represents the proportion of samples correctly predicted as positive to the total number of samples predicted as positive. The closer the values of recall and precision are to 1, the better the classification performance of the machine learning algorithm. The calculation formula is as follows:
  Recall   = T P T P + F N
  Precision   = T P T P + F P

4.2.4. Result Analysis

In order to improve the reliability of model evaluation, we adopted a repeated K-fold cross-validation method. Specifically, in the updated experiment, we divided the dataset into five subsets using RepeatedKFold and performed three repetitions on this basis to ensure the stability of the evaluation. At each iteration, one subset of the dataset is used as the test set, while the remaining four subsets are used to train the model. Through this approach, each sample has the opportunity to be a part of the test data, making the evaluation more comprehensive and accurate. After completing all segmentation and repetition, we calculated and output the average accuracy and average AUC values for all iterations to comprehensively evaluate the overall performance of the model. Generally, the AUC value ranges from 0.5 to 1. When the AUC value exceeds 0.75, it indicates that the model is acceptable. AUC values closer to 1.0 signify higher fitting performance of the model. The ROC curve is illustrated in Figure 4. From the figure, it is evident that the AUC of the XGBoost model combined with the improved Sparrow Search Algorithm is 0.923, which is very close to 1. This demonstrates that the model exhibits good performance in assessing credit risk, implying its significant ability to interpret customer performance status. The contributions of variables are fully considered, and the model possesses strong predictive capabilities.
Figure 4. ROC curve.
To comprehensively and deeply evaluate the performance of the XGBoost model combined with an improved Sparrow Search Algorithm in personal credit risk assessment, this paper selects four commonly used machine learning classification algorithms as benchmarks for comparison, namely logistic regression (Logistic Regression), random forest (Random Forest), support vector machine (Support Vector Machine, SVM), K-nearest neighbors (K-Nearest Neighbors, KNNs), Light Gradient Boosting Machine (LightGBM), and Categorical Boosting (CatBoost). The resulting model results are shown in Table 4.
Table 4. Comparison of evaluation indexes of each model.
The XGBoost model combined with the improved Sparrow Search Algorithm achieves an accuracy rate of 0.953, indicating that the model has high accuracy in distinguishing between high and low credit risks, and can provide precise risk assessments for most credit applications, thereby helping financial institutions effectively control credit risks and improve decision-making efficiency. In terms of precision, the model reaches 0.964, meaning that when the model predicts that a credit application is high-risk, the probability that the application is actually risky is extremely high, and it is almost certain to be a genuine risk case. Such high precision not only reduces potential losses caused by misjudgments but also enhances the trust of financial institutions in the model’s prediction results. However, the recall rate of the model is 0.685, which is also at a relatively high level, but compared to the accuracy and precision rates, there is still some room for improvement. The recall rate reflects the model’s ability to identify all actual high-risk cases. A lower recall rate may mean that some high-risk cases are missed by the model, which may pose certain risks in practical applications. To comprehensively evaluate the performance of the model, the F1 score metric is introduced. The F1 score is the harmonic mean of accuracy and recall rates, which can comprehensively reflect the model’s ability to balance the two. The F1 score of this model reaches 0.801, indicating that while maintaining high accuracy and precision rates, the model also achieves relatively good performance in terms of the recall rate, demonstrating excellent overall performance.
Meanwhile, comparing the evaluation data of other models in Table 4, it can be seen that the XGBoost model combined with the improved Sparrow Search Algorithm outperforms other models in all evaluation metrics, indicating that the risk assessment effect of the XGBoost model combined with the improved Sparrow Search Algorithm is superior to other models, which can effectively assist financial institutions in credit risk assessment to a certain extent, improving the scientificity and accuracy of decision making. Furthermore, the XGBoost model combined with the improved Sparrow Search Algorithm is superior to the classic XGBoost model, proving that the model with optimized hyperparameters can better identify the changing trends of specific data, thereby producing better prediction results.

5. Construction of Evaluation Model

This paper constructs an intelligent assessment model for personal credit risk under the background of big data, providing new methodological support for the intelligent assessment of personal credit risk. This model is based on the forest optimization feature selection algorithm, combined with initialization based on chi-square verification, adaptive global seeding, and greedy search strategy, which can automatically identify key risk factors accurately from high-dimensional data. It has made up for the shortcomings of low efficiency and easy omission in traditional manual selection methods, and improved the accuracy of risk factor identification. Then, the model takes personal credit risk factors as input features and uses the XGBoost algorithm to evaluate the customer’s credit risk level. Improve the traditional Sparrow Search Algorithm by using Tent chaotic mapping, sine cosine search, reverse learning, and Cauchy mutation strategy to enhance the optimization performance of algorithm parameters. This model can evaluate personal credit risk and predict customer default probability, making up for the shortcomings of traditional evaluation methods in data processing breadth and risk assessment accuracy, and improving the accuracy of customer default probability prediction.
Based on all the work of this paper, we can draw the following conclusions:
(1) Constructing an intelligent identification model for personal credit risk factors in the context of big data is an effective way to enhance the efficiency of credit risk management. By utilizing machine learning technology to automatically identify potential risk factors, the accuracy and efficiency of risk identification have been significantly improved. The model can extract key features from massive data, providing a solid data foundation and support for subsequent risk assessment and decision making. The establishment of an intelligent identification model not only reduces the time and cost of manual review but also enhances the scientific and systematic nature of risk management.
(2) Developing an intelligent assessment model for personal credit risk in the context of big data can significantly enhance the comprehensiveness and scientific nature of risk assessment. By quantitatively evaluating credit risk and incorporating various advanced algorithms and technologies, such as XGBoost and an improved Sparrow Search Algorithm, this model ensures the objectivity and accuracy of assessment results. It provides a more reliable and comprehensive basis for credit institutions’ risk management decisions, helping institutions better understand and control credit risk, thereby optimizing resource allocation and reducing the non-performing loan ratio.
The method proposed in this paper has broad potential for real-world applications, especially in financial institutions such as banks, which can effectively address the limitations of traditional risk assessment models in processing high-dimensional data and improve the accuracy and efficiency of credit assessment. Traditional credit risk assessment relies on limited historical data and manual experience judgment, making it difficult to comprehensively capture the subtle factors that affect credit risk. Financial institutions can use the method proposed in this paper to mine high-dimensional data, accurately identify key features that affect personal credit risk, and lay a solid foundation for subsequent risk assessment. Financial institutions can use intelligent risk assessment models based on big data to quickly process and analyze complex datasets. This efficiency not only shortens the loan approval cycle and improves user experience, but also provides personalized risk assessment results based on each applicant’s specific circumstances, such as credit history, income level, consumption behavior, etc. This not only enhances the accuracy of risk assessment, but also promotes the diversification and customization of credit products, better meeting the needs of different customer groups. However, potential ethical and technological challenges also need to be considered in practical applications. For example, algorithms may inherit unfair decision-making patterns due to historical biases in training data, which in turn can affect the credit evaluation results of certain groups. To ensure the fairness and transparency of the model, financial institutions should avoid using sensitive features when applying it and strengthen the monitoring and interpretation of the model’s decision-making process.
Although this study has made some progress in the field of personal credit risk assessment, there are still some limitations that need further exploration and improvement. First is the deepening and integration of theoretical frameworks. With the continuous advancement of technology and the accumulation of practical experience, the theoretical system of personal credit risk management in the future will be further deepened and refined. It will further integrate interdisciplinary perspectives such as behavioral economics, psychology, sociology, etc., to comprehensively understand and predict borrower behavior patterns and risk changes. Secondly, multimodal data fusion is an important direction for future research. This article is mainly based on traditional structured data, and in the future, we can further explore the fusion of various unstructured data such as text, images, and videos to more comprehensively capture customers’ credit status and behavioral characteristics. Third is further optimizing the performance of the model. By combining the advantages of deep learning and reinforcement learning, more complex and efficient models can be developed in the future to cope with more dynamic and complex credit risk environments. Fourthly, this article did not directly consider the impact of macro factors on individual credit risk. Future research can explore how to combine macro factors with individual characteristics to construct more comprehensive credit risk assessment models. In addition, there may be complex nonlinear relationships between macro factors and personal characteristics, which provides new research directions for the application of machine learning methods.

Author Contributions

Conceptualization, methodology, software, formal analysis, investigation, writing—original draft preparation, H.Y.; writing—review and editing, C.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Beijing Social Science Foundation, grant number 20GLB028.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data were obtained from https://www.kaggle.com (accessed on 14 August 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, H.Q.; Lin, H.Z.; Huang, F. Research on the influencing factors of personal micro credit risk. J. Appl. Stat. Manag. 2022, 41, 1105–1115. [Google Scholar] [CrossRef]
  2. Silva, F.B.G. Fiscal Deficits, Bank Credit Risk, and Loan-Loss Provisions. J. Financ. Quant. Anal. 2021, 56, 1537–1589. [Google Scholar] [CrossRef]
  3. Bernanke, B.S.; Gertler, M. Inside the Black Box: The Credit Channel of Monetary Policy Transmission. J. Econ. Perspect. 1995, 9, 27–48. [Google Scholar] [CrossRef]
  4. Schularick, M.; Taylor, A.M. Credit booms gone bust: Monetary policy, leverage cycles, and financial crises, 1870-2008. Am. Econ. Rev. 2012, 102, 1029–1061. [Google Scholar] [CrossRef]
  5. Govindarajan, V.; Ilyas, H.; Silva, F.B.G.; Srivastava, A.; Enache, L. How Companies Can Prepare for a Long Run of High Inflation. Harv. Bus. Rev. 2022. Available online: https://hbr.org/2022/05/how-companies-can-prepare-for-a-long-run-of-high-inflation (accessed on 2 February 2025).
  6. Wu, F.; Su, X.; Ock, Y.S.; Wang, Z. Personal credit risk evaluation model of P2P online lending based on AHP. Symmetry 2021, 13, 83. [Google Scholar] [CrossRef]
  7. Wang, J.X.; Ma, Y.L. Research on Personal Credit Risk Assessment Methods Based on Causal Inference. arXiv 2024, arXiv:2403.11217. [Google Scholar]
  8. Zhu, X.Q. Modeling and analysis of personal credit risk assessment of commercial banks. Sci. Technol. Ecnony Mark. 2023, 12, 86–88. [Google Scholar]
  9. Zhang, G.Z.; Chen, W.H.; Liu, C.H. Research on risk assessment of personal consumption credit of commercial banks based on logistic model. Financ. Theory Pract. 2015, 03, 53–57. [Google Scholar]
  10. Cai, W.X.; Luo, Y.H.; Zhang, G.X.; Zhong, H.L. Personal credit risk assessment model and empirical analysis based on the fusion of gbdt and logistic regression. Mod. Manag. 2017, 37, 1–4. [Google Scholar]
  11. Wu, J.; Chen, D.F.; Li, R.T.; Shi, J.Y. Research on personal credit risk assessment based on xgboost and shap method. Mod. Inf. Technol. 2024, 8, 146–150+155. [Google Scholar] [CrossRef]
  12. Liao, W.X.; Zeng, B.; Liang, T.K.; Xu, Y.; Zhao, J. Personal credit risk assessment method for high dimensional data. Comput. Eng. Appl. 2020, 56, 219–224. [Google Scholar]
  13. Wang, Y.; Wu, Z.; Gao, J.; Liu, C.; Guo, F. A multi-level classification based ensemble and feature extractor for credit risk assessment. PeerJ Comput. Sci. 2024, 10, e1915. [Google Scholar] [CrossRef]
  14. Zhang, H.; Zeng, R.; Chen, L.; Zhang, S. Research on personal credit scoring model based on multi-source data. J. Phys. Conf. Ser. 2020, 1437, 012053. [Google Scholar] [CrossRef]
  15. Chen, G.; Li, Z.L.; Zhu, L. Personal credit risk prediction based on multi scale depth feature fusion. Comput. Eng. Sci. 2023, 45, 2265–2273. [Google Scholar]
  16. Namakin, M.; Rouhani, M.; Sabzekar, M. A Multi-objective Feature Selection Method Considering the Interaction Between Features. Inf. Syst. Front. 2024, 1–16. [Google Scholar] [CrossRef]
  17. Mu, Q.; Qin, L.L. Statistical models and methods in financial risk management. Bus. News 2024, 13, 88–91. [Google Scholar]
  18. Zhang, J.L.; Guo, S.Y.; Ren, C.P.; Ma, Q. Research on personal credit scoring card model based on logistic regression. Mod. Inf. Technol. 2024, 8, 12–16. [Google Scholar] [CrossRef]
  19. Li, S.J.; Ji, X.J. Application of lasso Cox model in personal credit risk assessment. Resour. Dev. Mark. 2021, 37, 129–135. [Google Scholar]
  20. Xu, X.X. Research on personal credit risk based on Stochastic Forest Model. Audit. Financ. 2024, 9, 55–58. [Google Scholar] [CrossRef]
  21. Gao, X.; Yang, X.; Zhao, Y. Rural micro-credit model design and credit risk assessment via improved LSTM algorithm. PeerJ Comput. Sci. 2023, 9, e1588. [Google Scholar] [CrossRef] [PubMed]
  22. Li, H.; Cao, Y.; Li, S.; Zhao, J.; Sun, Y. XGBoost model and its application to personal credit evaluation. IEEE Intell. Syst. 2020, 35, 52–61. [Google Scholar] [CrossRef]
  23. Zhao, Y. Research on personal credit evaluation of internet finance based on blockchain and decision tree algorithm. EURASIP J. Wirel. Commun. Netw. 2020, 2020, 213. [Google Scholar] [CrossRef]
  24. Zhou, X.; Zhang, W.; Jiang, Y. Personal credit default prediction model based on convolution neural network. Math. Probl. Eng. 2020, 2020, 5608392. [Google Scholar] [CrossRef]
  25. Liang, L.Y.; Wang, H.Z. Prediction of personal credit risk based on graph convolution neural network. Comput. Eng. Appl. 2023, 59, 275–285. [Google Scholar]
  26. Li, S.J.; Mei, H. Application of deep cycle survival analysis in personal credit evaluation. China Collect. Econ. 2023, 33, 94–97. [Google Scholar]
  27. Zhong, J.H.; Shao, J.J.; Li, X.G. Research on personal credit risk assessment based on portfolio classification strategy. J. Hefei Univ. Technol. (Nat. Sci.) 2020, 43, 996–1002. [Google Scholar]
  28. Zhang, Z.M.; Lv, J.; Liu, Y. Research on personal credit risk assessment model based on optimized acgan-gbdt. Theory Pract. Financ. Econ. 2022, 43, 84–89. [Google Scholar] [CrossRef]
  29. Yan, Q.; Xu, H.Y. Research on customer credit risk of microfinance companies based on hybrid smote and RF model. Oper. Res. Manag. Sci. 2024, 33, 191–197. [Google Scholar]
  30. Shen, G.F. Research on personal credit risk assessment based on PCA and SVM. J. Yichun Univ. 2022, 44, 49–52+120. [Google Scholar]
  31. Rao, C.; Liu, Y.; Goh, M. Credit risk assessment mechanism of personal auto loan based on PSO-XGBoost Model. Complex Intell. Syst. 2023, 9, 1391–1414. [Google Scholar] [CrossRef]
  32. Chen, J.J.; Liu, S. Personal credit evaluation based on SVM optimized by improved cerambus algorithm. Comput. Technol. Dev. 2021, 31, 135–139. [Google Scholar]
  33. Gu, Q.H.; Song, S.Y.; Zhang, X.S.; Bao, Z.Q. Research on personal credit risk assessment method based on improved BS stacking model. Oper. Res. Manag. Sci. 2023, 32, 137–144. [Google Scholar]
  34. Charilaou, P.; Battat, R. Machine learning models and over-fitting considerations. World J. Gastroenterol. 2022, 28, 605. [Google Scholar] [CrossRef]
  35. Hamel, J.F.; Sébille, V.; Challet-Bouju, G.; Hardouin, J.B. Partial credit model: Estimations and tests of fit with pcmodel. Stata J. 2016, 16, 464–481. [Google Scholar] [CrossRef]
  36. Xie, Y.; Liu, G.J.; Yan, C.G.; Jiang, C.; Zhou, M.; Li, M. Learning Transactional Behavioral Representations for Credit Card Fraud Detection. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 5735–5748. [Google Scholar] [CrossRef]
  37. Xie, Y.; Liu, G.J.; Yan, C.G.; Jiang, C.; Zhou, M. Time-Aware Attention-Based Gated Network for Credit Card Fraud Detection by Extracting Transactional Behaviors. IEEE Trans. Comput. Soc. Syst. 2023, 10, 1004–1016. [Google Scholar] [CrossRef]
  38. Xie, Y.; Liu, G.J.; Zhou, M.C.; Wei, L.; Zhu, H.; Zhou, R.; Cao, L. A Spatial–Temporal Gated Network for Credit Card Fraud Detection by Learning Transactional Representations. IEEE Trans. Autom. Sci. Eng. 2024, 21, 6978–6991. [Google Scholar] [CrossRef]
  39. Zhou, J.W.; Shao, J.; Cao, S. Improved sparrow search point cloud registration algorithm. Comput. Eng. Appl. 2024, 1–13. Available online: https://kns-cnki-net.webvpn.cueb.edu.cn/kcms/detail/11.2127.TP.20240911.1553.012.html (accessed on 18 December 2024).
  40. Liu, P.L.; Lu, Q. Self learning sparrow search algorithm with escape mechanism. Comput. Digit. Eng. 2023, 51, 2769–2773. [Google Scholar]
  41. Xiao, S.H.; Du, C.J.; Wang, C.J. Improved sparrow search algorithm to optimize BP neural network prediction model of pipeline corrosion rate. Oil Gas Storage Transp. 2024, 43, 760–768+795. [Google Scholar]
  42. Wang, X.; Ru, X.W. Sand cat swarm optimization algorithm integrating sparrow search algorithm and Cauchy mutation strategy. J. Tonghua Norm. Univ. 2024, 45, 35–41. [Google Scholar]
  43. Wang, R.L.; Wang, H.B.; Li, Z.F.; Li, P.T.; Wen, H.; Liu, C.J. Improved Sparrow Search Algorithm with Multi Strategy Fusion. J. Jilin Inst. Chem. Technol. 2024, 41, 68–75. [Google Scholar] [CrossRef]
  44. Xue, S.R.; Zhang, C.S.; Ding, X.; Qian, B.; Hu, R. Integrating multiple strategies to improve sparrow search algorithm. J. Yunnan Univ. (Nat. Sci. Ed.) 2024, 1–12. Available online: https://kns-cnki-net.webvpn.cueb.edu.cn/kcms/detail/53.1045.n.20241021.0954.004.html (accessed on 11 December 2024).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.