Next Article in Journal
Application of Quantum—Markov Open System Models to Human Cognition and Decision
Previous Article in Journal
Intermittency and Critical Scaling in Annular Couette Flow
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Harmony Search–Artificial Intelligence Models in Credit Scoring

1
Laboratory of Computational Statistics and Operations Research, Institute for Mathematical Research, Universiti Putra Malaysia, Serdang, Selangor 43400, Malaysia
2
Department of Mathematics, Faculty of Science, Universiti Putra Malaysia, Serdang, Selangor 43400, Malaysia
3
Nottingham University Business School, University of Nottingham Malaysia, Semenyih, Selangor 43500, Malaysia
*
Author to whom correspondence should be addressed.
Entropy 2020, 22(9), 989; https://doi.org/10.3390/e22090989
Submission received: 23 June 2020 / Revised: 28 August 2020 / Accepted: 1 September 2020 / Published: 4 September 2020
(This article belongs to the Section Multidisciplinary Applications)

Abstract

:
Credit scoring is an important tool used by financial institutions to correctly identify defaulters and non-defaulters. Support Vector Machines (SVM) and Random Forest (RF) are the Artificial Intelligence techniques that have been attracting interest due to their flexibility to account for various data patterns. Both are black-box models which are sensitive to hyperparameter settings. Feature selection can be performed on SVM to enable explanation with the reduced features, whereas feature importance computed by RF can be used for model explanation. The benefits of accuracy and interpretation allow for significant improvement in the area of credit risk and credit scoring. This paper proposes the use of Harmony Search (HS), to form a hybrid HS-SVM to perform feature selection and hyperparameter tuning simultaneously, and a hybrid HS-RF to tune the hyperparameters. A Modified HS (MHS) is also proposed with the main objective to achieve comparable results as the standard HS with a shorter computational time. MHS consists of four main modifications in the standard HS: (i) Elitism selection during memory consideration instead of random selection, (ii) dynamic exploration and exploitation operators in place of the original static operators, (iii) a self-adjusted bandwidth operator, and (iv) inclusion of additional termination criteria to reach faster convergence. Along with parallel computing, MHS effectively reduces the computational time of the proposed hybrid models. The proposed hybrid models are compared with standard statistical models across three different datasets commonly used in credit scoring studies. The computational results show that MHS-RF is most robust in terms of model performance, model explainability and computational time.

1. Introduction

Credit risk evaluation is a crucial routine of risk management in financial institutions. Credit scoring models are the main tool utilized to make credit granting decisions where the probability of default resembles the entropy concept, i.e., probabilistic measure of uncertainty. Hence to better measure risk, more accurate classification models are needed. Though statistical models are usually the preferred option, Artificial Intelligence (AI) models are beginning to be favoured for their accuracy and flexibility in the face of the volume of data. Advances in these techniques have further increased their popularity particularly in risk assessments. Support Vector Machines (SVM) and Random Forest (RF) are the main AI classifiers used in this study as recommended in two large scale benchmark studies by [1,2], respectively, due to their competitive performance as compared to other classifiers. There are two issues to be considered when using SVM and RF, i.e., sensitivity to hyperparameters settings and the black-box property.
For hyperparameter tuning, Grid Search (GS) has always been the conventional tuning tool for both SVM and RF. Recently, the metaheuristic approaches (MA) have shown potential as a competitive tool to tune SVM hyperparameters [3]. Some works utilized Genetic Algorithm (GA) [4,5] and Particle Swarm Optimization (PSO) to tune SVM [6,7]. A recent work experimented with MA [8] by using Artificial Bee Colony (ABC). These MA techniques have reported competitive results, indicating the potential of MA to be used to tune SVM hyperparameters.
For RF hyperparameter tuning, the major approach is the repeated trial-and-error tuning which requires subjective judgement from researchers [9,10,11,12,13]. Some researches tune the hyperparameters by examining a certain input range which is available in some software toolbox [2,14]. The GS is still a popular technique to tune RF [15,16]. Reference [16] compared GS with Random Search and PSO, and pointed out the benefit of PSO. Despite manual tuning being the common approach, experiments with PSO [16] shows the potential of MA.
Solving the black-box property is a challenging task. For SVM, feature selection strategy to enable explanation on reduced features is frequently attempted. GA has shown its potential in developing a different wrapper GA-SVM with the ability to reduce the features of SVM, yet maintaining a good model performance. References [17,18] incorporated information from a filter technique as the input to the GA wrapper. References [19,20] proposed a hybrid GA-SVM to perform hyperparameter tuning and feature selection simultaneously, whereas [20] included feature weighting in the wrapper GA-SVM model.
RF has the advantage of being able to explain the attributes with the computed feature importance. References [9,13] provided attributes information based on the feature importance. Reference [12] used this benefit for feature screening. Reference [10] incorporated the feature importance with profit measures while [11,21] built new credit scoring models with the feature ranking.
Despite being the most common technique for hyperparameter tuning, GS is a rigid brute force technique that will search through all possible combinations of the hyperparameters. For continuous hyperparameters, the computational effort will increase as the granularity of the search range increases. In addition, using GS to conduct feature selection will tremendously increase the computational time due to the increased features search space. Thus, to address both hyperparameter tuning and model explainability simultaneously, MA is a potential candidate tool to be hybridized with SVM and RF, with GA being the most commonly used method in the past. Recently, Harmony Search (HS) has received attention to be hybridised with SVM [22,23,24] and RF [25] in various domains for the purpose of hyperparameter tuning or feature selection. The authors of [26] have reviewed works using HS to conduct feature selection along with the use of different machine learning algorithms across various domains. Despite the successful implementations of HS for hyperparameter tuning and feature selection, ref. [27] has been the only study utilizing HS for feature selection in the nearest neighbourhood credit scoring model.
To the best of our knowledge, HS has yet to be hybridized with SVM and RF for the purpose of simultaneous hyperparameter tuning and model explainability in credit scoring. The HS first developed by [28] is inspired by the music improvisation process, where musicians tune their instruments’ pitch to achieve perfect harmony in seeking for an optimal solution. Hence, two hybrid models, i.e., HS-SVM and HS-RF are proposed in this study. HS-SVM conducts hyperparameter tuning and feature selection simultaneously to select appropriate hyperparameters and explain the attributes based on the reduced features. The HS-RF conducts hyperparameter tuning to ensure good model performance to provide reliable feature importance. SVM and RF are then hybridized with a modified HS (MHS) to improve the computational efficiency yet maintaining a comparable performance of the GS and HS. Parallel computation is applied on MHS hybrid models to improve the computational time. Then, the proposed models are compared with standard statistical models across two well-known credit scoring datasets and a peer-to-peer lending dataset. The discussions are based on model performances, model explainability, and computational time. Competitive results of the proposed models highlight the flexibility of utilizing HS as compared to GS, i.e., the ability to conduct feature selection and search for continuous hyperparameters without the need to specify the granularity of the search range. In addition, the competitiveness of MHS hybrid models further demonstrates the flexibility of HS to be modified to improve computational efficiency.
Research in credit scoring studies have been continuously attempted with various AI techniques. Several recent studies have marked a paradigm shift towards the usage of non-linear advanced techniques [29] and ensemble models [30,31,32] to be the potential techniques to achieve good performance in handling various types of data patterns. Due to the property of decision tree models being the de-facto community standard in classification task [33], tree-based ensembles have received attention in [31,32] with competitive performances reported on credit scoring datasets. In line with the recommendation from the large scale benchmark studies [1,2] and the paradigm shift observed in recent literatures, the proposed models in this study which are based on SVM and RF are aligned with the current trend of utilizing non-linear AI techniques and tree-based models. Most of the recent studies with the new advanced AI techniques have been a direct application of the models to assess their performance on credit scoring datasets to investigate their potential, with only [31] included synthetic features to address model explainability. Instead of a direct application for investigative purpose, the proposed hybrid models provide new idea to improve performance via simultaneous hyperparameters tuning and features selection. Besides, usage of RF has the benefit compared to the other tree-based ensembles because the computed features ranking is an appropriate tool for model explanation.
This paper is organized as follows. Section 2 provides an overview of the HS algorithm and the numerical experiments to demonstrate the potential of HS hybrid models, leading to the intuition to develop MHS hybrid models. Section 3 details on the hybrid models’ formulation. Section 4 elaborates on the experimental setup. Then, Section 5 reports the computational results with detailed discussions. Finally, Section 6 concludes the study and provides possible future directions.

2. Harmony Search to Modified Harmony Search

2.1. Harmony Search

The HS metaheuristic is a random search technique guided by fitness function evaluations. The HS search process is controlled by explorative and exploitative operators to seek solutions from the search range. A standard HS algorithm [28] consists of five procedures as follows:
  • Definition of objective function and the required parameters. The parameters are Harmony Memory Size ( H M S ), Harmony Memory Considering Rate ( H M C R ), Pitch-Adjustment Rate ( P A R ), bandwidth ( b w ), and maximum iterations ( m a x _ i t e r ).
  • Initialization of Harmony Memory (HM). HM consists of the H M S number of possible candidate solutions with n dimensions which depends on the number of decision variables. All the candidate solutions in HM are generated from a uniform distribution that is based on the decision variables’ range.
  • Improvisation to generate new harmony. There are two main operators, i.e., H M C R to control the exploration and P A R to control the exploitation of the search process. H M C R is the probability of selecting a new harmony from HM, while its counterpart ( 1 H M C R ) is the probability to randomly generates a new harmony. A low H M C R indicates high explorative power of the search process, because the search process will continuously generate a new harmony out of HM by exploring different search spaces. P A R is the probability to improvise the selected harmony from HM by moving to the neighbouring values with a step size of b w (for continuous variables) or one step to the left or right (for discrete variables). A low P A R indicates low exploitation power to conduct local exploitation of the harmony around its neighbourhood.
  • Update H M . The new harmony from (3) is evaluated against the fitness function. Replace the worst solution in HM with the new harmony if it has a better fitness value.
  • Termination. Repeat (3) and (4) until m a x _ i t e r has reached.

2.2. Numerical Experiment Part I: Potential of HS Compared to GS

A numerical experiment is conducted to compare the performance of HS hybrids with the GS approach for hyperparameter tuning. For demonstration purposes, the numerical experiment is implemented on the German credit dataset and evaluated with the Area Under Receiver Operating Characteristics (AUC) from the average of 10-fold cross validation. Details of HS-SVM and HS-RF are enclosed in Section 3.1 and Section 3.2, while the details of the German dataset can be found in Section 4.1.
SVM hyperparameters search range follows the recommended settings by [8], i.e., log 2 C = [ 5 , 12 ] and log 2 γ = [ 12 , 5 ] . Both hyperparameters are continuous variables that can take any values within the range. For GS, it is important to determine the granularity of the grid. Thus, GS for SVM will first search at a coarse grid of log 2 C = [ 5 , 4 , 3 , , 11 , 12 ] and log 2 γ = [ 12 , 11 , 10 , , 4 , 5 ] then follow with a finer grid of a granularity of 0.05 around the best returned solution from the coarse grid. For HS-SVM, the search process will examine the whole search range by picking the values via a uniform distribution, without the need to set the granularity.
RF hyperparameters are discrete variables with the search range of n t r e e = { 100 , 200 , , 500 } and m t r y = { 1 , 2 , , a } , where a is the total number of attributes available. For both GS and HS, the search will be conducted across the search range but the granularity will not have to be determined as both are discrete variables.
In order to show that HS is a competent tool, the main parameters in HS, i.e., H M C R and P A R have to be robust across small perturbations. To set up for the experiment, a recommended range from [34] is applied ( H M C R = [ 0.70 , 0.95 ] , P A R = [ 0.10 , 0.35 ] ) for both parameters. Both parameters are perturbed at a small change of 0.05 to examine their robustness. The AUC performances and computational time are reported in Table 1.
In terms of the model performance, both HS-SVM and HS-RF have reported a high mean AUC with only 0.2% and 0.1% standard deviation, respectively, across all the different combinations of H M C R and P A R . Despite the small perturbations of H M C R and P A R , HS-SVM and HS-RF still result in a stable performance, indicating HS is a robust tool to be hybridized with SVM and RF for hyperparameter tuning. Besides this, the results from Table 1 also imply that the H M C R and P A R range recommended by [34] is reliable as all the models have reported a competitive AUC across the whole range of the two operators.
However, the computational time deviation for HS-SVM and HS-RF are slightly higher. This is due to the effect of different hyperparameters that require different computational power. During the search process, HS may explore different areas of the hyperparameters search space followed with exploitation in the neighbourhood areas. Hence, depending on the search areas led by the operators, the different combinations of the operators will lead to different search areas where some consume more computing power than the others.
GS-tuned SVM and RF are compared with HS-SVM and HS-RF, respectively (see Table 2). HS hybrid models have achieved a slightly better model performance (higher AUC). HS-SVM is effective in the computational time and show competitive AUC performance. The extra computational effort for GS approach is due to the continuous search space of SVM hyperparameters that require the GS process to first search on a coarse grid followed by a finer grid. This is the main advantage of HS which is able to save computational effort by not needing to determine the granularity of the continuous search space. Due to the discrete hyperparameters search space for RF, there is no huge difference in the computation time between the GS-tuned RF and HS-RF.
The HS has demonstrated its potential as a competent tool to tune SVM and RF, with m a x _ i t e r as the only stopping criteria. The competent AUC performance reported in Table 1 implies m a x _ i t e r = 100 is sufficient for the search. The b w assistant exploitation tool controls how far the new solution should be adjusted. HS-SVM has b w = 0.1 , which is an appropriate width to move around the SVM hyperparameters search range. Since HS-RF has discrete decision variables, settings of b w is not required.
Since the HS is more computationally efficient than the GS (depending on the m a x _ i t e r ), this inspired the idea to further enhance the computational efficiency by having the whole search process end with lesser number of iterations. However, it will be doubtful to set a low m a x _ i t e r as there may still be more space to be searched upon. Therefore, the convergence pattern of the hybrid HS models have to be observed at different levels of exploration and exploitation.

2.3. Numerical Experiment Part II: Intuition to Develop MHS Hybrid Models

2.3.1. Search Patterns of HS-SVM and HS-RF

Figure 1 and Figure 2 illustrate the search patterns for HS-SVM and HS-RF at four different combinations of H M C R and P A R operators for the first fold of the cross validation. H M C R = 0.70 and H M C R = 0.95 indicate higher and lower exploration, respectively, while P A R = 0.10 and P A R = 0.35 indicate for lower and higher exploitation, respectively. Generally, these combinations lead to only a slight difference in the final AUC performance. However, the search patterns of the different combinations show how exploration and exploitation control the search process. The search process is said to reach a convergence when ‘a plateau’ pattern is observed.
For both HS-SVM and HS-RF, the lower H M C R value has demonstrated higher explorative power compared to the higher H M C R value, given that P A R settings are constant ( H M C R = 0.70 , P A R = 0.10 / H M C R = 0.95 , P A R = 0.10 and H M C R = 0.70 , P A R = 0.35 / H M C R = 0.95 , P A R = 0.35 ). At a lower H M C R value, the curves show sharp increment patterns, indicating active movement to different search areas. On the other hand, at a higher H M C R value, the increment patterns have lower gradient during the transition to higher AUC. The sharper transitions show that global search takes more dominant role than local search in the process.
As for the P A R operator, the lower P A R value has demonstrated lower exploitative power compared to the higher P A R value, given that H M C R settings are constant ( P A R = 0.10 , H M C R = 0.70 / P A R = 0.35 , H M C R = 0.70 and P A R = 0.10 , H M C R = 0.95 / P A R = 0.35 , H M C R = 0.95 ). At a higher P A R value, the curves assist in the global search with more transition points before moving to another search space with a higher AUC. This indicates a more active local search at a higher P A R to improve the AUC by shifting to the neighbourhood search area.
Towards the end of the search process, ‘a plateau’ pattern is observed, indicating the process has reached convergence. Overall, all the different combinations have reached convergence, with certain combinations showing earlier convergence and the others. Thus, the main intuition for the development of a modified HS (MHS) is to achieve a comparable performance as the HS but at an earlier convergence to save up extra computational effort. The main modifications on HS to develop a MHS are as follows:
  • Elitism selection during memory consideration
    The selection of new harmony is no longer a random selection from H M , but with an objective to select a better quality harmony. Elitism selection leads the search process to focus on better quality candidates, thus enabling a faster convergence. Harmony vectors in H M are divided into two groups, i.e., elite ( g 1 ) and non-elite ( g 2 ), where g 1 consists of harmony vectors with better performance than g 2 .
    Each harmony vector in H M takes an index number from the sequence of { 1 , H M S } . Since H M is sorted in the order of best to worst performance, harmony vectors with lower index number indicate their potential as the candidates in the elite group. The first quartile, q 1 of the index sequence is computed as in Equation (1), with decimal places being rounded up because index values are discrete. The computed q 1 is the cutoff to divide H M into the elite and non-elite groups where g 1 { 1 , q 1 } and g 2 { ( q 1 + 1 ) , H M S } .
    q 1 = r o u n d ( 0.25 × ( H M S + 1 ) ) t h term .
    An extra parameter e l i t is included to allocate a proper weightage on the elite group. So, the selected new harmony has a higher probability to originate from the elite group. With a probability e l i t , a new harmony is selected from the elite group. If the selection is from the non-elite group, two harmonies will be picked. Then, the better one of the two will be the new harmony.
    By doing this, a better harmony is always selected. Note that a low quality harmony, when joining with other harmony or being adjusted, may also produce good harmony. Thus, e l i t cannot be too high to ensure a balance to seek from elite and non-elite group. The detailed selection process is illustrated in Algorithm 1.
    Algorithm 1
    selection( )
    g 1 { 1 , q 1 } /*refer Equation (1)*/
    g 2 { ( q 1 + 1 ) , H M S } /*refer Equation (1)*/
    if ( U ( 0 , 1 ) e l i t )
            i n d 1 = r o u n d ( g 1 m i n + U ( 0 , 1 ) × ( g 1 m a x g 1 m i n ) )
            i n d = i n d 1
    else
            i n d 2 = r o u n d ( g 2 m i n + U ( 0 , 1 ) × ( g 2 m a x g 2 m i n ) )
            i n d 3 = r o u n d ( g 2 m i n + U ( 0 , 1 ) × ( g 2 m a x g 2 m i n ) )
           if( i n d 2 i n d 3 )
                   i n d = i n d 2
           else
                   i n d = i n d 3
    return i n d
  • Dynamic H M C R and P A R with step function
    The numerical experiment demonstrates repeated trials with different combinations of H M C R and P A R for the hybrid HS models. The competitive results reported from the different combinations indicated that the recommended range by [34] ( H M C R = [ 0.70 , 0.95 ] and P A R = [ 0.10 , 0.35 ] ) is an appropriate range for both operators. Along with the elitism selection, it is important to ensure sufficient exploration and exploitation of the search process before reaching convergence. Thus, the H M C R and P A R is designated to be dynamic following an increasing and decreasing step function, respectively.
    The increasing and decreasing step function of H M C R and P A R cooperates with each other for a balance of exploration and exploitation. Initially, a lower H M C R provides an active global search to explore the search area and it works along with a higher P A R that provides an active local search to exploit the neighbourhood of the search area. Thus, H M consists of candidates scattering around the search area with its corresponding neighbourhood being well-exploited in the early stage of the search procedure. Following the step function, the global search exploration decreases and focuses in the search area stored in the H M , leading to the local exploitation to be focused in this specific search area. The dynamic settings of H M C R and P A R enable effective determination of the appropriate search area that lead to a more efficient convergence towards the final solutions.
    In utilizing the step function, several components, i.e., H M C R range, P A R range, H M C R increment, P A R decrement, and step size ( s t e p ) have to be determined. Based on the numerical experiment conducted earlier, the range of the operators are set as the recommended range. The interval for increment and decrement of H M C R and P A R , respectively, is set at 0.05 as this small interval is sufficient to cover the whole range for these two operators. The s t e p determines the number of iterations for H M C R and P A R to maintain before shifting to another value in the range until both operators reach a plateau. The setting of s t e p depends on the search range size with a smaller step preferable as the main aim is to have faster convergence with active exploration and exploitation in the early stage of the search. Thus, s t e p is set to enable both H M C R and P A R to reach a plateau within the first half of the total iterations. For the numerical experiment, MHS-SVM has s t e p = 10 while MHS-RF has s t e p = 5 . The smaller s t e p for MHS-RF is due to its smaller discrete search space than MHS-SVM with continuous search space.
  • Self-adjusted b w
    b w is an assistant tool for pitch adjustment and poses an effect on local exploitation. We suggest to replace the b w using a coefficient of variation ( c o e f ) (Equation 2) of the decision variable for every iteration, which will now be an auto-updated value in each iteration, thus enabling possible early convergence. This modification is only applicable for continuous decision variables as b w is not required for the adjustment of discrete decision variables.
    s i = j = 1 H M S ( x i j x i ¯ ) H M S 1 x i ¯ = j = 1 H M S x i j H M S c o e f i = s i / x i ¯ i = 1 , 2 , , n .
    This intuition comes from several past researches [35,36,37] that have proposed the improved HS with the b w modified. From these modifications, it is suggested that the dynamic b w should converge to smaller values as the iterations of the search process increases. Reference [35] recommended standard deviation ( s d ) as the appropriate replacement of b w . Reference [37] also utilized s d to replace b w , with an additional constant attached to control the local exploitation. Hence, using c o e f to substitute b w is an appropriate strategy because the division of s d with the mean is perceived as equivalent to the attached constant as in [37], yet has the benefit of being automatically updated in every iteration. Besides, c o e f can effectively scale the s d to ensure the search processes are maintained in an appropriate range. When the iterations increase, solutions in HM will converged, causing the c o e f to converge to smaller values.
  • Additional termination criteria
    The termination criteria used in this study are the maximum number of iterations ( m a x _ i t e r ), convergence of HM, and non-improvement on the best solution for a fixed number of consecutive iterations ( c o n s _ n o _ i m p ). Since the previous three modifications open up the possibility for faster convergence, both criteria are included to avoid redundant iterations to save computational effort. MHS procedure will stop when any one of the criteria is met.

2.3.2. Potential of MHS-SVM and MHS-RF

The numerical experiment in Section 2.3.1 is repeated with the MHS hybrid models. The search patterns of the MHS hybrid models are compared with the HS hybrid models in Figure 3 and Figure 4 to illustrate the effect of the modifications.
For both MHS-SVM and MHS-RF, the modifications lead to earlier convergence, with the ‘a plateau’ pattern achieved much earlier compared to the HS hybrid models. The vertical lines in the figures mark the number of iterations required for the search process to end, while the AUC after the vertical lines is the result if the search process is allowed to run for the full number of iterations. With the MHS, the increment of AUC has shown a faster transition towards ‘a plateau’ with fewer numbers of iterations as compared to the HS hybrids, yet maintaining a comparable AUC performance (even with the other different settings of HS hybrid models). This indicates the MHS hybrid models have active exploration and exploitation in the earlier stages of the search, fulfilling the objectives of the MHS hybrid models to reach convergence with lesser iterations. This is the benefit of MHS hybrid models which can help by not needing to perform additional efforts for different H M C R and P A R combinations to check the model performance.
Table 3 compares the three different approaches, i.e., GS, HS, and MHS for the hyperparameter tuning task; in terms of measuring the AUC and the resulting computational time from the required number of iterations to end the search process. Overall, the HS hybrid models have potential in reducing computational effort especially when the hyperparameters are of continuous variables. The MHS hybrid models save up more computational efforts by reducing the number of iterations and the AUC reported are comparable to that of the GS approach and the standard HS.

3. Hybrid Models

HS and MHS act as the assistant tool to solve both hyperparameter tuning and model explainability tasks of SVM and RF models. All the proposed hybrid models are supported by machine learning theory as the underlying technique to carry out the final classification which consist of the supervised learning algorithms of SVM and RF, with HS and MHS hybridized together to improve model performances.

3.1. HS-SVM and MHS-SVM

SVM seeks for an optimal hyperplane with a maximum margin as the decision boundary to separate the two different classes. Given a training set with labelled instance pairs ( x i , y i ) , where i is the number of instance i = 1 , 2 , 3 , , m , x i R n and y i { 1 , + 1 } . The decision boundary to separate two different classes in SVM is generally expressed as w · x i + b = 0 , which is the dot product between the weight vector w and data instances with the bias b.
The optimal hyperplane is found by solving the convex optimization problem as in Equation (3). The ϵ i is the slack variable introduced to account for misclassification, with C as the accompanied penalty cost. To handle non-linearity, this study utilizes SVM with the Radial Basis Function (RBF) kernel, exp { γ | | x i x j | | 2 } . Hence, the hyperparameters to be tuned for RBF-SVM are C and g a m m a .
m i n ϕ ( w , b ) = 1 2 | | w | | 2 + C i = 1 m ϵ i s . t . y i ( w · x i + b ) 1 .
HS-SVM and MHS-SVM are utilized to search for features subset and hyperparameters that can maximize the AUC of SVM. The full procedure of HS-SVM and MHS-SVM, as well as their differences are detailed as follows:
Step 1:
Define objective function and parameters of HS and MHS.
The objective function is to maximize the AUC of the SVM classification function with three decision variables. The first decision variable, x 1 is a binary (0,1) string of length a (number of features in dataset), second ( x 2 ) and third ( x 3 ) decision variables correspond to the SVM hyperparameters search range log 2 C = [ 5 , 12 ] and log 2 γ = [ 12 , 5 ] [38], respectively. The detailed parameters settings are enclosed in Section 4.3.
Step 2:
Initialization of Harmony Memory
Each harmony vector in H M has three decision variables. Every harmony vector is evaluated with the fitness function and sorted from the best to worst. Each decision variable is randomly initialized as in Equation (4). Both HS-SVM and MHS-SVM have the same H M .
x 1 j = b i n a r y ( 0 , 1 ) , x 2 j = m i n ( x 2 ) + U ( 0 , 1 ) × ( m a x ( x 2 ) m i n ( x 2 ) ) , x 3 j = m i n ( x 3 ) + U ( 0 , 1 ) × ( m a x ( x 3 ) m i n ( x 3 ) ) .
Step 3:
Improvisation
With probability H M C R , a new harmony is selected from the H M . The selected harmony is adjusted to the neighbouring values with a probability P A R . The two continuous variables ( x 2 , x 3 ), the hyperparameters of SVM are adjusted to neighbouring values of width b w . With probability 1 H M C R , a new harmony vector is generated as in Equation (4).
However, for the first decision variable, x 1 (which is the features), P A R operator acts as a flipping agent. When it is activated, the selected harmony will be flipped from 1 to 0 or vice versa. Note that not every feature is flipped as our aim is to adjust the harmony rather than randomize the harmony. The higher the fraction of features flipped, the more randomized is the harmony, causing it to resembles exploration instead of exploitation of the features, and altogether resulting in higher computational effort as the search process continuously explore other search space. Features fraction of more than half is considered as high randomization. On the other hand, the lower the fraction of features flipped, the lesser the harmony is being exploited. To ensure the functionality of the P A R operator as the exploitation tool, the midpoint between zero and half of the features to be flipped is selected. Thus, only a quarter of the features is flipped. This is controlled by f l i p , a random vector generating the feature numbers to be flipped.
MHS-SVM will have the three modifications, i.e., dynamic H M C R and P A R following the step function in Equation (5), elitism selection, and replacement of b w with c o e f .
( H M C R i t e r , P A R i t e r ) = ( 0.70 , 0.35 ) 1 i t e r < 20 ( 0.75 , 0.30 ) 20 i t e r < 40 ( 0.80 , 0.25 ) 40 i t e r < 60 ( 0.85 , 0.20 ) 60 i t e r < 80 ( 0.90 , 0.15 ) 80 i t e r < 100 ( 0.95 , 0.10 ) i t e r 100 .
The improvisation procedure for HS-SVM and MHS-SVM are summarized in Algorithms 2 and 3, respectively.
Step 4:
Update H M by replacing the worst solution in H M with the new harmony if it has a better fitness value. This procedure is the same for HS-SVM and MHS-SVM.
Step 5:
Repeat Steps 3 and 4 until m a x _ i t e r is reached for HS-SVM and for MHS-SVM when one of the two additional criteria, i.e., H M converges or n o _ c o n s _ i m p is reached.
Algorithm 2
if ( U ( 0 , 1 ) H M C R )
        i n d = integer ( U ( 0 , 1 ) × H M S ) + 1
        x 1 = x 1 i n d
       if ( U ( 0 , 1 ) P A R )
               f l i p = r a n d ( 1 : a , 0.25 × a )
               x 1 = 1 x 1 i n d ( f l i p )
else
        x 1 = b i n a r y ( 0 , 1 )
if ( U ( 0 , 1 ) H M C R )
        i n d = integer ( U ( 0 , 1 ) × H M S ) + 1
        x 2 = x 2 i n d
       if ( U ( 0 , 1 ) P A R )
               x 2 = x 2 i n d + U ( 1 , 1 ) × b w
else
        x 2 = m i n ( x 2 ) + U ( 0 , 1 ) × ( m a x ( x 2 ) m i n ( x 2 ) )
if ( U ( 0 , 1 ) H M C R )
        i n d = integer ( U ( 0 , 1 ) × H M S ) + 1
        x 3 = x 3 i n d
       if ( U ( 0 , 1 ) P A R )
               x 3 = x 3 i n d + U ( 1 , 1 ) × b w
else
        x 3 = m i n ( x 3 ) + U ( 0 , 1 ) × ( m a x ( x 3 ) m i n ( x 3 ) )
Note: a,b: Refer Equation (5)        c: Refer Algorithm 1        d: Refer Equation (2)
Algorithm 3
H M C R = H M C R i t e r a
P A R = P A R i t e r b
if ( U ( 0 , 1 ) H M C R )
        i n d = selection ( )c
        x 1 = x 1 i n d
       if ( U ( 0 , 1 ) P A R )
               f l i p = r a n d ( 1 : a , 0.25 × a )
               x 1 = 1 x 1 i n d ( f l i p )
else
        x 1 = b i n a r y ( 0 , 1 )
if ( U ( 0 , 1 ) H M C R )
        i n d = selection ( )c
        x 2 = x 2 i n d
       if ( U ( 0 , 1 ) P A R )
               x 2 = x 2 i n d ± C V 2 d
else
        x 2 = m i n ( x 2 ) + U ( 0 , 1 ) × ( m a x ( x 2 ) m i n ( x 2 ) )
if ( U ( 0 , 1 ) H M C R )
        i n d = selection ( )c
        x 3 = x 3 i n d
       if ( U ( 0 , 1 ) P A R )
               x 2 = x 2 i n d ± C V 2 d
else
        x 3 = m i n ( x 3 ) + U ( 0 , 1 ) × ( m a x ( x 3 ) m i n ( x 3 ) )
Note: a,b: Refer Equation (5)        c: Refer Algorithm 1        d: Refer Equation (2)

3.2. HS-RF and MHS-RF

Random Forest is an ensemble model with a collection of decision trees using the bootstrap aggregation technique. Trees are grown using a binary splitting algorithm with Gini Impurity, GI = 1 i = 1 k p i 2 as the splitting criteria; where i is the number of classes and p i is the proportion of instances belonging to the respective class. During the tree growing process, to avoid correlations in between the trees, only a subset of the variables are required for splitting. The end result of the classification is based on the majority of votes from all the collected trees in the forest. The two hyperparameters to be tuned in RF are the number of trees ( n t r e e ) and number of variables available for splitting ( m t r y ). HS-RF and MHS-RF are utilized to search for hyperparameters that can maximize the AUC of RF. The full procedure of HS-RF and MHS-RF, as well as their differences are detailed as follows:
Step 1:
Define objective function and parameters of HS and MHS.
The objective function is the RF classification function with two decision variables that corresponds to the two hyperparameters, i.e., n t r e e and m t r y . The search range for n t r e e is chosen to be discrete values of x 1 { 1 , 5 } , where these values are then converted to the corresponding hundred. This search range is selected as it is often attempted by researchers. The search range of the second decision variable is discrete values of x 2 { 1 , a } , where a is the total number of attributes available. This search range is chosen because the hyperparameter m t r y is the random subset of variables from the total available attributes. The detailed parameters are enclosed in Section 4.3.
Step 2:
Initialization of Harmony Memory
Each harmony vector in H M has two decision variables. Every harmony vector is evaluated with the fitness function and sorted from the best to worst. Since the decision variables to solve RF are discrete, the harmony vectors are sampled directly from the search range as in Step 1. Both HS-RF and MHS-RF have the same H M .
Step 3:
Improvisation
With probability H M C R , a new harmony is selected from H M . Then the selected harmony is adjusted to the neighbouring values with probability P A R . As there are only discrete variables, the new harmony is adjusted directly to the left or right; b w is not required to adjust the new harmony. Hence only two modifications are involved in MHS-RF, i.e., dynamic H M C R and P A R following Equation (6) and the elitism selection.
( H M C R i t e r , P A R i t e r ) = ( 0.70 , 0.35 ) 1 i t e r < 5 ( 0.75 , 0.30 ) 5 i t e r < 10 ( 0.80 , 0.25 ) 10 i t e r < 15 ( 0.85 , 0.20 ) 15 i t e r < 20 ( 0.90 , 0.15 ) 20 i t e r < 25 ( 0.95 , 0.10 ) i t e r 25 .
The improvisation procedures for HS-RF and MHS-RF are summarized in Algorithms 4 and 5, respectively.
Algorithm 4
fori in (1:2)
       if ( U ( 0 , 1 ) H M C R )
               i n d = i n t ( U ( 0 , 1 ) × H M S ) + 1
               x i = x i i n d
              if ( U ( 0 , 1 ) P A R )
                      x i = x i i n d ± 1
       else
               x i { m i n ( x i ) , m a x ( x i ) }
Note: a,b: Refer Equation (6)        c: Refer Algorithm 1
Algorithm 5
H M C R = H M C R i t e r a
P A R = P A R i t e r b
fori in (1:2)
       if ( U ( 0 , 1 ) H M C R )
               i n d = selection ( )c
               x i = x i i n d
              if ( U ( 0 , 1 ) P A R )
                      x i = x i i n d ± 1
       else
               x i { m i n ( x i ) , m a x ( x i ) }
Note: a,b: Refer Equation (6)        c: Refer Algorithm 1
Step 4:
Update H M by evaluating and comparing the fitness function of the new harmony with the worst harmony in H M . Replace the worst harmony if the new harmony has better fitness value. This procedure is the same for both HS-RF and MHS-RF.
Step 5:
Repeat Steps 3 and 4 until m a x _ i t e r is reached for HS-RF and for MHS-RF when one of the two additional criteria, i.e., H M converges or n o _ c o n s _ i m p is reached.

3.3. Parallel Computing

Both MHS-SVM and MHS-RF aim for quality results but faster convergence. Parallel computing with master-slave concept can be employed on the 10 independent tasks (from cross validation) to enhance the computational efficiency. Initially, the master generates sub-tasks via data preparation and splitting to be assigned to 10 slaves for independent and simultaneous execution. When done, each slave returns the required performance measures (refer Section 4.2 to compute the average. Algorithm 6 summarizes the parallel computation. Since the main aim is to save computational time, the same seeding is applied for both sequential and parallel execution to ensure identical model performance.
Algorithm 6
Master: Data preparation and partitioning
do_parallel
       for i in (1:10)
               Slave: Step 1-5 of MHS-RF
return AUC, ACC, ACC*
Master: mean(AUC), mean (ACC), mean (ACC*)

4. Experimental Setup

4.1. Credit Datasets Preparation

The datasets used in the experiments are the German and Australian datasets which are publicly available at the UCI repository (https://archive.ics.uci.edu/). Additionally, a peer-to-peer lending dataset downloaded from the Lending Club (LC) website (https://www.lendingclub.com/info/download-data.action) is also included.
For the experiment, only the sample of 60-month-term of the year 2012 is taken because less attention was given on the 60-month-term loan in the past literature. To prepare the LC dataset, this experiment focuses only on loan status that are fully paid and charged off. Variables having all empty values or more than 5% missing values are removed, and variables with less than 1% of missing value have the whole instance being removed as it is only a small loss of information. Missing data is imputed with the mean for numerical and mode for categorical attributes, respectively. Table 4 gives a summary of the datasets. Attributes descriptions for German and Australian are available online while the brief descriptions of the LC attributes are shown in Table 5.
Numerical attributes are standardized by subtracting the column mean and dividing the standard deviation. Categorical attributes are converted to numerical attributes with the weight-of-evidence (WOE) transformation. 10-fold cross validation is applied on the datasets, and a validation set is prepared for the hyperparameter tuning procedure to avoid the overfitting problem. In the experiment, the German and Australian datasets are relatively small, thus the validation set is an inner 5-fold cross validation, whereas the relatively larger LC dataset has a holdout set as the validation set.

4.2. Performance Measures

This study utilizes both threshold-variant and threshold-invariant performance measures to evaluate the model performances. Accuracy (ACC) and the F1 score (F1) are reported at the default threshold at the cutoff probability of 0.5. ACC is the proportion of correctly classified instances in the data. For a more reliable estimate when there is class imbalance, F1 computes the harmonic mean of precision and recall is reported together for model evaluation. The threshold-invariant measure, AUC gives a better picture on the discriminating ability of a model across all possible thresholds. The Friedman test is conducted to test the significance of AUC between the compared models across the 10 test sets (from cross validation) for each dataset. The Wilcoxon signed rank test is applied if there is a significant difference reported from Friedman test.
This study assigns a positive sign to non-defaulting customers and a negative sign to defaulting customers. The Type I error represents acceptance of an actual defaulting customer whereas Type II error represents the rejection of an actual non-defaulting customer. Both types of errors result in a different extent of losses, depending on the financial environment of the institution. Hence, a different cutoff probability is usually adjusted to achieve a balance in between both types of errors. High sensitivity (SEN) and specificity (SPE) are equivalent to low Type II and Type I error, respectively. SEN and SPE are reported at the cutoff probability of 0.5 for a further discussion on the model performance in achieving a balance between both error types.

4.3. Models Setup

To assess the performance of the proposed models, Logistic Regression (LOGIT), Backward Stepwise Logistic Regression (STEP) and Linear Discriminant Analysis (LDA) are included for comparison as they are the standard statistical models in the credit scoring domain. The standard SVM and RF are tuned with the conventional GS, using the same grid points described in Section 2.2. Considering the extensive computational effort due to the cross validation setup explained in Section 4.1, only a coarse GS is conducted for SVM. Thus, there are five comparison models to be compared with the proposed models.
The detailed parameters settings of all hybrid models are shown in Table 6. HS and MHS hybrid models have their parameters set in the same way as in the numerical experiment described in Section 2.2 and Section 2.3. Hence, across the three datasets, HS hybrid models have different parameters settings whereas MHS hybrid models save up the effort of repeated trial-and-error due to the modification of dynamic H M C R and P A R step function. The MHS hybrid models step function for H M C R and P A R are setup as in Equations (5) and (6) for MHS-SVM and MHS-RF, respectively.
The proposed models are coded in R 3.5.1 and executed on a 2.70 GHz Intel(R) Core(TM) i7-7500 CPU with 4.00 GB RAM under Windows 10 operating system. For parallel computation, the parallel environment is initiated with the ‘doParallel’ library in R 3.2.5 and executed on a Linux based operating system using IBM system X360 M4 server with ten nodes of 2.0 GHz Intel Xeon 6C processors.

5. Results and Discussions

This section reports the experimental results obtained from the different credit scoring models across the three credit datasets based on model performances, model explainability and computational time.

5.1. Model Performances

Table 7 reports the models’ performances across the three datasets. For the German and Australian dataset, the AI models are competitive with the statistical models with only a slight performance difference within 2%. On the other hand, for the LC dataset, the AI models consistently outperformed the statistical models. This indicates the flexibility of the AI models to account for various data patterns. Focusing in the SVM and RF families, the proposed hybrid models have slightly improved AUC compared to the GS-tuned models. While the hybrid models do not show consistent improvement of ACC and F1 compared to GS approach, the performance difference has been maintained in a less than 1% margin. Hence, the reported performance measures have implied the hybrid models are very competitive when compared to the GS tuning method.
Based on the three performance measures, the SVM family models have a wider gap of performance difference than the RF family models. This is due to the functionality of the HS-SVM and MHS-SVM to conduct simultaneous feature selection with hyperparameter tuning at a smaller granularity than the GS approach. Therefore, HS-SVM and MHS-SVM will have a different input features subset with the GS-tuned SVM that utilized the full features. In addition, the ability of HS-SVM and MHS-SVM to directly search the continuous hyperparameters space also results in a slightly better performance than the coarse GS for SVM tuning in this experiment. HS-RF and MHS-RF report only very slight performance difference with GS tuned RF because no feature selection is conducted and the hyperparameters are discrete which results in the same search space for the three models.
ACC and F1 are threshold-variant performance measures that will change depending on the threshold settings. Hence, the Friedman statistical test is only conducted based on the threshold-invariant AUC, reported in the last row of Table 7, with the respective p-values enclosed in the parentheses. For the German and Australian datasets, despite the numerical differences, the Friedman tests do not show statistical significant differences between all the experimented models. For the LC dataset, the Friedman test shows statistically significant differences between the models. The corresponding post-hoc test with the p-values is tabulated in Table 8. The pairs that show significant differences at α = 0.01 are marked in bold text.
The post-hoc Wilcoxon-signed ranked test shows statistically significant better AUC performance of both SVM and RF families than the statistical models. There is significant difference among the models in the SVM families, indicating the significant improvement of the proposed hybrid SVM models compared to the GS-tuned SVM. While there is no significant difference among the models from RF family, the slight difference in performance indicates that the proposed hybrid RF models are competitive to the GS-tuned RF. There is significant difference is reported between models from SVM and RF family, with RF family models having better performance.
Among the three performance measures, only AUC indicates consistent best performance from the RF family models and consistent improvements of proposed hybrid models compared to the GS tuning approach. Considering the ACC and F1, the performance ranking of the models are different across the three datasets. To consider all the performance measures together for a general overview evaluation of the models, each model is assigned an overall rank (ORank). In each dataset, the models are ranked based on their average rank computed across the three performance measures. The same reported performance will have a tied rank. All the rankings are tabulated in Table 9, with a lower value of ORank indicating better model performance.
According to the ORank, the RF family models take the best rank, followed by the SVM family models and lastly the statistical models. This order indicates the robustness of AI models compared to the statistical models. In addition, for both the RF and SVM family models, the proposed hybrid models always have a better ORank than the GS approach. Thus, HS is suitable to be hybridized with AI models for cautious hyperparameter tuning, and also feature selection; particularly SVM in this study. The hybrid models do not require specific settings of the granularity as in GS for continuous decision variables and at the same time is able to perform feature selection within the same time as a GS approach which only conducts hyperparameter tuning. The competitive performance of the proposed MHS hybrid models show that HS is very adaptable depending on user needs. MHS hybrid models effectively reduce computational effort yet at the same time maintains the quality of the solution (detailed discussions in Section 5.3).
Table 10 reports the model sensitivity and specificity across the three datasets. Instead of evaluating both measures separately, this study discusses both measures together as one single pair since a good model should not have dominance in only either one of it to achieve a balanced trade off between the two types of losses.
For the German and Australian dataset, both statistical and AI models have reported relatively similar sensitivity-specificity gaps, indicating a reasonable balance between Type II-Type I error. Models from the RF and SVM families have slight inclined priority towards a reduction of Type II error (due to higher sensitivity and lower specificity) in German and Australian datasets, respectively. For the LC dataset, the AI models report significant smaller sensitivity-specificity gaps compared to the statistical models. In contrast to statistical models that have extreme dominance in sensitivity, AI models have reported a reasonable balance between sensitivity and specificity, indicating that AI models have a better balance to reduce both Type I and Type II errors for the LC dataset. For the models of the RF family, the proposed hybrid models do not show consistent improvement across the three datasets compared to the GS-tuned RF but the reported performance have only a very slight difference, indicating the proposed models are competent. On the other hand, for the models from the SVM family, the proposed hybrid models have a slight improvement in the German and Australian dataset but significant improvement in the LC dataset compared to the GS-tuned SVM. This shows that HS and MHS have effectively improved the SVM performance by simultaneous feature selection and hyperparameter tuning.
Several recent literatures are outlined in Table 11 to highlight the main paradigm shift towards the usage of advanced AI techniques in credit scoring and the recent approach in credit scoring performance evaluation. Note that the abbreviations in Table 11 shall be referred to the original studies. Recent studies have much attention paid on advanced non-linear classifiers [29] and ensemble models, with tree-based ensembles [31,32] showing great potential because decision tree is perceived as the conventional classifier technique [33]. Besides, performance measures that is able to reflect the ability of the model in handling class imbalance is the recent trend where [33] have highlighted the usage of expected cost together with class imbalance for model evaluation and [29,30,31] have employed SEN, SPE and AUC for model evaluation. The summary from Table 11 implies the alignment of this study to fit with the recent paradigm, i.e., improvement of non-linear SVM and tree-based ensemble RF via a hybrid approach as well as model evaluation that takes into account for class imbalance via discussion on SEN and SPE.
Based on the summary in Table 11 and Table 12 compiled the studies that have experimented on the same dataset and utilized the same performance measures for results comparison with this study. Hence, only studies from [30,32] are included as comparison. Since [30] has proposed a novel approach in ranking the assessed models, only the top three models with the best rank are compiled. Due to the different assignment of positive sign for defaulting customers by [30], the true positive rate and true negative rate reported would be analogous to the SPE and SEN, respectively, in this study. For the study by [32], Australian dataset is the only data in common with our study with error rate as the single performance measure, thus ACC is computed from the error rate and reported in Table 12. Performance measures in bold texts indicate the best performance within that particular study.
The compiled results show that there is no obvious outperformance between the results reported by external studies with the proposed models. Across the two datasets over the four performance measures, the margin of difference have been maintained within 5% difference, which could not be considered as significant performance difference. This indirect comparison with external studies implies the competitiveness of the proposed models with the latest state-of-the-arts. It is worth to emphasize that model performance comparison with external studies would be difficult due to different experiment setup and varying proposed approaches to address different issues, the comparison in Table 12 aims to indicate the competency of the proposed models with current techniques instead of identifying a ‘winner’ among these experiments.

5.2. Model Explainability

For model explainability, the HS-SVM and MHS-SVM conduct feature selection and hyperparameter tuning simultaneously. The end user can then focus on the investigation of the reduced features subset. Table 13 reports the average number of reduced features across the 10-fold test sets. HS-SVM and MHS-SVM are compared only with STEP because STEP is the only model that conducted feature selection.
From Table 13, there is only a slight difference of the average number of reduced features between the three models for German and Australian datasets. However, for LC datasets, hybrid SVM models have more reduced features compared to STEP. For all the three datasets, the proposed hybrid SVM models have effectively reduced the features while maintaining a good performance as compared to the standard SVM that used the full features. This indicates that the proposed hybrid SVM models effectively reduced the original features but yet improved the standard SVM model.
For the RF models, this study recommends the use of the computed feature importance, i.e., the mean decrease in accuracy (mDA) and the mean decrease in Gini Impurity (mGI) for model explainability. Both mDA and mGI ranks the features from most to least important, thus providing the initial insight for the end user.

5.3. Computational Time

Table 14 reports the computational time of all the models utilized in Section 5.1, including the two parallel MHS hybrid models. Note that the performance measures for these two parallel hybrid models are identical, with only a difference in computational time, because of the same seeding applied.
Across the three datasets, the statistical models are very efficient, with only the STEP taking a longer time due to the feature selection process. The AI models are time-consuming due to the hyperparameter tuning process which is unavoidable as they are sensitive to the hyperparameters’ choice.
For both the SVM and RF families, similar computational effort can be perceived. In the experiments using the three datasets, the HS-hybrid models take the longest time, as compared to GS since it depends on the only termination criteria, m a x _ i t e r in the HS procedure to ensure the search space is sufficiently explored. In contrast, MHS-hybridized models are able to search for comparable solutions as with the HS and GS, but in a much shorter time. This saves up to half of the computational effort. Together with parallel computing, the MHS hybrid models have more efficient computational power. Nonetheless, the development of the MHS hybrid models require experimentation and additional computational efforts. Between both the SVM and RF families, the SVM models are extremely time-consuming when the dataset contains more instances, i.e., the LC dataset. This is due to the training time complexity of SVM O ( n 3 ) .
Despite the benefit of being efficient without the need of hyperparameter tuning, the standard statistical models face limitations in dealing with more complex data patterns. In cases where the standard statistical models can no longer account for the data pattern, this results in poor performance. Further data transformation or interaction terms have to be included for the statistical models building procedure. This additional procedure may be another time-consuming process.

6. Conclusions and Future Directions

In this study, HS and MHS are hybridized with both SVM and RF, forming four new models. The newly proposed MHS is to ensure an effective yet efficient searching process. HS-SVM and MHS-SVM tune hyperparameters and reduces the features for model explanation while HS-RF and MHS-RF tune hyperparameters and utilize the two types of feature importance for model explainability. This allows flexibility in modeling while having high accuracy in the classification, comparative to traditional statistical modeling. In addition to this, the computational time is also competitive.
All the proposed models, HS-SVM, MHS-SVM, HS-RF, and MHS-RF, are competitive in the German and Australian datasets, and have great improvement over the standard statistical models in the LC dataset. All the HS and MHS hybrid models consistently reported higher AUC than the standard SVM and RF, implying the effectiveness of the proposed hybrid models to improve model discriminating ability. The proposed models only show slight improvement without significant difference in the German and Australian datasets, but the RF family models have shown statistically significant better results in LC dataset, with hybrid RF models reporting the best performance.
HS-SVM and MHS-SVM have effectively shrunk down the number of features, enabling end users to focus on the reduced features. HS-RF and MHS-RF are well-tuned, thus the computed feature importance are believed to be reliable in ranking the features. These strategies can be useful in providing initial insight for the end users.
In terms of computational effort, the standard statistical models are efficient as there is no hyperparameter tuning procedure required, but they fail to achieve good performance in the LC dataset. HS hybrid models are time consuming while MHS hybrid models have shown significant time saving. Along with parallel computing, the computational effort is further reduced. MHS hybrid models are very competitive compared to HS hybrid models, with the benefit of the computational efficiency being improved. In consideration of good discriminating ability, model explainability and computational efficiency, MHS-RF is the recommended alternative credit scoring model.
There are some possible future directions that can be pointed out. Instead of the time consuming standard SVM, other versions of SVM such as Least Squares SVM can be attempted to form a hybrid model. The model explainability approach in this study does not solve the black-box property of the AI models. Rules extraction can be incorporated to solve the black-box problem.

Author Contributions

Conceptualization, L.S.L. and H.-V.S.; methodology, R.Y.G., L.S.L. and H.-V.S.; software, R.Y.G. and K.G.; validation, R.Y.G., L.S.L. and H.-V.S.; writing–original draft preparation, R.Y.G., L.S.L., H.-V.S. and K.G.; writing–review and editing, R.Y.G., L.S.L. and H.-V.S.; supervision, L.S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Geran Putra-Inisiatif Putra Siswazah (GP-IPS/2018/9646000) (Universiti Putra Malaysia).

Acknowledgments

The authors would like to thank reviewers for their time to thoroughly review and provide constructive comments for improvements of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest. The authors alone are responsible for the content and writing of the paper. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
ABCArtificial Bee Colony
ACCAccuracy
AFAverage Frequency
AIArtificial Intelligence
AUCArea Under Receiver Operating Characteristics
GAGenetic Algorithm
GSGrid Search
HMHarmony Memory
HMCRHarmony Memory Considering Rate
HMSHarmony Memory Size
HSHarmony Search
LDALinear Discriminant Analysis
LOGITLogistic Regression
MAMetaheuristic Algorithm
mDAmean Decrease Accuracy
mGImean decrease in Gini Impurity
MHSModified Harmony Search
PARPitch-Adjusting Rate
PSOParticle Swarm Optimization
RFRandom Forest
SENSensitivity
SPESpecificity
STEPBackward Stepwise Logistic Regression
SVMSupport Vector Machine

References

  1. Baesens, B.; Van Gestel, T.; Viaene, S.; Stepanova, M.; Suykens, J.; Vanthienen, J. Benchmarking state-of-the-art classification algorithms for credit scoring. J. Oper. Res. Soc. 2003, 54, 627–635. [Google Scholar] [CrossRef]
  2. Lessmann, S.; Baesens, B.; Seow, H.V.; Thomas, L.C. Benchmarking state-of-the-art classification algorithms for credit scoring: An update of research. Eur. J. Oper. Res. 2015, 247, 124–136. [Google Scholar] [CrossRef] [Green Version]
  3. Goh, R.Y.; Lee, L.S. Credit scoring: A review on support vector machines and metaheuristic approaches. Adv. Oper. Res. 2019, 2019, 30. [Google Scholar] [CrossRef]
  4. Yu, L.; Yao, X.; Wang, S.; Lai, K.K. Credit risk evaluation using a weighted least squares SVM classifier with design of experiment for parameter selection. Expert Syst. Appl. 2011, 38, 15392–15399. [Google Scholar] [CrossRef]
  5. Zhou, L.; Lai, K.K.; Yu, L. Credit scoring using support vector machines with direct search for parameters selection. Int. J. Inf. Technol. Decis. Mak. 2009, 13, 149–155. [Google Scholar] [CrossRef]
  6. Danenas, P.; Garsva, G. Selection of support vector machines based classifiers for credit risk domain. Expert Syst. Appl. 2015, 42, 3194–3204. [Google Scholar] [CrossRef]
  7. Garsva, G.; Danenas, P. Particle swarm optimization for linear support vector machines based classifier selection. Nonlinear Anal. Model. Control 2014, 19, 26–42. [Google Scholar] [CrossRef]
  8. Hsu, F.J.; Chen, M.Y.; Chen, Y.C. The human-like intelligence with bio-inspired computing approach for credit ratings prediction. Neurocomputing 2018, 279, 11–18. [Google Scholar] [CrossRef]
  9. Gorter, D. Added Value of Machine Learning in Retail Credit Risk. Ph.D. Thesis, Financial Engineering and Management, University of Twente, Enschede, The Netherlands, 2017. [Google Scholar]
  10. Òskarsdòttir, M.; Bravo, C.; Sarraute, C.; Vanthienen, J.; Baesens, B. The value of big data for credit scoring: Enhancing financial inclusion using mobile phone data and social network analytics. Appl. Soft Comput. 2019, 74, 26–39. [Google Scholar] [CrossRef] [Green Version]
  11. Van Sang, H.; Nam, N.H.; Nhan, N.D. A novel credit scoring prediction model based on feature selection approach and parallel random forest. Indian J. Sci. Technol. 2016, 9, 1–6. [Google Scholar] [CrossRef] [Green Version]
  12. Tang, L.; Cai, F.; Ouyang, Y. Applying a nonparametric random forest algorithm to assess the credit risk of the energy industry in China. Technol. Forecast. Soc. Chang. 2019, 144, 563–572. [Google Scholar] [CrossRef]
  13. Ye, X.; Dong, L.A.; Ma, D. Loan evaluation in P2P lending based on random forest optimized by genetic algorithm with profit score. Electron. Commer. Res. Appl. 2018, 32, 23–36. [Google Scholar] [CrossRef]
  14. Malekipirbazari, M.; Aksakalli, V. Risk assessment in social lending via random forests. Expert Syst. Appl. 2015, 42, 4621–4631. [Google Scholar] [CrossRef]
  15. Jiang, C.; Wang, Z.; Zhao, H. A prediction-driven mixture cure model and its application in credit scoring. Eur. J. Oper. Res. 2019, 277, 20–31. [Google Scholar] [CrossRef]
  16. He, H.; Zhang, W.; Zhang, S. A novel ensemble method for credit scoring: Adaption of different imbalance ratios. Expert Syst. Appl. 2018, 98, 105–117. [Google Scholar] [CrossRef]
  17. Jadhav, S.; He, H.; Jenkins, K. Information gain directed genetic algorithm wrapper feature selection for credit rating. Appl. Soft Comput. 2018, 69, 541–553. [Google Scholar] [CrossRef] [Green Version]
  18. Wang, D.; Zhang, Z.; Bai, R.; Mao, Y. A hybrid system with filter approach and multiple population genetic algorithm for feature selection in credit scoring. J. Comput. Appl. Math. 2018, 329, 307–321. [Google Scholar] [CrossRef]
  19. Huang, C.L.; Chen, M.C.; Wang, C.J. Credit scoring with a data mining approach based on support vector machines. Expert Syst. Appl. 2007, 33, 847–856. [Google Scholar] [CrossRef]
  20. Zhou, L.; Lai, K.K.; Yen, J. Credit scoring models with AUC maximization based on weighted SVM. Int. J. Inf. Technol. Decis. Mak. 2009, 8, 677–696. [Google Scholar] [CrossRef]
  21. Yeh, C.C.; Lin, F.; Hsu, C.Y. A hybrid KMV model, random forests and rough set theory approach for credit rating. Knowl.-Based Syst. 2012, 33, 166–172. [Google Scholar] [CrossRef]
  22. Huang, Y.F.; Wang, S.H. Movie genre classification using SVM with audio and video features. In International Conference on Active Media Technology; Springer: Berlin/Heidelberg, Germany, 2012; pp. 1–10. [Google Scholar]
  23. Ceylan, O.; Taskın, G. SVM parameter selection based on harmony search with an application to hyperspectral image classification. In Proceedings of the 2016 24th Signal Processing and Communication Application Conference (SIU), Zonguldak, Turkey, 16–19 May 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 657–660. [Google Scholar]
  24. Nekkaa, M.; Boughaci, D. Hybrid harmony search combined with stochastic local search for feature selection. Neural Process. Lett. 2016, 44, 199–220. [Google Scholar] [CrossRef]
  25. Sridharan, K.; Komarasamy, G. Sentiment classification using harmony random forest and harmony gradient boosting machine. Soft Comput. 2020, 24, 7451–7458. [Google Scholar] [CrossRef]
  26. Yusup, N.; Zain, A.M.; Latib, A.A. A review of harmony search algorithm-based feature selection method for classification. J. Phys. Conf. Ser. 2019, 1192, 012038. [Google Scholar] [CrossRef]
  27. Krishnavei, V.; Santhiya, K.G.; Ramesh, P.; Jaganathan, S. A credit scoring prediction model based on harmony search based 1-NN classifier feature selection approach. Int. J. Adv. Res. Sci. Eng. 2018, 7, 468–476. [Google Scholar]
  28. Geem, Z.W.; Kim, J.H.; Loganathan, G.V. A new heuristic optimization algorithm: Harmony search. Simulation 2001, 76, 60–68. [Google Scholar] [CrossRef]
  29. Lahmiri, S.; Bekiros, S. Can machine learning approaches predict corporate bankruptcy? Evidence from a qualitative experimental design. Quant. Financ. 2019, 19, 1569–1577. [Google Scholar] [CrossRef]
  30. Peng, Y.; Wang, G.; Kou, G.; Shi, Y. An empirical study of classification algorithm evaluation for financial risk prediction. Appl. Soft Comput. 2011, 11, 2906–2915. [Google Scholar] [CrossRef]
  31. Zięba, M.; Tomczak, S.K.; Tomczak, J.M. Ensemble boosted trees with synthetic features generation in application to bankruptcy prediction. Expert Syst. Appl. 2016, 58, 93–101. [Google Scholar] [CrossRef]
  32. Lahmiri, S.; Bekiros, S.; Giakoumelou, A.; Bezzina, F. Performance assessment of ensemble learning systems in financial data classification. Intell. Syst. Account. Financ. Manag. 2020, 27, 3–9. [Google Scholar] [CrossRef]
  33. Drummond, C.; Holte, R.C. C4.5, class imbalance, and cost sensitivity: Why under-sampling beats over-sampling. In Workshop on Learning from Imbalanced Datasets II; Citeseer: Washington, DC, USA, 2003; Volume 11, pp. 1–8. [Google Scholar]
  34. Yang, X.S. Harmony search as a metaheuristic algorithm. In Music-Inspired Harmony Search Algorithm: Theory and Applications; Geem, Z.W., Ed.; Springer: Berlin, Germany, 2009; Volume 191, pp. 1–14. [Google Scholar]
  35. Mukhopadhyay, A.; Roy, A.; Das, S.; Das, S.; Abraham, A. Population-variance and explorative power of harmony search: An analysis. In Proceedings of the Third International Conference on Digital Information Management (ICDIM 2008), London, UK, 13–16 November 2008; IEEE: Piscataway, NJ, USA; pp. 775–781. [Google Scholar]
  36. Kalivarapu, J.; Jain, S.; Bag, S. An improved harmony search algorithm with dynamically varying bandwidth. Eng. Optim. 2016, 48, 1091–1108. [Google Scholar] [CrossRef]
  37. Tuo, S.; Yong, L.; Deng, F.A. A novel harmony search algorithm based on teaching-learning strategies for 0-1 knapsack problems. Sci. World J. 2014, 637412. [Google Scholar] [CrossRef] [PubMed]
  38. Hsu, C.W.; Chang, C.C.; Lin, C.J. A Practical Guide to Support Vector Classification. Available online: http://www.csie.ntu.edu.tw/~cjlin/papers/guide/guide.pdf (accessed on 2 February 2020).
Figure 1. Search pattern of HS-SVM.
Figure 1. Search pattern of HS-SVM.
Entropy 22 00989 g001
Figure 2. Search pattern of HS-RF.
Figure 2. Search pattern of HS-RF.
Entropy 22 00989 g002
Figure 3. Search pattern comparison between HS-SVM with Modified HS (MHS-SVM).
Figure 3. Search pattern comparison between HS-SVM with Modified HS (MHS-SVM).
Entropy 22 00989 g003
Figure 4. Search pattern comparison between HS-RF with MHS-RF).
Figure 4. Search pattern comparison between HS-RF with MHS-RF).
Entropy 22 00989 g004
Table 1. Performance of Harmony Search (HS)-Support Vector Machines (SVM) and HS-Random Forest (RF) across small perturbations of the operators.
Table 1. Performance of Harmony Search (HS)-Support Vector Machines (SVM) and HS-Random Forest (RF) across small perturbations of the operators.
HS-SVMHS-RF
HMCRPARAUCTime (min)AUCTime (min)
0.700.100.81774.44810.825610.2604
0.700.150.81754.62110.827110.0100
0.700.200.82004.42030.827210.8761
0.700.250.81864.50080.830510.9019
0.700.300.81814.56140.825210.4817
0.700.350.81964.51220.825411.0123
0.750.100.81704.37730.825511.1008
0.750.150.81734.33230.825110.4266
0.750.200.82044.63570.828110.8729
0.750.250.81914.37020.82799.7096
0.750.300.81904.46400.825910.6458
0.750.350.81934.48100.826710.7759
0.800.100.81644.56140.825211.3625
0.800.150.81734.51980.825811.2265
0.800.200.81964.51380.82719.9028
0.800.250.81934.46980.83009.9752
0.800.300.81864.53570.826310.6068
0.800.350.81894.55410.82989.9372
0.850.100.81634.39970.825211.7332
0.850.150.81794.42580.825611.2218
0.850.200.81904.37160.824110.6202
0.850.250.81634.50280.827310.5372
0.850.300.81584.61220.826510.9275
0.850.350.81524.69280.825210.9275
0.900.100.81704.36480.827611.2554
0.900.150.81844.35020.826211.5087
0.900.200.81934.39500.824410.9275
0.900.250.81304.62320.827510.6972
0.900.300.81384.42450.827910.8565
0.900.350.81394.42600.825811.2159
0.950.100.81544.61320.825610.8480
0.950.150.81454.27170.821711.1239
0.950.200.81474.51640.824411.6584
0.950.250.81334.60130.828211.3777
0.950.300.81524.40700.827711.1307
0.950.350.81524.46300.827512.3353
mean0.81724.48170.826510.8754
sd0.00210.10000.00180.5785
Table 2. Comparison of Grid Search (GS) approach with HS hybrid models.
Table 2. Comparison of Grid Search (GS) approach with HS hybrid models.
AUCTime (min)
GS-SVM0.807823.9922
HS-SVM *0.81724.4817
GS-RF0.82149.0614
HS-RF *0.826510.8754
* mean results from Table 1.
Table 3. Comparison of MHS hybrid models with HS Hybrids and GS Approach.
Table 3. Comparison of MHS hybrid models with HS Hybrids and GS Approach.
AUCTime (min)Iterations
GS-SVM0.807823.9922614
HS-SVM *0.81724.4817100
MHS-SVM0.81973.150271
GS-RF0.82149.0614100
HS-RF *0.826510.8754100
MHS-RF0.82615.200849
* mean results from Table 1.
Table 4. Summary of benchmark datasets.
Table 4. Summary of benchmark datasets.
InstancesCategoricalNumericalDefault Rate
German100013730%
Australian6908644.45%
LC988741727.69%
Table 5. List of attributes in Lending Club (LC) dataset.
Table 5. List of attributes in Lending Club (LC) dataset.
AttributesTypeAttributesType
loan_amntNumericallast_credit_pull_d ***Numerical
emp_length *Numericalacc_now_delinqNumerical
annual_incNumericalchargeoff_within_12mthsNumerical
dtiNumericaldelinq_amntNumerical
delinq_2_yrsNumericalpub_rec_bankruptciesNumerical
earliest_cr_line **Numericaltax_liensNumerical
inq_last_6mthsNumericalhome_ownershipCategorical
open_accNumericalverification_statusCategorical
pub_recNumericalpurposeCategorical
revol_utilNumericalinitial_list_statusCategorical
total_accNumerical
The full name of the attributes details can be found in the LCDataDictionary.xls file in the LC website. * Transformed from categorical. ** Transformed to how many years since first credit line opened. *** Transformed to how many months since LC pulled credit.
Table 6. Parameters settings of HS and MHS hybrid models for the three credit datasets.
Table 6. Parameters settings of HS and MHS hybrid models for the three credit datasets.
ParametersGermanAustralianLending Club
HS-SVM H M S 303030
H M C R 0.700.800.80
P A R 0.300.100.30
b w 0.100.100.10
m a x _ i t e r 100010001000
MHS-SVM H M S 30
e l i t 0.70
H M C R {0.70, 0.95}
P A R {0.10, 0.35}
s t e p 20
m a x _ i t e r 1000
n o _ c o n s _ i m p 500
HS-RF H M S 101010
H M C R 0.700.700.80
P A R 0.300.300.10
m a x _ i t e r 100100100
MHS-RF H M S 10
e l i t 0.70
H M C R {0.70, 0.95}
P A R {0.10, 0.35}
s t e p 5
m a x _ i t e r 100
n o _ c o n s _ i m p 25
Table 7. Model performances of the proposed hybrid models with GS-tuned AI models and statistical models.
Table 7. Model performances of the proposed hybrid models with GS-tuned AI models and statistical models.
GermanAustralianLending Club
AUCACCF1AUCACCF1AUCACCF1
LOGIT0.79890.75900.83560.93080.87250.84620.62570.72390.8386
STEP0.79990.76200.83780.93210.87390.85500.62450.72380.8387
LDA0.80080.74700.83650.92860.86230.84730.62310.72380.8390
GS-SVM0.80060.74400.83150.92920.85360.84860.71680.72360.8393
HS-SVM0.80150.76200.84240.93130.86390.85790.82780.82510.8841
MHS-SVM0.80510.76200.84030.93100.85650.85240.82670.82030.8800
GS-RF0.79990.76400.84480.93540.87230.85980.86700.85800.9068
HS-RF0.80440.76400.84530.93660.87380.86140.86740.85710.9063
MHS-RF0.80530.75600.84100.93560.86950.85560.86790.85720.9064
χ friedman 2 = 4.3685 , (0.8224) χ friedman 2 = 5.8568 , (0.6633) χ friedman 2 = 75.05 , (4.826e-13)
Table 8. Post-hoc Wilcoxon-signed rank test for LC dataset of AUC performance.
Table 8. Post-hoc Wilcoxon-signed rank test for LC dataset of AUC performance.
LOGITSTEPLDASVMHS-SVMMHS-SVMRFHS-RFMHS-RF
LOGIT-
STEP1.31 × 10−1-
LDA8.40 × 10−23.23 × 10−1-
GS-SVM1.95 × 10−31.95 × 10−31.95 × 10−3-
HS-SVM1.95 × 10−31.95 × 10−31.95 × 10−31.95 × 10−3-
MHS-SVM1.95 × 10−31.95 × 10−31.95 × 10−31.95 × 10−33.23 × 10−1-
GS-RF1.95 × 10−31.95 × 10−31.95 × 10−31.95 × 10−31.95 × 10−31.95 × 10−3-
HS-RF1.95 × 10−31.95 × 10−31.95 × 10−31.95 × 10−31.95 × 10−31.95 × 10−36.25 × 10−1-
MHS-RF1.95 × 10−31.95 × 10−31.95 × 10−31.95 × 10−31.95 × 10−31.95 × 10−32.86 × 10−14.41 × 10−1-
Table 9. Ranking of model performances across the three credit datasets.
Table 9. Ranking of model performances across the three credit datasets.
GermanAustralianLending Club
AUCACCF1RankAUCACCF1RankAUCACCF1RankORank
LOGIT9687.77396.37797.77.3
STEP7.5465.84153.388.588.175.9
LDA5876.7978898.578.177.7
SVM6998897866666.4
HS-SVM4433.75634.744444.2
MHS-SVM2453.76866.755555.2
RF7.51.523.734233312.33.1
HS-RF31.511.831211.32232.31.9
MHS-RF17442543.71121.33.1
Table 10. Sensitivity and specificity analysis of the proposed hybrid models with GS-tuned AI models and statistical models.
Table 10. Sensitivity and specificity analysis of the proposed hybrid models with GS-tuned AI models and statistical models.
GermanAustralianLending Club
SENSPESENSPESENSPE
LOGIT0.87570.48670.86780.86290.99190.0462
STEP0.87860.49000.88480.85810.99300.0208
LDA0.87430.49670.91650.80700.99520.0150
GS-SVM0.90570.36670.91550.80430.99800.0069
HS-SVM0.90860.42000.91850.82000.92290.5695
MHS-SVM0.89570.45000.92520.80160.91300.5782
GS-RF0.91870.40000.86990.86740.95550.6034
HS-RF0.92290.39330.86340.88220.95640.5979
MHS-RF0.92290.36670.86350.87460.95650.5979
Table 11. Recent literature studies in credit scoring domain.
Table 11. Recent literature studies in credit scoring domain.
StudyProposed ApproachClassifierDatabasePerformance Measures
[33]Comparison of undersampling and oversampling to solve class imbalanceC4.54 UCI datasets* (A)expected cost
[30]3 MCDM methods to rank 9 techniques (Bayesian Network, Naive Bayes, SVM, LOGIT, k-nearest neighbour, C4.5, RIPPER, RBF network, ensemble)Top three: LOGIT, Bayesian network, ensemble2 UCI datasets* (G,A) and credit datasets representing 4 other countriesACC, AUC, SEN, SPE, precision
[31]Tree-based ensembles with synthetic features for features ranking and performance improvementExtreme Gradient Boosting to learn ensemble of decision treesEMIS database (Polish company)AUC
[32]Performance assessment of 5 tree-based ensemble modelsAdaBoost, LogitBoost, RUSBoost, Subspace, Bagging3 UCI datasets* (A)error rate
[29]Performance assessment of 4 neural network modelsBPNN, PNN, RBFNN, GRNN with RT as benchmark1 UCI datasetACC, SEN, SPE
* UCI dataset employed is the same as this study (G: German, A: Australian).
Table 12. Results comparison with external studies.
Table 12. Results comparison with external studies.
DataStudyModelACCAUCSENSPE
Australian[30]LOGIT0.86230.93130.85900.8664
Bayesian Network0.85220.91430.86560.7980
Ensemble0.85510.99000.87730.8274
[32]AdaBoost0.8725
LogitBoost0.8696
RUSBoost0.8551
Subspace0.7667
Bagging0.8939
HS-SVM0.86390.93130.91850.8200
MHS-SVM0.85650.93100.92520.8016
HS-RF0.87380.93660.86340.8822
MHS-RF0.86950.93560.86350.8746
German[30]LOGIT0.77100.79190.89000.4933
Bayesian Network0.72500.74100.88140.3600
Ensemble0.76200.79800.89430.4533
HS-SVM0.76200.80150.90860.4200
MHS-SVM0.76200.80510.89570.4500
HS-RF0.76400.80440.92290.3933
MHS-RF0.75600.80530.92290.3667
Table 13. Average number of reduced features.
Table 13. Average number of reduced features.
GermanAustralianLending Club
STEP14.6716.2
HS-SVM14.98.45.8
MHS-SVM13.996.1
Table 14. Computational time.
Table 14. Computational time.
GermanAustralianLending Club
LOGIT0.3698 s0.3624 s5.8037 s
STEP15.8467 s0.3822 s6.2699 min
LDA0.5280 s0.5339 s3.0937 s
GS-SVM49.3829 min20.1442 min3.6870 days
HS-SVM101.4499 min42.3236 min4.652 days
MHS-SVM36.8149 min17.7540 min1.4710 days
MHS-SVM (P)5.764 min3.405 min9.854 h
GS-RF49.4010 min15.8474 min1.9173 h
HS-RF57.0272 min30.0728 min2.3945 h
MHS-RF32.4922 min12.2331 min1.2369 h
MHS-RF (P)5.5027 min3.5496 min12.5525 min

Share and Cite

MDPI and ACS Style

Goh, R.Y.; Lee, L.S.; Seow, H.-V.; Gopal, K. Hybrid Harmony Search–Artificial Intelligence Models in Credit Scoring. Entropy 2020, 22, 989. https://doi.org/10.3390/e22090989

AMA Style

Goh RY, Lee LS, Seow H-V, Gopal K. Hybrid Harmony Search–Artificial Intelligence Models in Credit Scoring. Entropy. 2020; 22(9):989. https://doi.org/10.3390/e22090989

Chicago/Turabian Style

Goh, Rui Ying, Lai Soon Lee, Hsin-Vonn Seow, and Kathiresan Gopal. 2020. "Hybrid Harmony Search–Artificial Intelligence Models in Credit Scoring" Entropy 22, no. 9: 989. https://doi.org/10.3390/e22090989

APA Style

Goh, R. Y., Lee, L. S., Seow, H. -V., & Gopal, K. (2020). Hybrid Harmony Search–Artificial Intelligence Models in Credit Scoring. Entropy, 22(9), 989. https://doi.org/10.3390/e22090989

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop