Next Article in Journal
Identification of Auditory Object-Specific Attention from Single-Trial Electroencephalogram Signals via Entropy Measures and Machine Learning
Previous Article in Journal
Chaotic Attractors with Fractional Conformable Derivatives in the Liouville–Caputo Sense and Its Dynamical Behaviors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MOTiFS: Monte Carlo Tree Search Based Feature Selection

by
Muhammad Umar Chaudhry
* and
Jee-Hyong Lee
*
Department of Electrical and Computer Engineering, Sungkyunkwan University, Suwon 16419, Korea
*
Authors to whom correspondence should be addressed.
Entropy 2018, 20(5), 385; https://doi.org/10.3390/e20050385
Submission received: 13 April 2018 / Revised: 14 May 2018 / Accepted: 18 May 2018 / Published: 20 May 2018

Abstract

:
Given the increasing size and complexity of datasets needed to train machine learning algorithms, it is necessary to reduce the number of features required to achieve high classification accuracy. This paper presents a novel and efficient approach based on the Monte Carlo Tree Search (MCTS) to find the optimal feature subset through the feature space. The algorithm searches for the best feature subset by combining the benefits of tree search with random sampling. Starting from an empty node, the tree is incrementally built by adding nodes representing the inclusion or exclusion of the features in the feature space. Every iteration leads to a feature subset following the tree and default policies. The accuracy of the classifier on the feature subset is used as the reward and propagated backwards to update the tree. Finally, the subset with the highest reward is chosen as the best feature subset. The efficiency and effectiveness of the proposed method is validated by experimenting on many benchmark datasets. The results are also compared with significant methods in the literature, which demonstrates the superiority of the proposed method.

1. Introduction

In the current era of information overload, the size of datasets is growing extensively. This leads to the high dimensional datasets containing many redundant and irrelevant features, resulting in computationally expensive analysis and less accurate predictive modeling [1,2,3]. Feature selection comes to the rescue and aids in reducing dimensions. The feature selection algorithm looks for the optimal or most informative features by putting aside the redundant and irrelevant features, retaining accurate information and data structures where possible, resulting in efficient and more accurate predictive models. Feature selection has been studied for decades in various fields including machine learning [4,5,6], statistics [7,8], pattern recognition [9,10,11], and data mining [12,13].
When addressing the feature selection problem, there are two key aspects: search strategy and evaluation criterion. An efficient search strategy finds the best candidate subsets rather trying each and every possible subset, thus reducing the time complexity. A good evaluation criterion judges the goodness of candidate subsets and identifies the best one among them, thus improving performance in terms of accuracy. Based on the evaluation criterion, feature selection approaches are mainly classified as filter, wrapper, or hybrid approaches. In terms of the search strategy (ignoring evaluation criteria), feature selection algorithms can be classified into exhaustive search, heuristic search, or meta-heuristic search-based methods. Figure 1 shows the key aspects and classifications of feature selection methods.
Meta-heuristic approaches, often referred as Evolutionary Algorithms (EA), have recently gained much attention in feature selection [14]. Meta-heuristic algorithms dig the search space by keeping the good solutions and improving them (exploitation), as well as looking for the new ones in other areas through the search space (exploration). Examples of Evolutionary Algorithms are the Genetic Algorithm (GA) [15,16], Ant Colony Optimization (ACO) [17,18], Particle Swarm Optimization (PSO) [19,20,21], Multi-Objective Evolutionary Algorithms [22,23,24] and Bat Algorithms [25,26]. The use of these approaches is still in infancy stages with a debate on which approach is better than the others. Although the performance of meta-heuristic algorithms is pretty useful compared to traditional heuristic approaches, they are complex in that they need to be fine-tuned on many hyper parameters and need enough time to achieve convergence [27]. The tradeoff between the computational feasibility, model complexity and optimal features selection is still an unsolved puzzle among all these methods [14]. Therefore, vast room for improvement is available and new algorithms are immensely needed to overcome such issues, which can efficiently achieve high accuracy with less model complexity.
In this paper, we present a novel approach for feature selection which combines the robustness and dynamicity of Monte Carlo Tree Search (MCTS) with the accuracy of wrapper methods. We employ MCTS as an efficient search strategy within wrapper framework developing the efficient and effective algorithm, named as MOTiFS (Monte carlO Tree Search Based Feature Selection). MCTS is a search strategy which finds the optimal solutions probabilistically by using lightweight random simulations [28]. It takes random samples in the search space and builds the search tree accordingly. Currently, MCTS is successfully being deployed in games with huge search space [29]. However, its effectiveness is not well explored for feature selection problems, which is the major motivation of this study.
The proposed algorithm, MOTiFS, starts with an empty tree node, meaning no feature has been selected. The tree is then incrementally built by adding nodes one by one representing either of the two corresponding feature states: a feature is selected or not selected. Every iteration leads to the generation of a feature subset following the tree and default policies. The tree policy not only exploits the expanded feature space by searching for the features which have already shown good performance in the previous iterations, but also explores the new features by expanding the tree incrementally. The default policy, then, induces randomness by choosing the features randomly from the remaining set of yet unexpanded features. This perfect blend of tree search with random sampling accelerates the process and provides the opportunity to generate the best feature subset in a few iterations, even if the search tree is not fully expanded. MOTiFS uses the classification accuracy as a goodness of the current feature subset as well as the reward for the current iteration. The search tree is then updated by propagating the reward backwards through the selected nodes. Finally, the feature subset with highest accuracy is chosen as a best feature subset. For experimental purposes, the K-Nearest Neighbor classifier is employed as a reward function. MOTiFS is tested on 25 real-world datasets and the promising results prove its validity. The comparison with latest and state-of-the art methods shows the superiority of MOTiFS and serves as a proof of concept. The main contributions of this study are listed below:
  • The novel feature selection algorithm, MOTiFS, is proposed which combines the robustness of MCTS with the accuracy of wrapper methods.
  • MOTiFS searches through the feature space efficiently and find the best feature subset within a few iterations, relatively.
  • Only two hyper-parameters, scaling factor and termination criteria, are required to be tuned, making MOTiFS simple and flexible to handle.
  • MOTiFS is tested on 25 benchmark datasets and results are also compared with other established methods. The promising results demonstrate the superiority of MOTiFS.
The rest of the paper is organized as follows. The review of the literature is provided in Section 2. Section 3 provides the necessary background for the proposed method. Section 4 presents the demonstration of the proposed method (MOTiFS). The results and experimental details are presented in Section 5. Finally, the conclusions and future research directions are discussed in Section 6.

2. Literature Review

The key aspects of feature selection are illustrated above in Figure 1. This section presents a brief overview of various feature selection methods.
Filter methods are independent of the specific classification algorithm. They use the inherent properties, like distance and information gain, of the dataset and measure the importance of each feature with respect to the class label and rank them [30,31,32]. Filter-based methods are fast enough and can be used with any classification algorithm, but there is a major drawback in that they show a lower performance in terms of classification accuracy. In wrapper methods, the classification algorithm is directly related in a way that the accuracy of the classifier serves as a measure of goodness of the candidate feature subsets [33,34,35]. They are computationally expensive because they run the classifier repeatedly but deliver high accuracy compared to filter methods. Hybrid methods integrate the filter and wrapper methods in order to take advantage of both types [36,37]. Such methods use the independent metric and a learning algorithm in order to measure the goodness of each candidate feature subset in the search space.
Irrespective of the evaluation criterion, feature selection methods fall into one of the following search strategies: exhaustive, heuristic or meta-heuristic. In earlier literature, a few attempts at feature selection have been made, involving exhaustive searches [38]. However, applying an exhaustive search on datasets with many features is practically impossible due to the complexity involved, so they are seldom used. Hence, researchers have adopted heuristic search strategies, like greedy hill climbing and best first search, which use some heuristics to reach the goal rather traversing the whole search space [30,39]. Greedy hill climbing approaches include SFS (Sequential Forward Selection), SBS (Sequential Backward Selection) and bidirectional search algorithms. They look for the relevant features by evaluating all local changes in a search space. However, the major drawback associated with such algorithms is that whenever a positive change occurs—either a feature is added to the selected set in SFS or deleted from the selected set in SBF—this feature does not get a chance to be re-evaluated, and it becomes highly probable that the algorithm will deviate from optimality. Such a problem is referred as the nesting effect [14]. In efforts to tackle this major issue, researchers came up with some useful algorithms like SFFS (Sequential Forward Floating Selection) and SBFS (Sequential Backward Floating Selection).
Recently, meta-heuristic approaches like the Genetic Algorithm (GA) [15,16], Ant Colony Optimization (ACO) [17,18], Particle Swarm Optimization (PSO) [19,20,21], Multi-Objective Evolutionary Algorithms [22,23,24] and Bat Algorithms [25,26] have gained much attention [14]. However, the involvement of too many hyper parameters makes tuning the models for optimized performance too complex [27]. For example, in GAs, a sufficient population size is required with high enough generations to obtain the desired results. Obviously, this leads GAs to be computationally expensive. Also, many parameters are involved in GAs, like population size, number of generations, crossover probability, permutation probability, etc., which makes it more challenging to find the suitable model for effective feature selection. One earlier attempt using MCTS in feature selection is found in [40]. The method proposed in [40] maps the feature selection as an exhaustive search tree and, therefore, has a huge branching factor and is computationally very expansive with unacceptable bounds.
In this study, we proposed a novel feature selection algorithm based on MCTS and wrapper methods. We define the feature selection tree in a novel and incremental fashion, where exploration and exploitation are well balanced within limited computational bounds. The extensive experimentation on many benchmark datasets and comparison with state-of-the-art methods demonstrates the validity of the proposed method.

3. Background

This section presents the background concepts used in the proposed method.

3.1. Working Procedure of Monte Carlo Tree Search (MCTS)

MCTS is a heuristic search method which uses lightweight random simulations to reach a goal state [28]. Each MCTS iteration consists of four sequential steps: selection, expansion, simulation and backpropagation.
  • Selection: Starting from the root node, the algorithm traverses the tree by selecting nodes with the highest approximated values, until a non-terminal node with unexpanded children is reached.
  • Expansion: A new child node is added to expand the tree, according to the available set of actions.
  • Simulation: From the new child node, a random simulation is performed until the terminal node is reached, to approximate the reward.
  • Backpropagation: The simulation result (reward) is backpropagated through the selected nodes to update the tree.
The selection and expansion steps are performed using the tree policy, whereas the simulation step is performed with the default policy.

3.2. Upper Confidence Bounds for Trees (UCT) Algorithm

The tree policy uses the Upper Confidence Bounds for Trees (UCT) algorithm for node selection. The value of each node is approximated using the UCT algorithm, as shown in Equation (1). The tree policy then selects the nodes at each level which have the highest approximated values. This maintains a balance between exploiting the good solutions and exploring the new ones.
U C T v =   W v N v +   C   × 2   × l n ( N p ) N v ,
where, N v and N p represents number of times nodes v and its parent p are visited, respectively. W v represents the number of wining simulations (considering a games perspective) at node v . C is the exploration constant.

4. MOTiFS (Monte Carlo Tree Search Based Feature Selection)

The Monte Carlo Tree Search (MCTS) is used as a search strategy within a wrapper framework to develop a novel approach for feature selection. Using the efficient and meta-heuristic approach of MCTS and the predictive accuracy of the wrapper method, the goal is to find the best feature subset to give maximum classification accuracy. Figure 2 shows the depiction of the proposed method.
The preliminary step is to map the feature selection problem into some sort of game tree. In feature selection, either a feature is selected or not selected in a feature subset, and represented by 1 or 0 at the corresponding feature position in a total set of feature space. Using this intuition, we map the problem as a single player game where the goal is to select best possible features with the maximum accumulative reward. MOTiFS constructs a special tree where each node represents either of the two corresponding feature states: a feature is selected or not selected. The definition for the feature selection tree is provided below:
Definition 1.
For a feature set, F = { f 1 ,   f 2 ,   ,   f i ,   ,   f n } , the feature selection tree is a tree satisfying the following conditions:
  • The root is 0 , which means no features are selected.
  • Any node at level i 1 has two children, f i and i , where 0 < i < n .
Where, nodes f i and i represent the inclusion or exclusion of the corresponding feature, f i , in the feature subset, respectively. Any path from the root node to one of the leaves represents a feature subset. So, the goal is to find a path which gives the best reward (accuracy). We use MCTS for tree construction and traversal, and finally choose the path (feature subset) with best accuracy.
The search starts with an empty root node and incrementally builds the tree by adding nodes representing features states, one by one, with a random probability of being selected or not. At each turn, a subset of features is selected following the tree and default policies. The classification accuracy of the current feature subset is used as a reward, and the search tree is updated through backpropagation. The feature selection tree and four steps search procedure are graphically represented in Figure 3. Table 1 summarizes the notations used throughout the text.

4.1. Feature Subset Generation

A feature subset is generated during the selection, expansion and simulation steps, in each MCTS iteration.

4.1.1. Selection

In the selection step, one path is selected from the already expanded tree. The path selected is the one whose inclusion gave a high reward in the previous iterations. The features in the selected path are included in the feature subset of the current iteration. The algorithm traverses the already expanded tree following the tree policy until a node is reached which is non-terminal and has an unexpanded child. The UCT algorithm is used to decide which node to be chosen at each level. If the UCT algorithm selects node f i at level i , feature f i is included in the current feature subset. If it selects i , feature f i is not included. If f i is selected, this is based on an intuition that the inclusion of feature f i gave a high reward in previous iterations, so needs to be included in the current feature subset. On the other hand, if it is not selected, it is better not to choose feature f i   as it did not contribute much towards a better reward, previously.
The vanilla UCT algorithm approximates the reward at each node by dividing by the number of times the node is visited, as shown in Equation (1). This kind of approximation is most suitable in the game theoretic perspective where the reward is either a 1 (win) or 0 (loss), and the goal is to select the nodes (moves) which give the maximum number of wins in the minimum number of visits. However, feature selection is a different sort of a problem where the goal is to select the path which gives the maximum reward (accuracy). Using this intuition, instead of approximating the reward and penalizing by the number of visits, we used the maximum reward obtained at each node. The modified form of the UCT algorithm used in MOTiFS is shown in Equation (2). During tree traversal, the nodes which receive the highest scores from Equation (2) are selected until a non-terminal node with an unexpanded child is reached.
U C T v j =   max ( Q v j ) +   C   × 2   × l n ( N v i ) N v j ,
where, max ( Q v j ) is the maximum reward at node v j and C > 0 is a constant. N v j and N v i represent the number of times nodes v j and its parent v i are visited, respectively.

4.1.2. Expansion

During expansion, a new child node is added to the urgent node (the last selected node in the selection step). The addition of a new child node at node v i is also based on the UCT function. If U C T f i + 1 is larger than U C T i + 1 then child node f i + 1 is added, and thus, feature f i + 1 is included in the current feature subset. Conversely, child node i + 1 is added and feature f i + 1 is not included in the current feature subset.

4.1.3. Simulation

The simulation step induces randomness in feature subset generation following the default policy. It choses features from the remaining unexpanded features, with a uniform probability of being selected or not. If the recently expanded node is v i , a path from v i to a leaf node is randomly selected.
Assuming the current expanded node is v i , the inclusion of features from f 1 to f i into the current feature subset is determined in the selection and expansion steps, whereas the inclusion of the remaining features from f i + 1 to f n in the current feature subset is randomly determined in the simulation step. A tree search and a random search participate together in feature subset generation, thus giving the opportunity to obtain the best feature subset in fewer runs even if the search tree is not fully expanded.

4.2. Reward Calculation and Backpropagation

The classifier is then applied to evaluate the goodness of the feature subset. The classification accuracy of the current feature subset is also used as a simulation reward, Q s i m u l a t i o n , for the current selected nodes and propagated backwards to update the search tree.
Q s i m u l a t i o n =   A C C c l a s s i f i e r ( F s u b s e t )
where, A C C c l a s s i f i e r ( . ) represents the accuracy of the classifier on the current feature subset, F s u b s e t . If the accuracy of the current feature subset is better than the previous best, then the current feature subset becomes the best feature subset. This process goes on until stopping criteria is met.
For the purpose of this study, we employed the nearest neighbors (K-NN) classifier to evaluate the candidate feature subset and as a reward function. We used the simple and efficient nearest neighbors classifier as it is well-understood in the literature and works surprisingly well in many situations [41,42,43,44]. Moreover, many other similar studies and comparison methods in literature, mentioned in Section 5.3, have applied the nearest neighbors classifier and therefore, we considered it to be a better choice for the comparative analysis. However, any other classifier can be used within the proposed framework. The algorithm for MOTiFS is provided below as Algorithm 1.
Algorithm 1 The MOTiFS Algorithm
Load dataset and preprocess
Initialize SCALAR, BUDGET //Scaling factor & Number of MCTS simulations (hyper parameters)
function MOTiFS (featuresList)
  create rootNode
  maxReward, bestFeatureSubset ← UCTSEARCH (rootNode)
  return (maxReward, bestFeatureSubset)
function UCTSEARCH (rootNode)
  Initialize maxReward, bestFeatureSubset
  while within computational budget do
   frontNode ← TREEPOLICY (rootNode)
   reward, featureSubset ← DEFAULTPOLICY (frontNode.state)
   BACKUP (frontNode, reward)
   if reward is greater than maxReward then
    maxRewardreward
    bestFeatureSubsetfeatureSubset
  return (maxReward, bestFeatureSubset)
function TREEPOLICY (node)
  while node is non-terminal do
   if node not fully expanded then
    return EXPAND (node)
   else
    node ← BESTCHILD (node, SCALAR)
  return node
function EXPAND (node)
  choose a untried actions from A(node.state)
  add a newChild with f(node.state, a)
  return newChild
function BESTCHILD ( v , C )
  return max v   children   of   v max ( Q v ) + C 2   ×   l n ( v . v i s i t s ) v . v i s i t s
function DEFAULTPOLICY (state)
  while state is non-terminal do
   choose a A(state) uniformly at random
  state ← f(state, a)
traverse state.path
   if ai is equal to fi+1 then
    featureSubset ← INCLUDE (fi+1)
   reward ← REWARD (featureSubset)
   return (reward, featureSubset)
function BACKUP (node, reward)
  while node is not null do
   node.visitsnode.visits + 1
   if reward > node.reward then
    node.rewardreward
   nodenode.parent
  return

5. Experiment and Results

The efficacy of the MOTiFS was demonstrated by experimenting on many publicly-available benchmark datasets. Twenty-five benchmark datasets of varying dimensions were used and results were compared with other significant methods in the literature.

5.1. Datasets

Twenty-five benchmark datasets were used for validation and comparison purposes. Twenty-four datasets were taken from two publicly available repositories, LIBSVM [45] and UCI [46]. However, one dataset “Klekota Roth fingerprint (KRFP)” representing the fingerprints of the chemical compounds was taken from the 5-HT5A dataset to classify between active or inactive compounds [11,47]. The details of the datasets are summarized in Table 2. The datasets taken were of varying dimensions and sizes. In feature selection, literature categorizes datasets into three dimensional ranges, based on the total number of features (F): low dimension (0–19), medium dimension (20–49), and high dimension (50–∞) [48]. In the current study, 10 datasets were low dimensional, 5 datasets were medium dimensional, and 10 datasets were high dimensional.

5.2. Experimental Procedure and Parameter Setup for MOTiFS

We conducted 10-fold cross validation for the whole feature selection procedure. A dataset was equally divided into 10 random partitions. Then, a single partition was retained as a test set, while the remaining 9 partitions were used as a training set. This procedure was repeated 10 times (each partition behaved as a test set exactly once).
The significant advantage of MOTiFS is that only two hyper-parameters are required to be tuned. The parameter values used in the experiments are presented in Table 3. The “scaling factor”, C, maintains the balance between exploiting good solutions and exploring new ones in the search space. An excessively large value of C benefits the exploration part and slows down convergence, whereas, an insufficient C may cause the search to be stuck in a local optimum by penalizing the exploration and only sticking to locally good solutions. After careful examination and series of experimentation, we limited our choice of C to 0.1, 0.05, and 0.02. During training, we constructed three feature selection trees with different scaling factors (0.1, 0.05, 0.02) and selected one of them based on 5-fold cross validation accuracy. That is, the scaling factor was auto-tuned during the training process. For the “termination criteria”, we used the fixed number of iterations (MCTS simulations). This depends on the dimensional size of the dataset. We set the number of simulations to 500 if the total number of features dimensions was less than 20, otherwise it was set to 1000, excluding the “KRFP” dataset. For the very high dimensional dataset, “KRFP”, we used 10,000 simulations.

5.3. Comparison Methods

We compared MOTiFS with state-of-the-art methods in the literature. The comparison methods were diverse and varied between established sequential approaches, fuzzy rule-based, evolutionary and entropy reduction-based methods, as summarized in Table 4. Most of the comparison methods were wrapper based, except FS-FS and FR-FS (well established filter-based methods) which we included for generalized comparison of accuracies achieved with different datasets earlier. For fair comparison against each dataset we deployed the classifier reported in available comparison methods. For datasets “Spambase”, “WBC”, “Ionosphere”, “Arrhythmia”, “Multiple features”, “Waveform”, “WBDC”, “Glass”, “Wine”, “Australian”, “German number”, “Zoo”, “Breast cancer”, “DNA”, “Vehicle”, “Sonar”, “Hillvalley”, “Musk 1”, “Splice” and “KRFP”, the used classifier was 5-NN. For datasets “Soybean-small”, “Liver disorders”, “Credit”, “Tic-tac-toe” and “Libras movement”, the 3-NN classifier is deployed.

5.4. Results and Comparisons

We conducted 10-fold cross-validation for the whole feature selection procedure, as detailed in Section 5.2. As our method is heuristic, we ran our algorithm five times on every datasets and reported the average of five runs. We reported the average accuracy and the number of features selected. Table 5 and Table 6 present a detailed summary of the results and comparisons with other methods. The bold values in each row indicate the best among all the methods.
In Table 5, results are reported for 20 datasets using 5-NN as a classifier. Comparing classification accuracies, it is obvious that MOTiFS overall outperformed on 15 datasets, namely “Spambase”, “WBC”, “Ionosphere”, “Multiple features”, “WBDC”, “Glass”, “Wine”, “Australian”, “German number”, “Breast cancer”, “Vehicle”, “Sonar”, “Musk 1”, “Splice” and “KRFP”, compared to all other methods. Moreover, for two datasets, “Arrhythmia” and “Waveform”, MOTiFS ranked second among all the competitors. However, for three datasets, namely “Zoo”, “DNA” and “Hillvalley”, MOTiFS did not perform well.
Table 6 presents the result according to 3-NN on five datasets. Clearly, MOTiFS outperformed on four datasets, namely “Liver-disorders”, “Credit”, “Tic-tac-toe” and “Libras movement” in terms of classification accuracy. However, on the “Soybean-small” dataset, the average MOTiFS score was not 1.0, as reported by all other methods, although MOTiFS achieved an accuracy of 1.0 on three out of five independent runs.
We also reported the standard deviation of five independent runs of MOTiFS for each dataset in Table 5 and Table 6, according to the average accuracy. For almost all of the datasets, the standard deviation was too small to be negligible. Thus, the stability and reliability of MOTiFS was evident.
While comparing the number of selected features in Table 5, MOTiFS could not outperform because of marginal differences among all the methods. The reason is quite obvious; MOTiFS does not account for the selected features in reward evaluation and employs the classification accuracy only as the reward function. However, the DR (dimensional reduction) achieved by MOTiFS is presented in Figure 4. We calculate the DR on each dataset using Equation (4).
DR = 1   #   s e l e c t e d   f e a t u r e s #   t o t a l   f e a t u r e s
MOTiFS performed remarkably well in terms of accuracy and DR on all datasets compared to all other methods. On high dimensional datasets, “Ionosphere”, “Arrhythmia”, “Multiple features”, “Sonar”, “Musk 1”, “Splice”, “KRFP” and “Libras movement”, MOTiFS obtained DR values above 50%, achieving a high accuracy compared to other methods.
Summarizing the overall performance, MOTiFS outperformed the other methods on 19 out of 25 datasets, namely “Spambase”, “WBC”, “Ionosphere”, “Multiple features”, “WBDC”, “Glass”, “Wine”, “Australian”, “German number”, “Breast cancer”, “Vehicle”, “Sonar”, “Musk 1”, “Splice”, “KRFP”, “Liver-disorders”, “Credit”, “Tic-tac-toe” and “Libras movement”, compared to all other methods, in term of classification accuracy. For three datasets, “Arrhythmia”, “Waveform” and “Soybean-small”, MOTiFS ranked second among all methods. However, for three datasets, “Zoo”, “DNA” and “Hillvalley”, MOTiFS could not stand as high as the other methods. Overall, considering the accuracy and dimensional reduction together, MOTiFS performed remarkably well, as discussed above.

5.5. Discussion

We studied the effectiveness of MCTS in feature selection, which had previously been barely researched. Defining a feature selection problem as the proposed feature selection tree and applying MCTS to find best feature subset is a new concept, to the best of our knowledge. The proposed feature selection tree has the potential benefit of having less branching factors. A tree can grow to sufficient depth in a limited number of simulations, thus, taking adequate benefits from both the tree search and random sampling.
For the total number of features, n , if complexities in the node selection operation (UCT algorithm + random selection) and classifier are b and c , respectively, then the complexity of one MCTS simulation is O ( n b + c ) . However, the complexity of the node selection operation is a constant, so the complexity for s number of simulations is O ( s n + s c ) . If the number of simulations is fixed, the complexity of our proposed method is linear to the number of features, excluding the complexity of the classifier. Our proposed method finds the best feature subset within a limited number of simulations, as shown in the reported results and comparisons above.
We performed extensive experiments on 25 datasets of varying dimensions and sizes, with an aim of properly testing the performance of the proposed method. The results and comparisons with other state-of-the-art and evolutionary approaches showed the efficacy and usefulness of the proposed method. However, the performance could further be improved by careful examination of the datasets’ characteristics, modifying the reward functions and optimization of model parameters accordingly.
Future research directions may include experimentations on very large dimensional datasets and/or playing with different reward functions with an intention to further improve the performance in terms of both increasing accuracy and reducing dimensions.

6. Conclusions

In this paper, we proposed a novel feature selection algorithm, MOTiFS, which combines the robustness and dynamicity of MCTS with the accuracy of wrapper methods. MOTiFS searched the feature space efficiently by balancing between exploitation and exploration and found the best feature subset within a few iterations. Another significant feature of MOTiFS was the involvement of only two hyper-parameters: scaling factor and termination criteria, thus, making MOTiFS simple and flexible to handle. The K-NN classifier was used for experiments, and results were compared with the significant and state-of-the-art methods in the literature. Besides offering an improved classification accuracy on 25 real-world datasets, MOTiFS significantly reduced the dimensions of high dimensional datasets.

Author Contributions

M.U.C. and J.-H.L. conceived the idea and designed the experiments; M.U.C. performed the experiments; M.U.C. and J.-H.L. analyzed the data; M.U.C. wrote the paper.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIP) (No. NRF-2016R1A2B4015820). This research was also supported by Next-Generation Information Computing Development Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT (NRF-2017M3C4A7069440).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yu, L.; Liu, H. Efficient Feature Selection via Analysis of Relevance and Redundancy. J. Mach. Learn. Res. 2004, 5, 1205–1224. [Google Scholar]
  2. Gasca, E.; Sánchez, J.S.; Alonso, R. Eliminating redundancy and irrelevance using a new MLP-based feature selection method. Pattern Recognit. 2006, 39, 313–315. [Google Scholar] [CrossRef]
  3. Gheyas, I.A.; Smith, L.S. Feature subset selection in large dimensionality domains. Pattern Recognit. 2010, 43, 5–13. [Google Scholar] [CrossRef]
  4. Zheng, Y.; Kwoh, C.K. A Feature Subset Selection Method Based On High-Dimensional Mutual Information. Entropy 2011, 13, 860–901. [Google Scholar] [CrossRef]
  5. Robnik-Šikonja, M.; Kononenko, I. Theoretical and Empirical Analysis of ReliefF and RReliefF. Mach. Learn. 2003, 53, 23–69. [Google Scholar] [CrossRef]
  6. Sluga, D.; Lotrič, U. Quadratic mutual information feature selection. Entropy 2017, 19, 157. [Google Scholar] [CrossRef]
  7. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning. Elements 2009, 1, 337–387. [Google Scholar]
  8. Guo, Y.; Berman, M.; Gao, J. Group subset selection for linear regression. Comput. Stat. Data Anal. 2014, 75, 39–52. [Google Scholar] [CrossRef]
  9. Saganowski, S.; Gliwa, B.; Bródka, P.; Zygmunt, A.; Kazienko, P.; Kozlak, J. Predicting community evolution in social networks. Entropy 2015, 17, 3053–3096. [Google Scholar] [CrossRef]
  10. Reif, M.; Shafait, F. Efficient feature size reduction via predictive forward selection. Pattern Recognit. 2014, 47, 1664–1673. [Google Scholar] [CrossRef]
  11. Śmieja, M.; Warszycki, D. Average information content maximization-a new approach for fingerprint hybridization and reduction. PLoS ONE 2016, 11, e0146666. Available online: http://ww2.ii.uj.edu.pl/~smieja/aic/ (accessed on 5 May 2018). [CrossRef] [PubMed]
  12. Dash, M.; Choi, K.; Scheuermann, P.; Liu, H. Feature selection for clustering-a filter solution. In Proceedings of the 2002 IEEE International Conference on Data Mining, ICDM 2003, Maebashi City, Japan, 9–12 December 2002; pp. 115–122. [Google Scholar]
  13. Kim, Y.; Street, W.N.; Menczer, F. Feature selection in unsupervised learning via evolutionary search. In Proceedings of the 6th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Boston, MA, USA, 20–23 August 2000; pp. 365–369. [Google Scholar]
  14. Xue, B.; Zhang, M.; Browne, W.N.; Yao, X. A Survey on Evolutionary Computation Approaches to Feature Selection. IEEE Trans. Evol. Comput. 2016, 20, 606–626. [Google Scholar] [CrossRef]
  15. Hamdani, T.M.; Won, J.M.; Alimi, A.M.; Karray, F. Hierarchical genetic algorithm with new evaluation function and bi-coded representation for the selection of features considering their confidence rate. Appl. Soft Comput. J. 2011, 11, 2501–2509. [Google Scholar] [CrossRef]
  16. Hong, J.H.; Cho, S.B. Efficient huge-scale feature selection with speciated genetic algorithm. Pattern Recognit. Lett. 2006, 27, 143–150. [Google Scholar] [CrossRef]
  17. Kabir, M.M.; Shahjahan, M.; Murase, K. A new hybrid ant colony optimization algorithm for feature selection. Expert Syst. Appl. 2012, 39, 3747–3763. [Google Scholar] [CrossRef]
  18. Tabakhi, S.; Moradi, P. Relevance-redundancy feature selection based on ant colony optimization. Pattern Recognit. 2015, 48, 2798–2811. [Google Scholar] [CrossRef]
  19. Unler, A.; Murat, A.; Chinnam, R.B. Mr2PSO: A maximum relevance minimum redundancy feature selection method based on swarm intelligence for support vector machine classification. Inf. Sci. 2011, 181, 4625–4641. [Google Scholar] [CrossRef]
  20. Zhang, Y.; Gong, D.; Hu, Y.; Zhang, W. Feature selection algorithm based on bare bones particle swarm optimization. Neurocomputing 2015, 148, 150–157. [Google Scholar] [CrossRef]
  21. Xue, B.; Zhang, M.; Browne, W.N. Single feature ranking and binary particle swarm optimisation based feature subset ranking for feature selection. Conf. Res. Pract. Inf. Technol. Ser. 2012, 122, 27–36. [Google Scholar]
  22. Paul, S.; Das, S. Simultaneous feature selection and weighting—An evolutionary multi-objective optimization approach. Pattern Recognit. Lett. 2015, 65, 51–59. [Google Scholar] [CrossRef]
  23. Cordon, O.; Herrera, F.; del Jesus, M.J.; Villar, P. A multiobjective genetic algorithm for feature selection and granularity learning in fuzzy-rule based classification systems. In Proceedings of the IFSA World Congress and 20th NAFIPS International Conference, Vancouver, BC, Canada, 25–28 July 2001; Volume 3, pp. 1253–1258. [Google Scholar]
  24. Xue, B.; Fu, W.; Zhang, M. Multi-objective Feature Selection in Classification: A Differential Evolution Approach. In Proceedings of the 10th International Conference on Simulated Evolution and Learning, Dunedin, New Zealand, 15–18 December 2014; pp. 516–528. [Google Scholar]
  25. Nakamura, R.Y.M.; Pereira, L.A.M.; Costa, K.A.; Rodrigues, D.; Papa, J.P.; Yang, X.S. BBA: A binary bat algorithm for feature selection. In Proceedings of the Brazilian Symposium of Computer Graphic and Image Processing, Campinas, Brazil, 17–20 October 2012; pp. 291–297. [Google Scholar]
  26. Rodrigues, D.; Pereira, L.A.M.; Nakamura, R.Y.M.; Costa, K.A.P.; Yang, X.; Souza, A.N.; Papa, J.P. A wrapper approach for feature selection based on Bat Algorithm and Optimum-Path Forest. Expert Syst. Appl. 2014, 41, 2250–2258. [Google Scholar] [CrossRef]
  27. Montazeri, M. HHFS: Hyper-heuristic feature selection. Intell. Data Anal. 2016, 20, 953–974. [Google Scholar] [CrossRef]
  28. Browne, C.; Powley, E. A survey of monte carlo tree search methods. IEEE Trans. Intell. AI Games 2012, 4, 1–49. [Google Scholar] [CrossRef] [Green Version]
  29. Silver, D.; Huang, A.; Maddison, C.J.; Guez, A.; Sifre, L.; van den Driessche, G.; Schrittwieser, J.; Antonoglou, I.; Panneershelvam, V.; Lanctot, M.; et al. Mastering the game of Go with deep neural networks and tree search. Nature 2016, 529, 484–489. [Google Scholar] [CrossRef] [PubMed]
  30. Hall, M. Correlation-based Feature Selection for Machine Learning. Methodology 1999, 21i195-i20, 1–5. [Google Scholar]
  31. Senawi, A.; Wei, H.L.; Billings, S.A. A new maximum relevance-minimum multicollinearity (MRmMC) method for feature selection and ranking. Pattern Recognit. 2017, 67, 47–61. [Google Scholar] [CrossRef]
  32. Zhao, G.D.; Wu, Y.; Chen, F.Q.; Zhang, J.M.; Bai, J. Effective feature selection using feature vector graph for classification. Neurocomputing 2015, 151, 376–389. [Google Scholar] [CrossRef]
  33. Huang, C.L.; Wang, C.J. A GA-based feature selection and parameters optimizationfor support vector machines. Expert Syst. Appl. 2006, 31, 231–240. [Google Scholar] [CrossRef]
  34. Durbha, S.S.; King, R.L.; Younan, N.H. Wrapper-based feature subset selection for rapid image information mining. IEEE Geosci. Remote Sens. Lett. 2010, 7, 43–47. [Google Scholar] [CrossRef]
  35. Kohavi, R.; John, G.H. Wrappers for feature subset selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef]
  36. Bermejo, P.; Gámez, J.A.; Puerta, J.M. A GRASP algorithm for fast hybrid (filter-wrapper) feature subset selection in high-dimensional datasets. Pattern Recognit. Lett. 2011, 32, 701–711. [Google Scholar] [CrossRef]
  37. Solorio-Fernández, S.; Carrasco-Ochoa, J.A.; Martínez-Trinidad, J.F. A new hybrid filter–wrapper feature selection method for clustering based on ranking. Neurocomputing 2016, 214, 866–880. [Google Scholar] [CrossRef]
  38. Almuallim, H.; Dietterich, T.G. Learning Boolean concepts in the presence of many irrelevant features. Artif. Intell. 1994, 69, 279–305. [Google Scholar] [CrossRef]
  39. Li, F.; Zhang, Z.; Jin, C. Feature selection with partition differentiation entropy for large-scale data sets. Inf. Sci. 2016, 329, 690–700. [Google Scholar] [CrossRef]
  40. Gaudel, R.; Sebag, M. Feature Selection as a One-Player Game. In Proceedings of the International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 359–366. [Google Scholar]
  41. Szenkovits, A.; Meszlenyi, R.; Buza, K.; Gasko, N.; Lung, R.I.; Suciu, M. Feature Selection with a Genetic Algorithm for Classification of Brain Imaging Data; Springer: Cham, Switzerland, 2018. [Google Scholar]
  42. Buza, K.; Alexandros, N.; Lars, S.-T. Time-series classification based on individualized error prediction. In Proceedings of the IEEE 13th International conference on Computational Science and Engineering (CSE), Hong Kong, China, 11–13 December 2010; pp. 48–54. [Google Scholar]
  43. Chen, G.H.; Stanislav, N.; Devavrat, S. A latent source model for nonparametric time series classification. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2013; pp. 1088–1098. [Google Scholar]
  44. Devroye, L.; Gyorfi, L.; Krzyzak, A.; Lugosi, G. On the Strong Universal Consistency of Nearest Neighbor Regression Function Estimates. Ann. Stat. 1994, 22, 1371–1385. [Google Scholar] [CrossRef]
  45. Chang, C.; Lin, C. Retrieved from LIBSVM—A Library for Support Vector Machines. 2001. Available online: https://www.csie.ntu.edu.tw/~cjlin/libsvm/ (accessed on 18 July 2017).
  46. Machine Learning Repository. Retrieved from University of California, Irvine. Available online: http://archive.ics.uci.edu/ml/index.php (accessed on 18 July 2017).
  47. Klekota, J.; Roth, F.P. Chemical substructures that enrich for biological activity. Bioinformatics 2008, 24, 2518–2525. [Google Scholar] [CrossRef] [PubMed]
  48. Tahir, M.A.; Bouridane, A.; Kurugollu, F. Simultaneous feature selection and feature weighting using Hybrid Tabu Search/K-nearest neighbor classifier. Pattern Recognit. Lett. 2007, 28, 438–446. [Google Scholar] [CrossRef]
  49. Mitra, P.; Murthy, C.A.; Pal, S.K. Unsupervised feature selection using feature similarity. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 301–312. [Google Scholar] [CrossRef]
  50. Chen, Y.-C.; Pal, N.R.; Chung, I.-F. An Integrated Mechanism for Feature Selection and Fuzzy Rule Extraction for Classification. IEEE Trans. Fuzzy Syst. 2012, 20, 683–698. [Google Scholar] [CrossRef]
Figure 1. Key aspects of feature selection.
Figure 1. Key aspects of feature selection.
Entropy 20 00385 g001
Figure 2. The proposed method, MOTiFS (Monte Carlo Tree Search Based Feature Selection).
Figure 2. The proposed method, MOTiFS (Monte Carlo Tree Search Based Feature Selection).
Entropy 20 00385 g002
Figure 3. Feature selection tree and search procedure of MOTiFS.
Figure 3. Feature selection tree and search procedure of MOTiFS.
Entropy 20 00385 g003
Figure 4. Graphical representation of dimensional reduction (DR) achieved by MOTiFS on all the datasets.
Figure 4. Graphical representation of dimensional reduction (DR) achieved by MOTiFS on all the datasets.
Entropy 20 00385 g004
Table 1. Notations used in the proposed method.
Table 1. Notations used in the proposed method.
NotationInterpretation
F Original feature set
n Total number of features
v i Node v at tree level i
a ( v i ) Action taken at v i
Q simulation Simulation reward
Table 2. Summary of the selected datasets.
Table 2. Summary of the selected datasets.
#DatasetNo. of FeaturesNo. of InstancesNo. of Classes
1Spambase5747012
2WBC96992
3Ionosphere343512
4Arrhythmia19545216
5Multiple features649200010
6Waveform4050003
7WBDC305692
8Glass92146
9Wine131783
10Australian146902
11German number2410002
12Zoo171017
13Breast cancer106832
14DNA18020002
15Vehicle188464
16Sonar602082
17Hillvalley1006062
18Musk 11664762
19Splice6010002
20KRFP *48602152
21Soybean-small35474
22Liver disorders63452
23Credit156902
24Tic-tac-toe99852
25Libras movement9036015
* downloaded from [11].
Table 3. Parameters setup for MOTiFS.
Table 3. Parameters setup for MOTiFS.
ParameterValues Used for Different Datasets
Scaling factor, C(0.1, 0.05, 0.02)
Termination criteria(500, 1000, 10,000) iterations
Table 4. Summary of the methods for our comparison.
Table 4. Summary of the methods for our comparison.
MethodDescription
SFS, SBSSequential Forward Selection and Sequential Backward Selection * [22] (2015)
FS-FSFeature Similarity Technique [49] (2002)
FR-FSFuzzy Rule Based Technique [50] (2012)
SFSWAn Evolutionary Multi-Objective Optimization Approach [22] (2015)
DEMOFSDifferential Evolution Based Multi-Objective Feature Selection * [22] (2014)
BABat Algorithm and Optimum-Path Forest Based Wrapper Approach [26] (2014)
PSOParticle Swarm Optimization Based Method * [26] (2014)
SCE, CCEShannon’s Entropy Reduction, Complementary Entropy Reduction * [39] (2016)
PDE-2Partition Differential Entropy Based Method [39] (2016)
* Reported from the mentioned reference.
Table 5. Comparison of MOTiFS with other methods, according to 5-NN.
Table 5. Comparison of MOTiFS with other methods, according to 5-NN.
DatasetAvg. Acc.
# Sel. Feat.
MOTiFSSFSW [22]SFS [22]SBS [22]FS-FS [49]FR-FS [50]DEMOFS [22]BA [26]PSO [26]
Spambase0.907 ± 0.0030.8850.8740.8700.900
31.526.035.737.329.0
WBC0.968 ± 0.0010.9610.9600.9510.956
5.524.26.47.34.0
Ionosphere0.889 ± 0.0070.8830.8870.8590.7880.844 0.7800.790
12.3211.51.29.116.04.3321.014.0
Arrhythmia0.650 ± 0.0030.6580.5990.5800.589
94.4100.089.449.2100.0
Multiple features0.980 ± 0.0010.9790.9030.9120.783
321.84270.0210.0305.0325.0
Waveform0.816 ±0.0020.8370.7780.7850.752
19.4216.018.418.320.0
WBDC0.967 ± 0.0040.9410.9010.898 0.936
15.4213.513.917.82.14
Glass0.705 ± 0.0030.6780.6310.636 0.615
4.804.45.87.06.96
Wine0.963 ± 0.0040.9610.9140.914 0.9550.897
7.526.96.07.54.386.0
Australian0.850 ± 0.0020.8460.8300.828 0.773
6.984.73.73.04.0
German number0.725 ± 0.0080.7130.6820.658 0.701
11.4610.512.210.81.0
Zoo0.920 ± 0.0220.9540.9490.980 0.954
9.0611.09.013.011.0
Breast cancer0.967 ± 0.0030.9650.9510.949 0.9400.930
6.144.36.106.105.05.0
DNA0.810 ± 0.0060.8310.8220.823 0.7600.760
89.2671.818.820.696.091.0
Vehicle0.721 ± 0.0080.6530.6860.673
10.149.110.810.7
Sonar0.850 ± 0.0020.827 0.7290.786
28.9620.05.8510.0
Hillvalley0.535 ± 0.0030.575 0.605
45.1840.026.0
Musk 10.852 ± 0.0030.815 0.835
81.3459.358.0
Splice0.778 ± 0.002 0.6800.670
25.6628.028.0
KRFP0.896 ± 0.001 0.842 * 0.884 *
2390.26.01866
* evaluated using weka library.
Table 6. Comparison of MOTiFS with other methods, according to 3-NN.
Table 6. Comparison of MOTiFS with other methods, according to 3-NN.
DatasetAvg. Acc.
# Sel. Feat.
Avg. Acc.
MOTiFSSCE [39]CCE [39]PDE-2 [39]
Soybean-small0.988 ± 0.0151.0001.0001.000
12.18
Liver-disorders0.645 ± 0.0120.6020.5920.590
3.94
Credit0.845 ± 0.0070.6460.6540.659
8.22
Tic-tac-toe0.794 ± 0.0060.7740.7470.757
7.18
Libras movement0.807 ± 0.0110.5380.5520.554
44.94

Share and Cite

MDPI and ACS Style

Chaudhry, M.U.; Lee, J.-H. MOTiFS: Monte Carlo Tree Search Based Feature Selection. Entropy 2018, 20, 385. https://doi.org/10.3390/e20050385

AMA Style

Chaudhry MU, Lee J-H. MOTiFS: Monte Carlo Tree Search Based Feature Selection. Entropy. 2018; 20(5):385. https://doi.org/10.3390/e20050385

Chicago/Turabian Style

Chaudhry, Muhammad Umar, and Jee-Hyong Lee. 2018. "MOTiFS: Monte Carlo Tree Search Based Feature Selection" Entropy 20, no. 5: 385. https://doi.org/10.3390/e20050385

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop