Next Article in Journal
Editorial Board Members’ Collection Series: Biomimetic Design, Constructions and Devices in Times of Change I
Previous Article in Journal
Autonomous Robot Task Execution in Flexible Manufacturing: Integrating PDDL and Behavior Trees in ARIAC 2023
Previous Article in Special Issue
YOLO-LFPD: A Lightweight Method for Strip Surface Defect Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multimodal Multi-Objective Feature Selection Method for Intelligent Rating Models of Unmanned Highway Toll Stations

1
Intelligent Transportation System Research Center, Southeast University, Nanjing 211189, China
2
Logistics Research Center, Shanghai Maritime University, Shanghai 201306, China
*
Author to whom correspondence should be addressed.
Biomimetics 2024, 9(10), 613; https://doi.org/10.3390/biomimetics9100613
Submission received: 19 August 2024 / Revised: 29 September 2024 / Accepted: 8 October 2024 / Published: 10 October 2024

Abstract

:
To facilitate the intelligent classification of unmanned highway toll stations, selecting effective and useful features is pivotal. This process involves achieving a tradeoff between the number of features and the classification accuracy while also reducing the acquisition costs of features. To address these challenges, a multimodal multi-objective feature selection (MMOFS) method is proposed in the current study. In the MMOFS, we utilize a multimodal multi-objective evolutionary algorithm to choose features for the unmanned highway toll station classification model and use the random forest method for classification. The primary contribution of the current study is to propose a feature selection method specifically designed for the classification model of unmanned highway toll stations. Experimental results using actual data from highway toll stations demonstrate that the proposed MMOFS outperforms the other two competitors in terms of PSP, HV, and IGD. Furthermore, the proposed algorithm can provide decision-makers with multiple equivalent feature selection schemes. This approach achieves a harmonious balance between the model complexity and the classification accuracy based on actual scenarios, thereby providing guidance for the construction of unmanned highway toll stations.

1. Introduction

Highways are important components of the transportation system. However, congestion on expressways can lead to significant environmental and socio-economic issues [1]. Analyzing highway congestion reveals that severe congestion is often caused by the presence of manual toll booths [2]. Numerous studies have been studying the different aspects of highway tolls. For example, Velarde et al. [3] set motorway toll prices to control traffic flow and reduce congestion. Moreover, Gui et al. [4] adjusted the capacity of toll booths to control traffic flow. Ling et al. [5] proposed an optimization model for toll station lane configuration to minimize the total cost, which contains toll station operating cost and delay cost, combined with the queuing theory. The results demonstrate that the proposed method is effective in reducing congestion at toll booths. Rota et al. [6] highlighted that the intelligent traffic system (ITS) combined with emerging unmanned ground vehicles (UGVs) and novel tolling methods can significantly alleviate these bottlenecks. Souza [7] pointed out that the free-flow system, an electronic system for automatic vehicle identification, improved toll collection efficiency by electronically identifying and charging passing vehicles automatically. The results indicate that this system can improve toll collection efficiency. Therefore, implementing free-flow tolling is an effective approach to reduce highway congestion and improve traffic efficiency. However, due to various special situations at highway toll stations, such as vehicles benefiting from the “green channel” preferential policy and incorrect vehicle information, it is not realistic to have all highway toll stations fully unmanned. This underscores the need for intelligent rating models using various machine learning methods to assess the unmanned level of highway toll stations and determine the types of each unmanned highway toll station. Additionally, in the context of future developments in intelligent transportation methods and technologies, the intelligent classification model of unmanned highway toll stations can be valuable for prioritizing the implementation of free-flow tolling.
To improve the classification performance of machine learning methods in the field of intelligent transportation, feature selection is a fundamental step, as this process effectively reduces data dimensionality, shortens learning time, and improves classification performance [8]. The feature selection methods have been widely used in the field of intelligent transportation. For example, Kandiri et al. [9] used an interconnected optimization algorithm for feature selection to identify the best set of features, thereby improving the prediction accuracy of travel time in intelligent transportation systems. Zheng et al. [10] proposed a hybrid intelligent algorithm-based feature selector to optimize original state vectors. The results verify that this feature selection method can significantly enhance the predicted accuracy of the model. Liu et al. [11] proposed a community-based dandelion algorithm (CDA) to solve the traffic flow prediction problem. The experiment results show that the proposed method can select suitable features via the coding and decoding strategies and improve the prediction accuracy by 5–16%. For a large training set, Pareek et al. [12] proposed a feature selection method using a hybrid slime procedure and whale optimization, which is used for the selection of important variables and the elimination of duplicates, and then the extreme machine learning (ELM) is used for prediction. The experimental results demonstrated that the proposed method has good prediction performance and high prediction accuracy. Therefore, the feature selection can improve classification accuracy. Similar to the above studies, Wahab et al. [13] introduced a feature selection algorithm to delete irrelevant and abnormal features in advance and to cluster the intelligent features of the vehicle’s self-organizing network. Almutlaq et al. [14] proposed a hybrid feature selection method that combines three filter-based methods and two wrapper-based methods. The proposed method aids in eliminating irrelevant features in the intrusion detection of intelligent transportation systems. The results show that the method has faster prediction speed and improved classification accuracy. Kavitha et al. [15] proposed an optimized YOLOv2 model for vehicle detection and classification, incorporating a multi-layer feature fusion strategy to improve the effectiveness of feature selection. The experimental results demonstrate that the proposed method can achieve accurate vehicle classification. In addition to using feature selection methods to improve the classification performance of models, many studies also investigate classification models in the field of transportation [16]. For instance, Arinaldi et al. [17] proposed a classification method based on support vector machines (SVMs) and faster region convolutional neural networks (RCNNs) to identify vehicles. Sarikan et al. [18] used a classification method based on K-nearest neighbors (KNNs) and decision trees to classify vehicles. Like the above studies, Barreyro et al. [19] utilized an AlexNet convolutional neural network to identify automatic vehicles. The results show that the classification accuracy of the proposed algorithm can meet actual requirements. Machine learning classifiers have been applied for vehicle classification. Trivedi et al. [20] pointed out that the application of machine learning methods for vehicle classification is effective.
From the above-mentioned studies, feature selection is often treated as a single-objective optimization problem. However, it can also be viewed as a multi-objective combinatorial optimization problem, mainly involving two objectives: (1) the number of selected features and (2) the classification accuracy. To solve this problem, Liang et al. [21] proposed a feature selection algorithm based on a multi-objective evolutionary algorithm (FS-MOEA), in which the information gain (IG)–analytic hierarchy process (AHP) is used to prioritize the search for the best feature subset. The experimental results verify that the proposed algorithm achieves higher classification accuracy and lower computational complexity, making it suitable for use in intrusion detection systems (IDSs) in vehicular self-organizing networks (VANETs). Bohrer et al. [22] proposed a hybrid feature selection approach using a multi-objective genetic algorithm. The proposed method uses a multi-objective genetic algorithm to generate a feature set by combining and optimizing the feature sets generated by other conventional feature selection methods. The experimental results demonstrate that the proposed algorithm is highly effective. To further improve the algorithm performance in solving multi-objective feature selection problems, Zhang et al. [23] proposed a novel information gain-based evolutionary algorithm for multi-objective feature selection. Note that the proposed algorithm employs the information gain as a metric to evaluate the contribution of features. Dong et al. [24] proposed a many-objective optimization-based multi-label feature selection algorithm (MMFS) in which the real number encoding method is used to design new crossover operators and mutation operators. The results demonstrate the superiority of the algorithm. Xue et al. [25] proposed a multi-objective binary genetic algorithm integrating an adaptive operator selection mechanism (MOBGA-AOS) for feature selection. To fully utilize the search capabilities of different crossover operators in the genetic algorithms (GAs), the proposed algorithm uses five different crossover operators to generate a set of optimal feature subsets characterized by small size and high classification accuracy. The experimental results show that the proposed algorithm has significant advantages in handling large-scale datasets.
Besides conflicting objectives, real-world feature selection also exhibits multimodality [26,27]. In other words, the feature selection can be considered as a multimodal multi-objective optimization problem. Compared with ordinary multi-objective optimization problems, multimodal multi-objective optimization not only requires finding a high-quality Pareto front approximation in the objective space but also demands finding sufficient equivalent Pareto solutions in the decision space [28]. To solve the above issue, Yue et al. [29] proposed a multimodal multi-objective algorithm for feature selection. The experiments verify that the proposed algorithm can provide a high-quality feature subset without significantly reducing the classification accuracy. Hu et al. proposed a novel multimodal niching particle swarm optimization (MNPSO) algorithm to select features. The proposed algorithm employs the crowding distance and species clustering methods to divide the population to improve population diversity. The experimental results demonstrate that the proposed algorithm can effectively find more multimodal feature selection solutions [30]. Subsequently, Liang et al. [31] proposed a multimodal multi-objective genetic algorithm for feature selection, which can successfully find equivalent feature subsets on different datasets. Wang et al. [32] proposed a multi-objective differential evolution approach in which a novel feature relevance-based population initialization method is proposed to improve the search performance. The experimental results show that the proposed algorithm can obtain a greater number of higher-quality feature subsets. Jha et al. [33] used a multimodal multi-objective optimization algorithm based on a ring topology structure to perform filtering feature selection based on factors such as mutual information and redundancy between features, ultimately selecting feature subsets with minimum redundancy and maximum correlation. The experimental results show that the proposed algorithm can not only provide a larger number of equivalent feature subsets but also has better or similar prediction accuracy when compared with other feature selection methods.
From existing studies, although multimodal multi-objective feature selection methods have been used in various fields, they have rarely been utilized for feature selection problems in intelligent transportation systems, especially for highway toll stations. To find more equivalent feature selection schemes for the unmanned highway toll rating model and investigate the relationship between the model classification accuracy and the number of features, a multimodal multi-objective feature selection (MMOFS) method is introduced in the current study. In the MMOFS, a competitive multimodal multi-objective evolutionary algorithm is employed to explore diverse feature selection schemes in the objective space, elucidating the relationship between the model classification accuracy and the number of selected features. Moreover, it is employed to discover more equivalent solutions in the decision space, suitable for adapting to different scenarios and requirements. Additionally, the random forest (RF) algorithm [34] is used to evaluate the effectiveness of feature selection solutions. The performance of the MMOFS is compared with that of two selected algorithms on a real-world dataset. The results demonstrate that the proposed algorithm can provide decision-makers with high-quality and diverse feature selection schemes.
The main contributions of the present study are as follows:
(1)
To reduce highway congestion and achieve free-flow tolling, this study uses artificial intelligence methods to develop an intelligent rating model for unmanned highway toll stations based on real-world toll station data. Moreover, the level of unmanned highway toll stations is categorized into three levels.
(2)
A multimodal multi-objective feature selection method is employed to facilitate feature selection, providing multiple high-quality and equivalent feature selection schemes. This approach aids decision-makers in model development under various conditions while reducing modeling costs, thereby providing a reliable basis for the construction of unmanned toll stations.

2. Methodology

In multi-objective feature selection problems, there is often a conflict between the number of selected features (f1) and the classification accuracy of the model (f2). Therefore, they can be considered as two conflicting optimization objectives, which can be defined as follows:
min f 1 x = 1 D i = 1 D x i min f 2 x = E R x ,
where x represents a feature subset; D denotes the total number of original features; ER is the classification error rate. For the multi-objective feature selection, different feature subsets may be equivalent in some scenarios. This means that selecting different feature schemes may not affect the classification performance of models. Note that this equivalence can significantly reduce the cost of acquiring features. Therefore, the feature selection problems are treated as multimodal multi-objective problems to effectively select features for the intelligent rating model of highway toll stations. Additionally, compared with the filter and embedded methods, the wrapper method can select superior feature subsets [35]. Therefore, the wrapper method is utilized to select features.

2.1. Encoding and Decoding Methods

Because the particle swarm optimization, used as the search engine, is suitable for solving continuous optimization problems, it cannot be used to select features in the present study. To solve this issue, like the individual encoding method in Refs. [36,37], the current study also uses the real number encoding method. Namely, each decision variable ranges between 0 and 1. If a decision variable is greater than or equal to 0.5, the corresponding feature is selected; otherwise, this feature will be removed. The specific encoding and decoding process for individuals is illustrated in Figure 1. As shown in Figure 1, this individual has six features. Among them, features x2, x3, and x4 are greater than or equal to 0.5, indicating that these three features will be selected. Conversely, features x1, x5, and x6 are less than 0.5 and are thus not selected.

2.2. IDMMPSO

Since the improved discrete multimodal multi-objective particle swarm optimization (IDMMPSO) algorithm [36] is both effective for solving discrete multimodal multi-objective optimization problems and easily accessible, it is used to address the multimodal multi-objective feature selection problems. The main steps of the IDMMPSO are as follows (Algorithm 1): Line 1 is to generate an initial population. Subsequently, compute the fitness function values of all initial individuals (line 2). The INSCD method [36], which uses the Hamming distance to compute the crowding distance in the non-dominated_scd_sort method, is employed to sort all individuals into two archives, i.e., HOA and NOA (line 5). Select the first individuals from HOA{i}, denoted as pbesti. Moreover, choose the first individual from NOA{i} and denote it as nbesti. The velocity formula is used to update the i-th particle (see line 7). Lines 8–14 are to convert real-number into 0–1 integer. After generating i-th individuals, calculate its fitness function value and save it to HOA{i}. Line 16 is to use the INSCD method to update the HOA{i}. Subsequently, the environmental selection method is utilized to choose non-dominated individuals from HOA{i − 1}, HOA{i}, and HOA{i + 1}, and save them to NOA{i}. Steps in lines 4–19 are repeated until G = Gmax. Finally, output all non-dominated individuals in the NOA.
Algorithm 1 IDMMPSO
Input: the population size, NP; the maximum number of generations, Gmax; the dimension of individual, D; the historical optimal archive, HOA; the neighbor optimal archive, NOA.
1:          Generate an initial population P0; set G = 1;
2:          Compute the fitness function values of all individuals in P0;
3:          while G < Gmax do
4:                     for i = 1: NP do
5:                          The INSCD method is used to sort all individuals in both HOA and 
                             NOA;
6:                          Select the first individuals from HOA{i} and NOA{i}, respectively, and 
                             denote them as pbesti and nbesti;
7:                          the velocity of the i-th particle is updated via 
                              v i G + 1 = ω v i G + c 1 r 1 ( p b e s t i G x i G ) + c 2 r 2 ( n b e s t i G x i G ) ;
8:                          for j = 1: D do
9:                                if  r a n d   <   logsig ( v i , j G + 1 ) then
10:                                   x i , j G + 1 = 1 ;
11:                             else
12:                                   x i , j G + 1 = 0 ;
13:                             end if
14:                        end for
15:                        Calculate the fitness function value of the  x i G + 1  and save it to HOA{i};
16:                        The INSCD method is utilized to update the HOA{i};
17:                        The environmental selection method is used to choose non-dominated
                              individuals from HOA{i − 1}, HOA{i}, and HOA{i + 1} and save them to
                              NOA{i};
18:                  end for
19:                      G = G + 1;
20:        end while 
Output:  All non-dominated individuals in NOA.

2.3. Random Forest

The RF proposed by Breiman [38] is a competitive and effective classifier method. The pseudocode of the RF is shown in Algorithm 2. The first step is to generate M randomized training sets (see line 1). For the k-th dataset (denoted as Dk), randomly select a subset of features (i.e., F′) from original features. Note that  F  should be less than N. Subsequently, line 4 is to build a decision tree using Dk and F′. After building all the decision trees, a voting mechanism is employed to aggregate all their predicted results.
Algorithm 2 RF
Input: the dataset, D; the number of trees, M; the number of features in the original dataset, N
1:for k = 1: M do
2:      The bootstrap sampling method is used to generate a training set, Dk;
3:        Randomly select a subset of features, F′;  F  < N 
4:        Build a decision tree (Tk) based on the Dk and F′;
5:  end for
6:  Aggregate predicted results from all decision trees;
7:Output: the final classification predicted result. 

2.4. Overall Implementation of the Proposed Algorithm

In this study, a multimodal multi-objective feature selection method (MMOFS) is proposed to effectively develop an intelligent classification model for unmanned highway toll stations and to obtain multiple equivalent feature subsets that accommodate diverse scenario demands and varying feature acquisition costs. In the proposed MMOFS, its main steps are the same as the IDMMPSO, with the primary differentiation found in the computation of objective functions. Specifically, one of the objective functions in the MMOFS is the classification accuracy obtained from the RF algorithm. The main steps of the MMOFS are shown in Algorithm 3 in Section 2.2. Line 1 aims to produce an initial population. The objective function values (i.e., classification accuracy) of all individuals in the P0 are obtained by the RF algorithm. Note that the other objective function is the number of selected features. Lines 3–19 describe the use of IDMMPSO to identify feature selection schemes for the highway toll station rating model, as detailed in Section 2.2. Finally, output all feature selection schemes.
Algorithm 3 MMOFS
Input: the population size, NP; the maximum number of generations, Gmax; the dimension of individual, D; the historical optimal archive, HOA; the neighbor optimal archive, NOA.
1:          Generate an initial population P0; set G = 1;
2:          Compute the fitness function values of all individuals in P0 via the RF;
3:          while G < Gmax do
4:          for i = 1: NP do
5:                The INSCD method is utilized to rank all individuals in both HOA and NOA;
6:                Choose the first individual from HOA{i} and NOA{i}, respectively, and denote them as pbesti and nbesti;
7:                 v i G + 1 = ω v i G + c 1 r 1 ( p b e s t i G x i G ) + c 2 r 2 ( n b e s t i G x i G )  is used to update the velocity of
                   the i-th individual;
8:             for j = 1: D do
9:                     if  r a n d   <   logsig ( v i , j G + 1 ) then
10:                             x i , j G + 1 = 1 ;
11:                     else
12:                             x i , j G + 1 = 0 ;
13:                     end if
14:             end for
15:                Calculate the fitness function value of the  x i G + 1  using the RF algorithm and save it to HOA{i};
16:                All individuals in the HOA{i} are ranked using the INSCD method. Moreover, a 
                     certain number of individuals are selected and saved to HOA{i};
17:                Select non-dominated individuals from HOA{i − 1}, HOA{i}, and HOA{i + 1}
                     using the environmental selection method and store them in NOA{i};
18:             end for
19:          end while
Output: All the non-dominated individuals in NOA

3. Experimental Results and Analyses

3.1. Datasets

Real highway toll station data are used in the present study, which includes data from 13 highway toll stations, comprising 117 data samples and 23 features. Note that we only use 12 features listed in Table 1 due to the lack of relevant data for 11 out of the 23 features. Moreover, the unmanned highway toll stations are classified into three levels: (1) unmanned exit or entrance (U1); (2) unmanned exit and entrance (U2); (3) unmanned highway toll station (U3). For U1, the highway toll station has either the exit or entrance unmanned, with remote monitoring employed for the unattended exit or entrance. U2 involves an unmanned exit and entrance, with staff conducting remote monitoring from a smart cloud storage facility either in the station building or nearby. In contrast, U3 toll stations do not have a monitoring center.

3.2. Parameter Settings

The actual dataset is randomly split into training and test sets, with the training set comprising 80% of the total samples and the test set comprising the remaining 20%. Moreover, the three-fold cross-validation is used to train the model. For the RF algorithm, the number of trees and the minimum number of samples required for a leaf node are set to twenty and two, respectively. Additionally, the parameter settings in the IDMMPSO algorithm are the same as those in Ref. [36].

3.3. Comparisons with Competitive Algorithms

To demonstrate the feature selection performance of the proposed algorithm, it is compared with two other algorithms: MOPSO-based [39] and NSGA-II-based [40] random forest algorithms, named MOPSO-RF and NSGA-II-RF. Moreover, the performances of these compared algorithms are evaluated using three commonly used multi-objective performance metrics: the Pareto set proximity(PSP) [41], the hypervolume (HV) [42], and inverted generational distance (IGD) [43]. Note that the IGD value is computed using the Pareto front (PF) approximation obtained from each compared algorithm and a reference PF, which is selected from all the non-dominated solutions of all the compared algorithms via the environmental selection method [36]. Additionally, each compared algorithm is run five times on this feature selection problem. According to Ref. [44], the unpaired t-test method is used to analyze the performance of all the compared algorithms since the sample is unpaired and meets the normal distribution. The symbols “+”, “−”, and “≈” represent that the proposed MOPSO-RF is superior to, inferior to, and comparable with its competitors, respectively.
The mean and standard deviation values for all compared algorithms are shown in Table 2. It can be observed from Table 2 that the performance of the MOPSO-RF is significantly better than that of the other two competitors in terms of the PSP. Because the PSP is mainly used to assess the performance of the compared algorithm in the decision space, we can conclude that the proposed algorithm can find more equivalent feature selection solutions when compared with the MOPSO-RF and the NSGA-II-RF. Moreover, Table 2 indicates that the MOPSO-RF outperforms two compared algorithms in a statistical method in terms of HV and IGD. This means that the proposed algorithm is able to achieve higher-quality feature selection schemes. Therefore, it can be concluded that the proposed algorithm is an effective and competitive approach to selecting features for decision-makers.
To visually demonstrate the performance of the MMOFS, the PFs and Pareto sets (PSs) obtained by the three compared algorithms are shown in Figure 2 and Figure 3. From Figure 2, it is evident that, except for cases where the number of selected features is one, two, or three, the proposed algorithm can achieve a high-quality PF approximation when compared with the NSGA-II-RF. Moreover, the proposed algorithm consistently identifies high-quality solution sets in all cases when compared with the MOPSO-RF. Finally, Figure 2 also reveals that, for the intelligent rating model of unmanned highway toll stations, there is an inherent trade-off between the model accuracy and the number of selected features. From Figure 3, we can observe that, compared with the MOPSO-RF and the NSGA-II-RF, the MMOFS can find more equivalent feature subsets in most cases when the number of selected features is the same. The main reason may be that the INSCD method can assist the proposed algorithm in preserving the population diversity and locating more equivalent solutions. This indicates that the proposed algorithm has stronger feature selection capability and provides decision-makers with more equivalent feature selection schemes for use in different scenarios. Therefore, it can be concluded that the proposed algorithm is an effective approach to select features.

3.4. Multimodal Analysis of Feature Selection Schemes

To further illustrate the effectiveness of the proposed algorithm, this experiment analyzes the multimodality of feature selection schemes. All feature selection schemes obtained by the proposed algorithm are presented in Table 3. We can observe from Table 3 that for the same number of selected features and model classification accuracy, the proposed algorithm can provide equivalent feature selection schemes in most cases. For example, feature selection schemes {x1, x4, x7, x9, x11} and {x4, x5, x7, x11, x12} are equivalent when the number of features is five, and the ER value is 0.02. However, given that the acquisition cost of feature x12 (i.e., peak daily volume) is higher than that of other features in practice, we can adopt feature selection solution {x1, x4, x7, x9, x11} to build the intelligent rating model of highway toll stations without increasing its complexity and compromising its classification performance. This significantly reduces the cost of feature acquisition. Additionally, not all highway toll stations have complete special condition data; thus, equivalent feature selection schemes can be used to evaluate their levels. It can be concluded that the proposed algorithm can find equivalent feature selection schemes, thereby reducing modeling costs.

3.5. Influence of the Selected Performance Indicators

In the original MMOFS, the ER is used to assess the effectiveness of features. To analyze the influence of performance metrics, the Kappa coefficient method and macro-averaged precision, which are used to evaluate the classification performance of algorithms on multi-class problems, are employed in the MMOFS, named as the MMOFS-1 and MMOFS-2. Moreover, the parameter settings of the MMOFS-1 and the MMOFS-2 are the same as in Section 3.2.
The results of the MMOFS and its two variants are illustrated in Figure 4. We can observe from Figure 4 that using different performance indicators in the proposed algorithm can lead to different feature selection schemes. Therefore, the performance of the proposed algorithm is influenced by the selected performance metric. In other words, like other feature selection methods, the proposed MMOFS is also sensitive to specific models. Additionally, Figure 4 indicates that the ER can help the proposed algorithm find more feature selection solutions; thus, the ER is used in the proposed algorithm.

4. Discussion

The unmanned highway toll station can greatly improve traffic efficiency, but its construction must determine the unmanned level based on actual conditions. Therefore, developing a classification model for unmanned highway toll stations is essential, as its performance can be substantially affected by feature selection schemes. In this study, a multimodal multi-objective feature selection (MMOFS) approach is introduced to provide a set of feature selection schemes for decision-makers.
The findings in Section 3.3 demonstrate that the MMOFS is superior to the NSGA-II and the MOPSO in finding feature selection solutions. The results can be discussed from two perspectives: (1) For the decision space, the PSP values in Table 2 show that the MMOFS can assist the RF in finding more equivalent feature schemes when compared with the other two competitors. This can not only improve the algorithm’s adaptability to different scenarios, such as missing data, but also reduce the cost of feature acquisition, including the use of expensive data collection equipment. (2) In terms of the objective space, we can observe from Table 2 that the proposed MMOFS can help the RF achieve better feature selection solutions, which can influence the classification accuracy and the model complexity. Therefore, the MMOFS is an effective method for selecting features for the classification model of unmanned highway toll stations. Additionally, the application of the MMOFS is not limited to highway toll stations; it is also suitable for other classification challenges in the transportation field and beyond.
The results in Section 3.4 show that the proposed algorithm not only achieves high classification accuracy but also finds more equivalent feature schemes. Clearly, the RF method, incorporated into the MMOFS, exhibits high classification accuracy, providing a solid foundation for constructing unmanned highway toll stations. This allows decision-makers to easily determine the unmanned level of highway toll stations via actual conditions. Moreover, the proposed MMOFS can locate more equivalent feature selection schemes, which can improve the adaptability and applicability of classifiers, as well as their decision-making efficiency. Additionally, for different highway toll stations facing issues of insufficient historical data, we can adopt equivalent feature selection schemes based on existing data for modeling. This can significantly reduce the investment costs of additional equipment and shorten the modeling time. Finally, we can observe that the number of selected features conflicts with the classification performance; specifically, the classification performance of the algorithm improves as the number of selected features increases. Considering that each highway toll station has different data collection devices, each station can select a feature scheme specifically tailored to its unique conditions, allowing for a customized approach to constructing the classification model.
The MMOFS is an effective and competitive feature selection method. However, it has some limitations: (1) Although the proposed MMOFS can identify a larger quantity and higher quality of feature selection schemes when compared with the other two competitors, it fails to find any solutions when the number of selected features is seven or eight. Therefore, the search performance of the proposed algorithm needs further improvement. (2) Data imbalance presents a significant challenge for classification problems, and the proposed algorithm does not currently address this issue, which may affect its overall applicability in such scenarios. (3) Different fields require distinct feature selection methods. However, this study focuses solely on a particular scenario, limiting the capability of the MMOFS to evolve autonomously.

5. Conclusions

To find more equivalent and diverse feature selection schemes for unmanned highway toll station rating models that can adapt to changing scenarios and reduce feature acquisition costs, the multimodal multi-objective feature selection (MMOFS) method is proposed in the current study. In the MMOFS, an improved multimodal multi-objective evolutionary algorithm is utilized to find feature selection schemes. Moreover, the RF algorithm is used to evaluate the effectiveness of these feature selection schemes. Based on real-world highway toll station data, the proposed algorithm is compared with two other algorithms. The experimental results demonstrate that the proposed algorithm can identify more high-quality feature selection schemes due to its effective maintenance of population diversity. This provides decision-makers with diverse feature selection options and can reduce the cost of feature acquisition, thereby offering valuable references for the construction of unmanned highway toll stations.

Author Contributions

Conceptualization, Z.G., H.M. and Q.F.; methodology, Z.G., H.M., Z.Y. and Q.F.; investigation, Z.G., H.M., Z.Y. and Q.F.; writing—original draft preparation, Z.G.; writing—review and editing, Z.G. and Q.F.; visualization, Z.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the Humanities and Social Sciences on Planning Foundation of the Ministry of Education in China (Research on cross-domain collaborative maritime unmanned search and rescue methods and strategies, 23YJAZH029), the Shanghai Pujiang Program (No. 22PJD030), and the National Nature Science Foundation of China (No. 61603244).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data supporting reported results are available from the authors upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ha, P.; Chen, S.; Dong, J.; Labi, S. Leveraging vehicle connectivity and autonomy for highway bottleneck congestion mitigation using reinforcement learning. Transp. A Transp. Sci. 2023, 1–26. [Google Scholar] [CrossRef]
  2. Petrović, A.; Nikolić, M.; Bugarić, U.; Delibašić, B.; Lio, P. Controlling highway toll stations using deep learning, queuing theory, and differential evolution. Eng. Appl. Artif. Intell. 2023, 119, 105683. [Google Scholar] [CrossRef]
  3. González Velarde, J.; Camacho-Vallejo, J.; Pinto Serrano, G. A scatter search algorithm for solving a bilevel optimization model for determining highway tolls. Comput. Y Sist. 2015, 19, 5–16. [Google Scholar] [CrossRef]
  4. Yong, G.; Zhang, J.; Su, N.; Feng, X. Capacity design and pareto improvement of highway toll plaza in a competitive transport system. IEEE Access 2021, 9, 113642–113652. [Google Scholar] [CrossRef]
  5. Lin, P.; Liang, Y. Modeling and solving for lane type setting problem in highway toll station. J. Transp. Syst. Eng. Inf. Technol. 2020, 20, 152. [Google Scholar]
  6. Rota, B.; Simic, M. Traffic flow optimization on freeways. Procedia Comput. Sci. 2016, 96, 1637–1646. [Google Scholar] [CrossRef]
  7. Souza, B.; da Costa, G.; Szejka, A.; Freire, R.; Gonzalez, G. A deep learning-based approach for axle counter in free-flow tolling systems. Sci. Rep. 2024, 14, 3400. [Google Scholar] [CrossRef] [PubMed]
  8. Gheyas, I.; Smith, L. Feature subset selection in large dimensionality domains. Pattern Recognit. 2010, 43, 5–13. [Google Scholar] [CrossRef]
  9. Kandiri, A.; Ghiasi, R.; Nogal, M.; Teixeira, R. Travel time prediction for an intelligent transportation system based on a data-driven feature selection method considering temporal correlation. Transp. Eng. 2024, 18, 100272. [Google Scholar] [CrossRef]
  10. Zheng, L.; Zhu, C.; Zhu, N.; He, T.; Dong, N.; Huang, H. Feature selection-based approach for urban short-term travel speed prediction. IET Intell. Transp. Syst. 2018, 12, 474–484. [Google Scholar] [CrossRef]
  11. Liu, X.; Qin, X.; Zhou, M.; Sun, H.; Han, S. Community-based dandelion algorithm-enabled feature selection and broad learning system for traffic flow prediction. IEEE Trans. Intell. Transp. Syst. 2023, 25, 2508–2521. [Google Scholar] [CrossRef]
  12. Pareek, P.; Vidyalakshmi, K.; Nayana, S.; Gangambika, G.; Deepak, H. Forecasting of Traffic Flow Using Feature Selection with ML Model. In Proceedings of the 2023 International Conference on Data Science and Network Security (ICDSNS), Tiptur, India, 28–29 July 2023; pp. 1–6. [Google Scholar]
  13. Wahab, O.; Mourad, A.; Otrok, H.; Bentahar, J. CEAP: SVM-based intelligent detection model for clustered vehicular ad hoc networks. Expert Syst. Appl. 2016, 50, 40–54. [Google Scholar] [CrossRef]
  14. Almutlaq, S.; Derhab, A.; Hassan, M.M.; Kaur, K. Two-stage intrusion detection system in intelligent transportation systems using rule extraction methods from deep neural networks. IEEE Trans. Intell. Transp. Syst. 2022, 24, 15687–15701. [Google Scholar] [CrossRef]
  15. Kavitha, N.; Chandrappa, D. Optimized YOLOv2 based vehicle classification and tracking for intelligent transportation system. Results Control. Optim. 2021, 2, 100008. [Google Scholar]
  16. Ghosh, A.K. On nearest neighbor classification using adaptive choice of k. J. Comput. Graph. Stat. 2007, 16, 482–502. [Google Scholar] [CrossRef]
  17. Arinaldi, A.; Pradana, J.A.; Gurusinga, A.A. Detection and classification of vehicles for traffic video analytics. Procedia Comput. Sci. 2018, 144, 259–268. [Google Scholar] [CrossRef]
  18. Sarikan, S.S.; Ozbayoglu, A.M.; Zilci, O. Automated vehicle classification with image processing and computational intelligence. Procedia Comput. Sci. 2017, 114, 515–522. [Google Scholar] [CrossRef]
  19. Barreyro, J.; Yoshioka, L.; Marte, C. A non-intrusive category identification method based on the binary image of profile vehicles and CNN classification algorithm. In Proceedings of the 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), Indianapolis, IN, USA, 19–22 September 2021; pp. 1803–1808. [Google Scholar]
  20. Trivedi, J.; Devi, M.; Solanki, B. Step towards intelligent transportation system with vehicle classification an recognition using speeded-up robust features. Arch. Tech. Sci. Arh. Za Teh. Nauk. 2023, 1, 39–56. [Google Scholar] [CrossRef]
  21. Liang, J.; Ma, M. FS-MOEA: A novel feature selection algorithm for IDSs in vehicular networks. IEEE Trans. Intell. Transp. Syst. 2020, 23, 368–382. [Google Scholar] [CrossRef]
  22. Bohrer, J.; Dorn, M. Enhancing classification with hybrid feature selection: A multi-objective genetic algorithm for high-dimensional data. Expert Syst. Appl. 2024, 255, 124518. [Google Scholar] [CrossRef]
  23. Zhang, B.; Wang, Z.; Li, H.; Lei, Z.; Cheng, J.; Gao, S. Information gain-based multi-objective evolutionary algorithm for feature selection. Inf. Sci. 2024, 677, 120901. [Google Scholar] [CrossRef]
  24. Dong, H.; Sun, J.; Sun, X.; Ding, R. A many-objective feature selection for multi-label classification. Knowl. Based Syst. 2020, 208, 106456. [Google Scholar] [CrossRef]
  25. Xue, Y.; Zhu, H.; Liang, J.; Słowik, A. Adaptive crossover operator based multi-objective binary genetic algorithm for feature selection in classification. Knowl. Based Syst. 2021, 227, 107218. [Google Scholar] [CrossRef]
  26. Fan, Q.; Ersoy, O. Zoning search with adaptive resource allocating method for balanced and imbalanced multimodal multi-objective optimization. IEEE/CAA J. Autom. Sin. 2021, 8, 1163–1176. [Google Scholar] [CrossRef]
  27. Miao, Z.; Huang, W.; Jiang, Q.; Fan, Q. A novel multimodal multi-objective optimization algorithm for multi-robot task allocation. Trans. Inst. Meas. Control 2023, 01423312231183588. [Google Scholar] [CrossRef]
  28. Fan, Q.; Yan, X. Solving multimodal multiobjective problems through zoning search. IEEE Trans. Syst. Man Cybern. Syst. 2019, 51, 4836–4847. [Google Scholar] [CrossRef]
  29. Yue, C.; Liang, J.; Qu, B.; Yu, K.; Song, H. Multimodal multiobjective optimization in feature selection. In Proceedings of the 2019 IEEE Congress on Evolutionary Computation (CEC), Wellington, New Zealand, 10–13 June 2019; pp. 302–309. [Google Scholar]
  30. Hu, X.; Zhang, S.; Li, M.; Deng, J. Multimodal particle swarm optimization for feature selection. Appl. Soft Comput. 2021, 113, 107887. [Google Scholar] [CrossRef]
  31. Liang, J.; Yang, J.; Yue, C.; Li, G.; Yu, K.; Qu, B. A multimodal multiobjective genetic algorithm for feature selection. In Proceedings of the 2022 IEEE Congress on Evolutionary Computation (CEC), Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar]
  32. Wang, P.; Xue, B.; Liang, J.; Zhang, M. Multiobjective differential evolution for feature selection in classification. IEEE Trans. Cybern. 2021, 53, 4579–4593. [Google Scholar] [CrossRef]
  33. Jha, K.; Saha, S. Incorporation of multimodal multiobjective optimization in designing a filter based feature selection technique. Appl. Soft Comput. 2021, 98, 106823. [Google Scholar] [CrossRef]
  34. Li, X.; Chen, W.; Zhang, Q.; Wu, L. Building auto-encoder intrusion detection system based on random forest feature selection. Comput. Secur. 2020, 95, 101851. [Google Scholar] [CrossRef]
  35. Sánchez-Maroño, N.; Alonso-Betanzos, A.; Tombilla-Sanromán, M. Filter methods for feature selection–a comparative study. In Proceedings of the International Conference on Intelligent Data Engineering and Automated Learning, Birmingham, UK, 16–19 December 2007; pp. 178–187. [Google Scholar]
  36. Li, X.; Li, M.; Yu, M.; Fan, Q. Fault reconfiguration in distribution networks based on improved discrete multimodal multi-objective particle swarm optimization algorithm. Biomimetics 2023, 8, 431. [Google Scholar] [CrossRef]
  37. Wang, Y.; Huang, J.; Fan, Q. A two-stage feature selection method based on causal model and multimodal multi-objective optimization. J. Shaanxi Norm. Univ. Nat. Sci. Ed. 2023, 51, 25–34. [Google Scholar]
  38. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  39. Coello, C.; Lechuga, M. MOPSO: A proposal for multiple objective particle swarm optimization. In Proceedings of the 2002 Congress on Evolutionary Computation. CEC’02 (Cat. No. 02TH8600), Honolulu, HI, USA, 12–17 May 2002; pp. 1051–1056. [Google Scholar]
  40. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef]
  41. Yue, C.; Qu, B.; Liang, J. A multiobjective particle swarm optimizer using ring topology for solving multimodal multiobjective problems. IEEE Trans. Evol. Comput. 2017, 22, 805–817. [Google Scholar] [CrossRef]
  42. Guerreiro, A.; Fonseca, C.; Paquete, L. The hypervolume indicator: Computational problems and algorithms. ACM Comput. Surv. CSUR 2021, 54, 1–42. [Google Scholar] [CrossRef]
  43. Mohammadi, A.; Omidvar, M.; Li, X. A new performance metric for user-preference based multi-objective evolutionary algorithms. In Proceedings of the 2013 IEEE Congress on Evolutionary Computation, Cancun, Mexico, 20–23 June 2013; pp. 2825–2832. [Google Scholar]
  44. Derrac, J.; García, S.; Molina, D.; Herrera, F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evol. Comput. 2011, 1, 3–18. [Google Scholar] [CrossRef]
Figure 1. The encoding and decoding method (red color indicates the selected feature).
Figure 1. The encoding and decoding method (red color indicates the selected feature).
Biomimetics 09 00613 g001
Figure 2. PF approximations of three compared algorithms.
Figure 2. PF approximations of three compared algorithms.
Biomimetics 09 00613 g002
Figure 3. The number of feature subsets obtained by three compared algorithms under different cases.
Figure 3. The number of feature subsets obtained by three compared algorithms under different cases.
Biomimetics 09 00613 g003
Figure 4. The number of feature subsets obtained by the MMOFS using different performance indicators.
Figure 4. The number of feature subsets obtained by the MMOFS using different performance indicators.
Biomimetics 09 00613 g004
Table 1. 12 Selected features.
Table 1. 12 Selected features.
DefinitionType
x1ETC card mismatchDiscrete
x2U-type special conditionDiscrete
x3Vehicle type mismatchDiscrete
x4OverlimitDiscrete
x5Cash transaction countDiscrete
x6ETC malfunctionDiscrete
x7Green priority exceptionDiscrete
x8Manual gate barrierDiscrete
x9No ETCDiscrete
x10Weighing faultDiscrete
x11Axle load modificationDiscrete
x12Peak daily volumeDiscrete
Table 2. Mean and standard deviation values of three feature selection methods in terms of PSP, HV, and IGD.
Table 2. Mean and standard deviation values of three feature selection methods in terms of PSP, HV, and IGD.
NSGA-II-RF
Mean(std)
MOPSO-RF
Mean(std)
MMOFS
Mean(std)
PSP8.49 × 101 (3.94 × 101)+7.04 × 101 (9.76 × 100)+1.89 × 102 (3.99 × 101)
HV1.15 × 102 (3.67 × 100)+1.19 × 102 (4.37 × 100)+1.29 × 102 (2.04 × 100)
IGD1.31 × 100 (2.96 × 10−1)+7.64 × 10−1 (3.06 × 10−1)+7.37 × 10−1 (2.54 × 10−1)
+ 3 3
0 0
0 0
Table 3. Feature subsets obtained by the proposed algorithm.
Table 3. Feature subsets obtained by the proposed algorithm.
The Number of Selected FeaturesER ValueFeature Subset
10.12x7
20.08x7, x11
x7, x9
x9, x12
30.04x2, x7, x11
x7, x9, x11
40.03x4, x6, x7, x12
x1, x7, x9, x11
x1, x7, x10, x11
x1, x3, x7, x11
x4, x5, x7, x11
50.02x1, x4, x7, x9, x11
x4, x5, x7, x11, x12
x2, x4, x7, x8, x11
60.016x1, x5, x7, x10, x11, x12
x1, x3, x4, x7, x9, x11
x2, x3, x4, x7, x8, x11
x1, x4, x7, x9, x10, x11
x1, x2, x3, x7, x9, x11
90.008x1, x3, x4, x5, x6, x7, x10, x11, x12
x1, x2, x4, x5, x7, x8, x9, x10, x11
x1, x4, x5, x6, x7, x8, x10, x11, x12
100x2, x3, x4, x5, x6, x7, x9, x10, x11, x12
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, Z.; Mo, H.; Yan, Z.; Fan, Q. A Multimodal Multi-Objective Feature Selection Method for Intelligent Rating Models of Unmanned Highway Toll Stations. Biomimetics 2024, 9, 613. https://doi.org/10.3390/biomimetics9100613

AMA Style

Gao Z, Mo H, Yan Z, Fan Q. A Multimodal Multi-Objective Feature Selection Method for Intelligent Rating Models of Unmanned Highway Toll Stations. Biomimetics. 2024; 9(10):613. https://doi.org/10.3390/biomimetics9100613

Chicago/Turabian Style

Gao, Zhaohui, Huan Mo, Zicheng Yan, and Qinqin Fan. 2024. "A Multimodal Multi-Objective Feature Selection Method for Intelligent Rating Models of Unmanned Highway Toll Stations" Biomimetics 9, no. 10: 613. https://doi.org/10.3390/biomimetics9100613

Article Metrics

Back to TopTop