1. Introduction
In the past decade, the domain of computational intelligence has undergone significant advancements, spurred by factors such as enhanced computational capabilities, data proliferation, and algorithmic innovation. Concurrently, feature projection, a vital component of computational intelligence, has experienced substantial progress, with researchers and practitioners striving to optimize model performance and interpretability. Data projection, as a crucial discipline, concentrates on the collection, processing, storage, and analysis of data. The advent of the big data era has facilitated rapid progress and garnered considerable attention for data engineering in recent years. The swift growth of technologies such as the Internet and social media has resulted in an exponential increase in the volume of data generated worldwide. These data encompass structured, unstructured, and semi-structured types, presenting a diverse array of application scenarios for data engineering while simultaneously posing significant challenges. The main concept of feature projection is to establish a concise and more understandable data architecture, enhance its performance, and minimize computing overhead [
1,
2,
3].
The Group Method of Data Handling (GMDH) is a class of inductive algorithms devised by Alexey G. Ivakhnenko in the late 1960s for constructing mathematical models from empirical data. Employed across various domains, such as pattern recognition, time series prediction, system identification, and optimization problems, GMDH is a self-organizing technique that generates polynomial network models through iterative processing layers [
4]. GMDH offers numerous advantages, including automatic model selection, which obviates the need for manual trial and error and streamlines the modeling process. The robustness of this method allows it to handle noisy data and outliers effectively, while its inherent regularization and model selection mechanisms prevent overfitting and bolster the generalization capabilities of the resulting models. Moreover, the ability of GMDH to manage datasets with a large number of input variables renders it suitable for high-dimensional problems, and its hierarchical structure enables efficient parallel processing on contemporary hardware. Finally, the adaptability allows for integration with other techniques, such as fuzzy logic, neural networks, or evolutionary algorithms, to tackle specific modeling challenges and enhance performance.
Fuzzy neural networks (FNNs) synergistically integrate the adaptability of neural networks and the interpretability of fuzzy logic, thereby providing a potent instrument for addressing intricate problems.
Fuzzy set theory is a mathematical framework designed to address uncertainty, ambiguity, and imprecision in data. In contrast to classical set theory, which employs crisp boundaries to distinguish elements that belong to a set from those that do not, fuzzy set theory allows for partial membership [
5]. Central to fuzzy set theory is the concept of fuzzy membership functions, which represent the degree of membership of an element in a fuzzy set. Various types of membership functions, such as triangular, trapezoidal, Gaussian, and sigmoid, can be utilized to model different levels of uncertainty and imprecision in data. Throughout the years, fuzzy set theory has been applied across a broad range of disciplines, including control systems, decision making, pattern recognition, artificial intelligence, and data mining. The capacity of fuzzy set theory to handle vagueness and imprecision in data renders it particularly suitable for real-world applications, where accurate information is often unavailable or challenging to acquire.
In recent years, FNNs have experienced significant advancements in terms of their development and application. FNNs combine the learning capabilities of neural networks with the interpretability and adaptability of fuzzy logic, making them suitable for handling uncertain or imprecise data. This hybrid approach has led to improvements in various fields, such as control systems, pattern recognition, data mining, and forecasting. A salient feature of FNNs is their capacity to manage imprecise and uncertain data, rendering them highly applicable for real-world scenarios [
5,
6]. Cutting-edge progress in FNNs encompasses the formulation of hybrid FNNs, which amalgamate multiple FNN models to augment their performance. In a noteworthy study conducted in 2013, Chen et al. design an adaptive tracking control for a class of nonlinear stochastic systems with unknown functions [
7]. Subsequently, in 2022, Zhang et al. introduced a nonstationary fuzzy inference systems model for the variation in opinions of individual experts and expert groups [
8]. These advancements exemplify the ongoing innovation and potential applications of FNNs in diverse fields.
In this study, motivated by the contemporary challenges and opportunities in the field of computational intelligence, we reconsider and restructure data processing and fuzzy models from various angles.
The key aspects of the proposed FANN model can be concisely summarized as follows:
Enhanced performance through synergy of proposed methods: This study is the first to combine feature projection and clustering algorithms in the design of fuzzy aggregation models, addressing the limitations of existing FNNs. The FANN employs the FSU to generate essential inputs through clustering and correlation operations, while the SSM provides additional neurons with significant features.
Flexibility in generating significant subsets between input and output spaces: Conventional FNNs often have limited interpretability due to their inability to incorporate important features from the original data smoothly. This can negatively impact model performance. In contrast, the proposed FANN retains many original features from raw data, preserving their physical significance and enhancing the interpretability of the proposed model.
Alleviation of overfitting and optimization of the convergence process: The FANN, combined with the SSM and the ℓ2-norm regularization learning method, mitigates overfitting by selecting subsets derived from low-complexity raw data and introducing penalty terms through regularization.
Methodologically, to mitigate overfitting and enhance modeling accuracy, we modify the conventional FNN structure and integrate it with a feature self-enhancement unit (FSU) and statistical selection methods (SSMs). The FSU employs statistical methods and a gaussian mixture model (GMM) to reduce input dimensionality and noise influence. By analyzing the probability distribution of the feature space, the FSU autonomously adjusts and optimizes feature weights. From an experimental analysis perspective, we offer structural analyses and performance comparisons of typical machine learning models, recent FNN-based models, and the proposed FANN.
Figure 1 shows the overall idea of the FANN development process. The idea behind developing a fuzzy adaptive neural network (FANN) stems from the desire to address the limitations of conventional fuzzy neural networks (FNNs) and improve their performance in handling complex, nonlinear, and uncertain data. By introducing the FSU, SSMs, and ℓ2-norm regularization, the FANN combines the strengths of fuzzy logic and artificial neural networks, offering advanced capabilities for processing imprecise and uncertain information. By using the statistical selection method, the proposed model can generate extra neurons of variables with significant statistical features and evaluate node performance. The SSM is a vital and widely used feature processing method, applicable in various situations. Furthermore, the SSM can simultaneously assess multiple model terms, enabling the comparison of different linear model fits. This contrasts with most other methods, which can only evaluate one feature at a time. Concurrently, the proposed model, based on the FSU, introduces a strategy for processing input variables through the correlation relationship between input and output spaces. As a novel node selection method, the proposed methods also enhance the robustness and generalization of the proposed model.
The main contributions of the proposed FANN can be summarized as follows:
- (1)
A design methodology incorporating the FSU to enhance the robustness and performance of the model. By preserving important features from raw data, the FANN maintains the physical significance of the original features, resulting in improved interpretability and readability.
- (2)
Effective mitigation of overfitting by integrating the SSM and ℓ2-norm regularization learning techniques. The selected input subsets, derived from low-complexity raw data, and the regularization approach with penalty terms contribute to a more robust model.
- (3)
Synergistic design: The FANN combines feature projection, clustering algorithms, and regularization techniques to address the limitations.
The paper is structured as follows: In
Section 2, we present the details of the proposed FSU and the SSM. In addition, the theoretical explanation of the original model is supplemented. Comprehensive experimental results and experimental analysis are presented in
Section 3 while conclusions, a few limitations, and future work that require more attention are drawn in
Section 4.