Next Article in Journal
Mobile Mental Health Screening in EmotiZen via the Novel Brain-Inspired MCoG-LDPSNet
Next Article in Special Issue
Red-Crowned Crane Optimization: A Novel Biomimetic Metaheuristic Algorithm for Engineering Applications
Previous Article in Journal
High-Value Patents Recognition with Random Forest and Enhanced Fire Hawk Optimization Algorithm
Previous Article in Special Issue
Evolution and Trends of the Exploration–Exploitation Balance in Bio-Inspired Optimization Algorithms: A Bibliometric Analysis of Metaheuristics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Enhanced MIBKA-CNN-BiLSTM Model for Fake Information Detection

1
International Business School, Jilin International Studies University, Changchun 130117, China
2
School of Management Science and Information Engineering, Jilin University of Finance and Economics, Changchun 130117, China
3
Office of Scientific Research, Jilin University of Finance and Economics, Changchun 130117, China
4
Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
*
Author to whom correspondence should be addressed.
Biomimetics 2025, 10(9), 562; https://doi.org/10.3390/biomimetics10090562
Submission received: 4 July 2025 / Revised: 17 August 2025 / Accepted: 21 August 2025 / Published: 23 August 2025
(This article belongs to the Special Issue Nature-Inspired Metaheuristic Optimization Algorithms 2025)

Abstract

The complexity of fake information and the inefficiency of parameter optimization in detection models present dual challenges for current detection technologies. Therefore, this paper proposes a hybrid detection model named MIBKA-CNN-BiLSTM, which significantly improves detection accuracy and efficiency through a triple-strategy enhancement of the Black Kite Optimization Algorithm (MIBKA) and an optimized dual-channel deep learning architecture. First, three improvements are introduced in the MIBKA. The population initialization process is restructured using circle chaotic mapping to enhance parameter space coverage. The conventional random perturbation is replaced by a random-to-elite differential mutation strategy (DE/rand-to-best/1) to balance global exploration and local exploitation. Moreover, a logarithmic spiral opposition-based learning (LSOBL) mechanism is integrated to dynamically explore the opposition solution space. Second, a CNN-BiLSTM dual-channel feature extraction network is constructed, with hyperparameters such as the number of convolutional kernels and LSTM units optimized by MIBKA to enable adaptive model structure alignment with task requirements. Finally, a high-quality fake information dataset is created based on social media platforms, including CCTV. The experimental results show that our model achieves the highest accuracy on the self-built dataset, which is 3.11% higher than the optimal hybrid model. Additionally, on the Weibo21 dataset, our model’s accuracy and F1-score increased by 1.52% and 1.71%, respectively, compared to the average values of all baseline models. These findings offer a practical and effective approach for detecting lightweight and robust false information.

1. Introduction

With the Internet’s widespread adoption and rapid development, social media platforms have become convenient spaces for users to access information, share opinions, and communicate daily [1,2]. Users have transformed from passive information consumers to active content creators and distributors, accelerating the rapid dissemination of fake information across social networks [3,4]. Fake information could be defined as intentionally or unintentionally disseminated content that violates authenticity, including false statements, manipulated facts, and misleading narratives, with the potential to cause societal harm [5,6]. In 2025, fake news continues to pollute the Internet at an alarming scale, with 62% of online content now deemed false. A staggering 86% of global citizens have been exposed to misinformation, while 40% of content shared on social media is fake [7]. False information spread by social media has amplified its disruptive effects, disturbed daily life, and posed serious challenges to public safety and governance. Therefore, exploring rapid and efficient methods for fake information detection has become a persistent research focus in academia.
Traditional fake information detection methods rely on manually defined linguistic features, such as sentiment polarity, lexical complexity, and propagation rule models [8,9]. However, these approaches struggle with significant limitations in generalization when confronted with adversarially generated fake texts. The advancement of pre-trained language models has introduced new characteristics in fake information, including logical closed loops and contextual adaptation [10]. These developments present dual challenges for detection technologies. Detection methods must effectively capture semantic contradictions within texts while modeling the evolving patterns during information dissemination [11,12]. Deep learning-based detection methods have emerged as a crucial technological approach to tackle these difficulties due to their powerful feature abstraction capabilities [13,14].
In the domain of fake information detection, the integration of Convolutional Neural Networks (CNN) and bidirectional long short-term memory networks (BiLSTM) has demonstrated unique value [15]. CNN accurately captures local semantic patterns through hierarchical convolution operations [16]. These operations effectively identify unconventional modifier stacking and abnormal anaphoric relations, which are micro-linguistic features essential for precise detection. BiLSTM models bidirectional temporal dependencies and effectively tracks the logical coherence of textual context [17]. As a result, they are particularly well-suited for analyzing the progressive semantic distortion that occurs during the propagation of fake information. Compared to commonly used Transformer architectures, this combination offers irreplaceable advantages in specific scenarios. In data-scarce contexts, such as detecting fake information during emerging events, Lightweight recurrent neural networks such as BiLSTM can compress the number of parameters to less than 1/50 of the Transformer model while maintaining a performance level of over 90% in text classification tasks [18]. Additionally, the inference speed of CNNs in short text classification tasks is 3 to 5 times faster than that of Transformer-based models, and the accuracy is comparable in scenarios with high-frequency vocabulary [19]. Nevertheless, the performance of such models heavily depends on hyperparameter configuration [20]. Traditional grid search methods may require thousands of GPU hours when optimizing complex parameters, resulting in substantial human and computational costs. Gradient-based optimization strategies struggle with discrete parameter combinations, leading to high tuning costs and poor cross-domain robustness in practical deployments [21].
Researchers are exploring swarm intelligence algorithms to improve efficiency in optimizing deep learning model parameters. From classical particle swarm optimization (PSO) and genetic algorithms (GA) to emerging Black Kite Algorithm (BKA), these methods simulate collective behaviors in biological populations and demonstrate unique advantages in tackling high-dimensional, non-convex optimization problems [22,23,24]. Taking BKA as an example, it simulates the circling foraging behavior of black kites by dividing population individuals into explorers and followers and dynamically adjusting their movement strategies during the search. Compared with traditional algorithms, BKA exhibits faster convergence in continuous parameter optimization tasks. However, finding optimal solutions can still be challenging. This difficulty largely stems from imbalances between global exploration and local search capabilities, as well as BKA’s limited global search capability and vulnerability to local optima. Therefore, it is necessary to enhance BKA.
Here, we can obtain the following two motivations for this study:
(1) The dual-channel integrates local and global semantic understanding, enabling comprehensive characterization of fake information’s diverse linguistic features.
(2) The improved BKA model can optimize the hyperparameters of the hybrid deep learning model, enabling it to achieve higher accuracy.
Thus, this study proposes an improved multi-strategy Black Kite Algorithm (MIBKA) and constructs the MIBKA-CNN-BiLSTM hybrid model. The innovations are reflected in the following aspects.
(1) The paper proposes three strategies to enhance the BKA. The circle chaotic mapping is used for population initialization to make the initial population distribution more uniform. A differential elite mutation strategy is integrated during the kite’s attack phase for position updates to balance global exploration and local exploitation. An opposition-based learning mechanism is introduced for an individual foraging to dynamically explore the op-position solution space.
(2) A dual-channel feature extraction network combining CNN and BiLSTM is proposed. The CNN branch employs multi-scale convolution kernels to capture local textual anomaly patterns, while the BiLSTM branch models contextual logical relations through hierarchical state propagation. Then, the improved MIBKA is used to optimize hyperparameters, including the number of 1D convolution filters, convolution kernel sizes, and the number of BiLSTM units.
(3) A fake information dataset is constructed and validated. Extensive comparative experiments with various single and hybrid deep learning models demonstrate the superiority of the MIBKA-CNN-BiLSTM model in fake information detection tasks.

2. Related Work

2.1. Traditional Techniques-Driven Fake Information Detection

Traditional fake information detection techniques focus on two main approaches, manual rule construction and shallow feature analysis, corresponding to two foundational research frameworks driven by expert knowledge and data statistics [25].
Rule-based methods depend on domain experts to systematically extract linguistic and logical features, thereby constructing detection systems [26]. The core lies in developing a rule base encompassing multidimensional linguistic anomaly patterns. Specifically, the rule-based methods can be categorized into three distinct types. The first aspect focuses on the syntactic level, identifying contradictory rhetorical patterns where absolutist statements coexist with unsupported claims [27,28]. Thus, rules are established that capture common linguistic exaggerations and subjective judgments to detect emotionally charged fake information. The second type addresses the logical level by identifying gaps in causal relations caused by insufficient links between premises and conclusions [29]. It extracts structured patterns that indicate broken reasoning chains, thereby improving the detection of argumentative fallacies in misleading content. The third point examines inconsistencies in time and events [30]. It identifies conflicts in the events’ timing and sequences by finding contradictory adverbial phrases. It also sets up rules to analyze event consistency to spot misinformation based on fake timelines. Verified through manually annotated samples, these rule sets demonstrate precise localization capabilities and strong interpretability in identifying typical linguistic anomalies. Although rule-based methods provide explainability advantages, their passive knowledge acquisition remains a fundamental limitation. As creators of fake information adopt more flexible strategies, experts must frequently review new fake texts and enhance their guidelines, increasing maintenance costs as the types of information grow.
Research has gradually shifted toward shallow feature analysis to overcome the shortcomings of rule-based methods in generalization and scalability. This approach utilizes machine learning techniques to convert text into quantifiable feature vectors, enabling automated detection [31,32]. Some researchers often prepare texts by breaking them down into smaller parts and analyzing the role of each word. They look at features such as word frequency, sentence complexity, and the overall tone of the text. After this analysis, they create models to detect patterns using tools like Support Vector Machines and Random Forests [33]. Other researchers have identified distinct patterns over time in the detection features that differentiate genuine information from fake information. They suggest using dynamic temporal structures to capture the evolution of these time-sensitive features. They employ machine learning algorithms such as decision trees, random forests, and support vector machines (SVM) to classify this information. The primary advantages of shallow feature analysis lie in relatively controllable modeling processes, low computational complexity, and high training efficiency with strong interpretability, particularly on medium-scale datasets [34,35]. Furthermore, the sensitivity of shallow models to feature weights provides clear discriminative criteria, which is beneficial in situations that require explicit explanations of detection logic. However, when false information is hidden through semantic operations like local data tampering and restructuring narrative logic, traditional feature engineering often fails to identify the deep logical contradictions within the text effectively. This limitation stems from over-reliance on handcrafted features and the inability to model nonlinear semantic relationships between linguistic units. Therefore, deep learning techniques offer promising new pathways to break traditional detection performance bottlenecks.

2.2. Deep Learning-Driven Fake Information Detection

Advances in deep learning have transformed fake information detection from basic, feature-based classification to deep semantic understanding through end-to-end neural architectures. This evolution has led to three major technical branches: single models, hybrid architectures, and pre-training enhanced models [36,37,38].
Initial deep learning explorations in fake information detection focus on constructing structurally clear and task-specific single models, including Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN), Long Short-Term Memory networks (LSTM), Gated Recurrent Units (GRU), and early Transformer variants [39,40,41,42]. CNNs excel at extracting local syntactic and phrase-level features, while RNNs capture temporal dependencies to identify semantic continuity anomalies in context. LSTMs mitigate the vanishing gradient problem in long-distance dependencies via forget and memory gates, showing a strong ability to maintain semantic coherence. GRUs simplify gating mechanisms further, achieving faster convergence and comparable performance in small-data scenarios. Transformers possess specific global attention capabilities suitable for loosely structured social texts. Due to their structural simplicity and controllable parameterization, single-model architectures offer advantages such as stable training, fast convergence, and ease of debugging. These characteristics make them well-suited for early-stage scenarios with limited sample size or when interpretability is a primary concern. In addition, such models still demonstrate a strong ability to detect specific types of fake content, particularly in patterned rumors or highly repetitive texts. Nevertheless, single models typically concentrate on a limited range of semantic features due to their simplistic structure. They cannot capture the diverse manifestations of misinformation that may arise locally and globally across multiple granularities and contexts. Consequently, they struggle to meet the demands of comprehensive, multi-perspective modeling of high-dimensional deceptive patterns.
To overcome the dimensional limitations of single models, researchers propose various hybrid neural network structures that integrate different model advantages, enhancing the modeling capacity for complex semantics of fake information. Among these hybrid models, local-global fusion architectures are most prevalent. CNN-RNN or CNN-BiLSTM synergistically models local phrase features and global temporal dependencies [43,44]. CNN layers efficiently identify lexical-level anomalies, while BiLSTM captures inter-sentence logical relations and long-term semantic evolution, improving contextual sensitivity and robustness against complex fake information. Additionally, attention mechanisms have also been incorporated into hybrid models. For instance, the CNN-BiGRU-Attention model focuses on key semantic regions post-local feature extraction, improving recognition of covert manipulative language [45]. Some studies have explored the combination of Graph Neural Networks (GNN) with sequential models by constructing user-content interaction graphs to extract propagation structure features and model the semantic evolution in dissemination [46,47]. This approach enhances the depiction of fake information spread patterns and content dynamics. The advantages and limitations of the above models are shown in Table 1. In summary, hybrid architectures integrate multi-level semantic information deeply while ensuring computational efficiency. They significantly enhance the detection of logical inconsistencies, emotional polarization, and false citations, establishing themselves as valuable methods in fake information detection.
With the rise in pre-trained language models like Bidirectional Encoder Representation from Transformers (BERT), researchers have used BERT to analyze contextual semantics and develop deep semantic vector representations. BERT successfully captures long-range dependencies and implicit semantic connections by implementing self-attention mechanisms [48,49]. Improving the model enhances its comprehension of intricate text structures and demonstrates robustness in detecting disguised and semantically ambiguous misinformation. Some studies integrate BERT with BiLSTM and similar architectures to enhance modeling effectiveness [50]. These improvements result in hybrid networks that combine contextual modeling with temporal semantic mechanisms, thereby increasing detection accuracy and generalization capabilities. However, it is noteworthy that such large-scale language models still face significant challenges in practical deployment. Their massive parameter counts result in high computational costs, requiring powerful GPU clusters or cloud resources [51,52]. Additionally, adjustments in hyperparameters like the number of convolutional kernels, the depth of recurrent layers, and the number of attention heads can lead to significant performance variations. Therefore, creating efficient hyperparameter optimization methods for consistent performance improvements continues to be a key area of research.

3. Methodology

The paper proposes the MIBKA-CNN-BiLSTM hybrid model for fake information detection. First, the original BKA is enhanced using three strategies. Then, the MIBKA is employed to optimize the hyperparameters of the CNN-BiLSTM model. Finally, the trained MIBKA-CNN-BiLSTM model is used to classify the crawled information as real or fake. The overall architecture of the proposed model is illustrated in Figure 1.

3.1. Black Kite Optimization Algorithm (BKA)

The black kite is a medium-sized bird of prey known for its exceptional agility in hovering, strategic hunting maneuvers, and adaptive migratory patterns. These traits enable it to explore and exploit resources in diverse environments efficiently. The Black Kite Optimization Algorithm (BKA) is a nature-inspired optimization method derived from these distinctive behaviors. The BKA consists of three main stages: population initialization, attack behavior, and migratory behavior.

3.1.1. Population Initialization

In BKA, the first step is to create a set of random solutions to initialize the population. The following matrix can be used to represent the position of each Black Kite (BK).
B K = B K 1,1 B K 1,2 B K 1 , d i m B K 2,1 B K 2,2 B K 2 , d i m B K p o p , 1 B K p o p , 2 B K p o p , d i m
Here, p o p denotes the population size, and d i m refers to the dimensionality of the search space. B K i , j represents the position of the i black kite in the j dimension. The positions are initialized according to Equation (2).
B K i , j = B K l b + rand B K u b B K l b
i is an integer between 1 and p o p . B K u b and B K l b represent the lower and upper bounds of the j dimension, and r a n d is a randomly generated number within the range [0, 1]. During the initialization phase, the individual with the best fitness is selected as the leader of the initial population, as shown in Equation (3).
f b e s t = min f ( X i )

3.1.2. Attack Behavior

Black kites, while flying, adjust the angles of their wings and tails based on the wind speed. This flexibility allows them to hover silently and observe their prey before diving quickly to attack. The attack behavior is modeled using global exploration and local exploitation. The position update rule is formulated in Equation (4).
X t + 1 i , j = X t i , j + n 1 + s i n ( r ) X t i , j       p < r X t i , j + n 2 r 1 X t i , j                           e l s e  
The first case simulates scenarios where the black kite hovers in the air and then dives rapidly toward its prey. The second case simulates hovering and gradual movement. X t + 1 i , j and X t i , j denote the j dimension position of the i black kite at iterations t and t + 1 . r [0, 1] is a random number, and p is a constant set to 0.9. The parameter n is a nonlinear control variable, defined in Equation (5), where T is the maximum number of iterations, and t is the current iteration number. This parameter is designed to adjust the search scope dynamically.
n = 0.05 × exp 2 × t / T 2

3.1.3. Migratory Behavior

In nature, black kites migrate to new habitats for better survival conditions and resources. When the current leader fails to find optimal environments, the group dynamically replaces it with more capable individuals, ensuring continuous movement toward better habitats. This process is similar to the dissemination process of false information, where key communicators guide the spread of the information. Therefore, the migration behavior of the black kites can inspire the identification of the adaptive transmission patterns of false information. Migration is typically led by a dominant bird whose navigation ability is critical for group success. BKA adopts a migratory strategy based on the assumption that if the fitness of the current individual is worse than that of a randomly selected individual, the leader should relinquish control and join the migratory group. Otherwise, the leader continues to guide the population toward the optimal solution. The migration-based position update rule is shown in Equation (6).
X t + 1 i , j = X t i , j + C 0,1 × X t i , j L t j                             F i < F r i X t i , j + C 0,1 × L t j m × X t i , j           e l s e              
m = 2 × sin r + π / 2
Here, L t j denotes the j dimension of the leader’s position at iteration t (i.e., the current global best solution). F i is the fitness value of the current individual, and F r i is the fitness value of a randomly selected individual. C 0,1 refers to a value sampled from the Cauchy distribution. The probability density function of the Cauchy distribution is defined in Equation (8). When δ = 1 and μ = 0, the distribution simplifies to its standard form, as shown in Equation (9).
f x , δ , μ = 1 π δ δ 2 + x μ 2 ,   < x <
f x , δ , μ = 1 π 1 x 2 + 1 ,   < x <

3.2. Multi-Strategy Improved Black Kite Optimization Algorithm (MIBKA)

Although BKA demonstrates fast convergence speed and high optimization precision, global exploration and local exploitation are imbalanced. Specifically, its global search capability is relatively weak, making it prone to falling into local optima. Three improvement strategies are adopted in this study to address this issue.

3.2.1. Introducing Circle Chaotic Mapping for Population Initialization

In swarm intelligence algorithms, the quality of the initial solution significantly impacts performance and convergence. The standard BKA utilizes a uniform random initialization strategy that may lead to uneven distribution and boundary aggregation in mixed-parameter spaces, thereby diminishing the effectiveness of global search. Alternative methods such as Latin hypercube sampling, low-discrepancy sequences like Halton, and chaotic mappings are commonly used to mitigate this issue. Chaotic mappings are preferred due to their ergodicity, randomness, and adaptability to complex parameter spaces.
This study evaluates 21 chaotic mappings, categorized into three types based on their dynamic characteristics, as summarized in Table 2.
To determine the most suitable chaotic map for BKA initialization, each of the 21 strategies is tested by integrating it into BKA and evaluating the performance on the CEC2017 benchmark suite. The experiments are conducted with a population size of 50, a dimensionality of 30, and a maximum of 300 iterations. Each algorithm variant is run five times, and the average results are used for ranking. Evaluation metrics include optimal fitness, convergence iterations, and global optimality hit rate. Selected results are shown in Figure 2.
The experimental results show that the circle chaotic map outperforms other mappings on most benchmark functions. For instance, in the F1 test function, BKA with circle mapping achieves an average fitness of 8.3388 × 102, reducing the original BKA result of 1.1078 × 103 by 24.77%. The number of convergence iterations has decreased from 213 to 127, reflecting a 40.38% improvement in performance. Moreover, circle mapping ranks first in 21 out of 30 runs and exhibits premature convergence in only two cases. Based on these results, circle chaotic mapping is selected to improve the initialization phase of BKA.
The circle map generates pseudo-random sequences using a phase-shifting mechanism. Its iterative equation is defined in Equation (10). Here, θ = 0.5 is the rotation angle controlling the global traversal direction, and δ = 0.2 is the perturbation strength regulating local randomness.
x k + 1 = x k + θ δ 2 π sin 2 π x k mod 1
The improved initialization process is described as follows.
  • Chaotic sequence generation. For each individual, a chaotic sequence x 1 , x 2 , , x D is generated from a random seed x 0   ϵ [0, 1). The sequence is preheated for 100 iterations to eliminate transient effects. Here, D represents the number of parameters to be optimized.
  • Handling discrete parameters. For integer parameters such as the number of network layers L   ϵ 1 , 2 , , 10 or the number of convolution kernels K   ϵ 16 , 32 , 64 , floor mapping is applied to convert chaotic values. Specifically, given a normalized chaotic value x c h a o s , the transformation is defined in Equation (11). L m a x and L m i n are the upper and lower bounds of the parameters. For example, if x c h a o s = 0.73 , the convolution kernels K = 0.73 64 16 + 16 = 51 is rounded and adjusted to 64.
    L = x c h a o s L m a x L m i n + L m i n
  • Preserving continuous parameters. For continuous variables such as learning rate α   ϵ [0.0001, 0.1] and dropout rate β   ϵ [0.2, 0.6], the original distribution of the chaotic sequence is retained through linear scaling, as shown in Equation (12). For instance, when x c h a o s = 0.15, then α = 0.0151.
    α = x c h a o s α m a x α m i n + α m i n
  • Dynamic boundary reflection. If x k + 1 exceeds the defined domain boundaries, a reflection mechanism is used to keep it within the feasible range, as defined in Equation (13). This strategy ensures valid solutions and maintains population diversity in the adequate search space.
    x = 2 U B x k + 1 , i f x k + 1 > U B 2 L B x k + 1 , i f x k + 1 < L B
The BKA achieves better coverage and faster convergence in mixed parameter spaces by reconstructing the initialization phase using the circle chaotic mapping.

3.2.2. Using Random-to-Elite Differential Mutation for Attack Phase

The core search mechanism of BKA differentiates roles between explorers and followers during the attack phase. Explorers expand the search space through stochastic movements, while followers focus on local exploitation around the current best solution. However, in complex high-dimensional mixed-parameter spaces, the original random perturbation strategy lacks guidance, challenging balancing global exploration and local exploitation. Some researchers have attempted to tune BKA’s internal parameters to address this limitation or integrate it with other metaheuristic algorithms. Although such methods can improve performance, they often face difficulties in parameter coordination. Compared with other approaches, mutation strategies introduce random or probabilistic perturbations to break population homogeneity. These strategies can be designed according to the specific problem and the characteristics of the algorithm. By incorporating historical information or elite individuals, they can guide the search direction more effectively. This approach helps achieve a better balance between global search and local exploitation but only requires a linear complexity of time.
This study investigates 13 mutation strategies, including Gaussian mutation (Guass1), elite Gaussian mutation (Guass2), Cauchy mutation (Cauchy1), inverse Cauchy mutation (Cauchy2), t-distribution mutation (t), adaptive t-distribution mutation (Self-t), normal cloud mutation (Cloud), periodic mutation (Periodic), elite differential mutation (DE/best/1), random-to-elite differential mutation (DE/rand-to-best/1), random differential mutation (DE/rand/2), best/2 mutation (DE/best/2), and non-uniform mutation (H). These strategies are categorized into four types based on their mathematical mechanisms, as shown in Table 3.
To identify the most effective strategy for improving BKA’s attack phase, each of the 13 mutation strategies is embedded into BKA and tested on the CEC2017 benchmark functions. Experiments are conducted with a population size of 50, a dimensionality of 30, and a maximum of 300 iterations. Every improved algorithm is executed five times, and average performance is recorded. Evaluation metrics included best fitness and number of convergence iterations. Selected results are illustrated in Figure 3.
Experimental results show that the random-to-elite differential mutation (DE/rand-to-best/1) performs best on 19 test functions. For instance, in the F10 benchmark test, the average fitness of this strategy reached 1.9581 × 109, which was 80.42% lower than the original BKA’s 1.0002 × 1010. Differential evolution-based strategies balance exploration and exploitation by combining random exploration with elite guidance. Due to its random perturbation and elite direction characteristic, DE/rand-to-best/1 is integrated into BKA’s attack phase to compensate for the lack of guided search in the original design.
The random-to-elite differential mutation originates from the Differential Evolution (DE) algorithm, a population-based global optimization method. DE generates mutation vectors by computing individuals’ differences and combining them with crossover and selection to search for the global optimum. The DE/rand-to-best/1 strategy builds on this by introducing randomness to enhance diversity while leveraging elite individuals to accelerate convergence toward the global optimum.
In the BKA framework, this mutation strategy is embedded into the attack phase to replace the original stochastic disturbance formula. The updated position equation is defined in Equation (14).
X t + 1 i , j = X t i , j + n 1 + s i n ( r ) X i , j t + F X t r 1 , j X t r 2 , j + λ ( X t e l i t e , j X t i , j )       p < r X t i , j + n 2 r 1 X t i , j + F X t r 1 , j X t r 2 , j + λ ( X t e l i t e , j X t i , j )                       e l s e  
Here, X t r 1 , j and X t r 2 , j represent two randomly selected individuals from the population. X t e l i t e , j refers to the elite individual (with the best fitness) in the j dimension at iteration t . F is the differential scaling factor controlling the impact of the difference vector X t r 1 , j X t r 2 , j . λ is the elite learning factor, which determines how much influence the elite solution exerts on the updated position.
By incorporating the DE/rand-to-best/1 strategy into BKA’s attack phase, the algorithm significantly reduces its susceptibility to local convergence. The enhanced equation combines random perturbation with elite-driven guidance, improving search efficiency and robustness across complex optimization tasks.

3.2.3. Utilizing Logarithmic Spiral Opposition Learning for Position Update Phase

After completing the attack and migration phases, BKA only retains the superior individuals through fitness comparisons and lacks active exploration of the potential reverse solution space, resulting in a limited search range. In high-dimensional mixed parameter spaces, especially when discrete network layers coexist with continuous learning rates, forward search is prone to miss high-quality solutions in the reverse regions. Opposition-based learning (OBL) has been widely adopted among various improvement techniques due to its simplicity and effectiveness. The core idea of OBL is to generate opposite solutions relative to the current candidates and explore whether better solutions may lie in the opposite direction. This approach expands the search space, helps the algorithm avoid premature convergence, and enhances global optimization capability without significantly modifying the original algorithm structure. This study evaluates 12 opposition-based learning strategies and classifies them into five categories based on their core mechanisms, as shown in Table 4.
To identify the most suitable opposition learning strategy for enhancing the BKA population update phase, 12 strategies are individually integrated into BKA and tested on the CEC2017 benchmark set. Experiments are configured with a population size of 50, 30-dimensional search space, and 300 maximum iterations. Each modified algorithm is executed five times, and the average performance is recorded. Evaluation metrics include best fitness and number of convergence iterations. Partial results are illustrated in Figure 4.
Experimental results indicate that logarithmic spiral opposition-based learning (LSOBL) performs best on 24 benchmark functions. For the F23 function, the BKA-LSOBL algorithm converges in 102 iterations, while the original BKA takes 225 iterations, resulting in a 54.67% improvement. Its average fitness decreased from 1.204 × 104 to 5.824 × 103, representing a 51.63% reduction. Overall, LSOBL demonstrates the most significant enhancement among all tested strategies.
LSOBL is inspired by observing logarithmic spiral patterns in nature, which exhibit self-similarity and infinite scalability. These properties allow the algorithm to search effectively at multiple scales. In the context of opposition learning, the LSOBL strategy introduces a mathematical model based on the logarithmic spiral to generate opposite solutions. Compared to conventional OBL strategies, LSOBL enables wider yet directed exploration, accelerating convergence to the global optimum.
In the original BKA, the update mechanism retains the fitter individuals without exploring the solution space in the opposite direction. LSOBL introduces a spiral-based and dynamically bounded opposition learning model, mathematically defined in Equation (15).
x s p i r a l = x b e s t k θ cos θ + a t + b t x 1 k θ
Here, x b e s t denotes the best individual in the current population. θ = 2 π t / T m a x is the spiral angle that linearly increases with iteration. k   = 0.2 controls the spiral’s tightness. a t = μ t δ t and a t = μ t + δ t define the dynamic opposition boundaries. δ t = δ 0 e λ t is a decay function that gradually reduces the boundary range.
This strategy is embedded into the population update phase of BKA and consists of three steps. (1) Generating opposite solutions based on a logarithmic spiral. (2) Discretizing and applying boundary truncation for discrete parameters. (3) When conducting a competitive update, only the better individual, either the original or the spiral-opposite candidate, is retained. The spiral mechanism dynamically balances global and local search. In the early stages, θ is small, and the spiral radius is large, promoting global exploration. In the later stages, as θ close to 2 π , the spiral becomes tighter, focusing on local exploitation. The elite-guided term x b e s t biases the search toward high-quality regions, avoiding ineffective exploration. A reflection boundary handling mechanism is applied when solutions exceed the defined domain for continuous parameters.
Incorporating LSOBL into BKA’s update mechanism significantly improves the algorithm’s global search ability and robustness. This enhancement effectively mitigates the limitations of the original BKA in dealing with unexplored solution spaces and premature convergence.

3.2.4. Time Complexity Analysis of MIBKA

Time complexity is a critical metric for evaluating the efficiency of an algorithm, representing its performance in the worst-case scenario concerning input size. The core procedure of the MIBKA includes four major phases: chaotic initialization, attack, migration, and population update. Let N denote the population size, D the dimensionality of the search space, and T m a x the maximum number of iterations.
The time complexity of MIBKA is jointly determined by the baseline operations of the BKA and the additional overhead from the proposed improvement strategies. MIBKA employs the circle chaotic map to generate the initial population during the chaotic initialization phase. For each individual, 100 warm-up iterations are conducted to eliminate transient effects. Therefore, the initialization phase has a time complexity of O 100 × N × D , simplifying to O N × D as the constant factor is negligible in asymptotic analysis. During the attack phase, mutation operations inspired by differential evolution, combined with boundary reflection handling and discrete parameter rounding, are applied to each individual across all dimensions, resulting in a time complexity of O N × D . The updates to positions during migration require traversing all individuals and computing dimensional updates, resulting in another O N × D complexity. The phase of population update involves generating logarithmic spiral-based opposition solutions, conducting validation filtering, and comparing elite steps, all of which contribute to the complexity of O N × D . Fitness comparison requires O N , which is asymptotically dominated by the higher-dimensional operations. Thus, the overall complexity remains O N × D .
In summary, the total time complexity per iteration of the MIBKA is O N × D . Considering the entire optimization process over T m a x iterations, the overall time complexity becomes O T m a x × N × D . Although MIBKA introduces additional procedures such as chaotic mapping, differential mutation, and opposition learning, these improvements only marginally increase the per-iteration computation. The algorithm retains the same polynomial-level complexity as the original BKA. Thus, MIBKA maintains strong scalability and practical efficiency for large-scale optimization problems.

3.2.5. The Steps of MIBKA

MIBKA initializes the search space using the circle chaotic map, which ensures a uniformly distributed population. It employs an elite differential mutation strategy to balance global exploration and local exploitation during the optimization process. Furthermore, it integrates logarithmic spiral-based opposition learning to extend the search boundaries and avoid premature convergence. Throughout the optimization process, the model continuously evaluates the fitness of current candidate solutions, updates the direction of search based on elite individuals, and gradually converges towards the near-optimal solution. The complete pseudocode of MIBKA is presented in Algorithm 1.
Algorithm 1: MIBKA (Multi-Strategy Improved Black Kite Algorithm)
Input: Population size N, Dimensionality D, Max iterations T_max
Output: Best solution x_best and its fitness f_best
1:  Initialize population X = {x1, x2, ..., x_N} using circle chaotic map
2:  Evaluate fitness f(xi) for each xi ∈ X
3:  Identify initial leader x_best with best fitness f_best
4:  for t = 1 to T_max do
5:   for each individual xi ∈ X do
6:    Generate rand ∈ [0, 1]
7:    if rand < p_global then
8:     // Global exploration
9:     Update xi using leader-based exploration rule
10:   else
11:    // Local exploitation
12:    Update xi near x_best using local refinement rule
13:   end if
14:   Apply random-elite differential mutation to xi
15:   Evaluate new fitness f(xi)
16:   if f(xi) better than f_best then
17:    Update x_best ← xi, f_best ← f(xi)
18:   end if
19:  end for
20:  // Migration phase
21:  Generate new candidate x_new
22:  Evaluate f(x_new)
23:  if f(x_new) better than f_best then
24:   Update x_best ← x_new, f_best ← f(x_new)
25:  end if
26:  // Spiral opposition-based population update
Input: Population size N, Dimensionality D, Max iterations T_max
27:  for each xi ∈ X do
28:   Generate spiral-opposition solution xi_opp
29:   Evaluate f(xi_opp)
30:   if f(xi_opp) better than f(xi) then

3.3. CNN-BiLSTM Model

To effectively capture local patterns and global semantic structures from textual data, we construct a dual-branch CNN-BiLSTM hybrid feature extraction model. This model extracts local linguistic features using a Convolutional Neural Network (CNN). At the same time, a bidirectional long short-term memory (BiLSTM) network is employed to model sequential semantic dependencies. The outputs of these two branches are concatenated and used for final classification.
The input text is first preprocessed and converted into a sequence of word embeddings with dimensionality d . Given an input text length of L , the embedded representation can be formulated as a matrix.
X R L × d
Multiple one-dimensional convolutional kernels with varying sizes in the CNN branch are applied to capture multi-scale n-gram features. The feature map generated by the i convolutional kernel is computed as
C i = R e L U W i X + b i
W i and b i denote the weight and bias of the i convolutional filter, respectively, and represents the one-dimensional convolution operation. ReLU is the activation function applied elementwise. All convolutional outputs are subject to max pooling along the temporal dimension and concatenated to form a fixed-dimensional local representation vector.
In the BiLSTM branch, the same input matrix X is fed into a bidirectional LSTM network to learn temporal dependencies. Let h denote the number of hidden units in each direction. The hidden state at time step t is computed as
H t = h t h t
The bidirectional hidden states are aggregated via average pooling or attention mechanisms to obtain a global semantic representation vector.
The outputs from the CNN and BiLSTM branches are concatenated to form the final text representation Z .
Z = C N N o u t B i L S T M o u t
Then, the representation vector Z is passed through a fully connected layer, followed by a softmax to generate the predicted label.
The CNN-BiLSTM model involves several structural hyperparameters that significantly affect its performance, including the kernel sizes and counts in CNN, the number of LSTM layers and hidden units, the dropout rate, and the learning rate. These hyperparameters form a high-dimensional mixed search space, and their optimal configuration is crucial for the model’s predictive capability.

3.4. MIBKA-CNN-BiLSTM Model

To enhance the adaptability and parameter tuning efficiency of the CNN-BiLSTM model, we introduce a multi-strategy improved Black Kite Algorithm (MIBKA) to optimize the model’s hyperparameters jointly.
In this framework, the optimized hyperparameters include the number of convolutional kernels, the number of hidden units in the BiLSTM layer, learning rate, batch size, dropout rate, and additional parameters. MIBKA considers cross-validation accuracy to be the fitness function. During the initialization phase, the circle chaotic map is employed to enhance the diversity of initial solutions. A random-to-best differential mutation strategy is utilized during the optimization process to enhance local perturbation capabilities. An elite-driven updating mechanism is implemented to steer the population towards promising areas in the solution space. Moreover, logarithmic spiral-based opposition learning is embedded in the position update phase to improve the quality of solutions further. During each iteration, MIBKA uses the hyperparameter configuration represented by each individual to train a CNN-BiLSTM model and evaluates its performance on a validation set. Through continuous iteration, the algorithm converges toward a near-optimal set of hyperparameters. The final output of MIBKA is used to construct the optimized CNN-BiLSTM model, which is then evaluated on the test set.
This optimization framework enables the model to achieve improved training efficiency and predictive performance highlighting the practical value of MIBKA in complex, high-dimensional optimization scenarios.

4. Experiments

4.1. Experimental Setup

All experiments use an NVIDIA RTX 4090 GPU with 24 GB of VRAM in the MATLAB 2024b environment. The CNN-BiLSTM hybrid model is implemented using the Deep Learning Toolbox, and the optimization process is enhanced using the Parallel Computing Toolbox. The specific parameter settings are detailed in Table 5.

4.2. Evaluation Metrics

To assess the effectiveness of the proposed model, we select accuracy, precision, recall, and F1-score as metrics. The formulas for these metrics are shown in Equations (20)–(23). Table 6 uses TR and TF to denote True Real and True Fake, identifying the model and correctly predicting real and fake examples. Conversely, FR and FF represent False Real and False Fake, indicating that the model incorrectly predicts real and fake examples.
A c c u r a c y = T R + T F T R + F R + T F + F F
P r e c i s i o n = T R T R + F R
R e c a l l = T R T R + F F
F 1 s c o r e = 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l

4.3. Data Collection and Preprocessing

This study collects data from CCTV.com and various fact-checking platforms using web scraping technology to validate the model’s effectiveness. The dataset comprises verified real information published on CCTV.com between April 2023 and November 2024. After removing irrelevant or meaningless content, we obtain 2817 real information. Fake information comes from platforms such as the China Internet Joint Rumor Debunking Platform, Science Rumor Debunking, Popular Science China, and Kedou Wuxianpu, covering debunked information from August 2019 to December 2023. After applying the same processing methods to the real information, 3605 pieces are retained.

4.4. Experimental Analysis

4.4.1. Comparison of Singal Models

This study compares several single models using a self-built dataset to evaluate the superiority of the proposed MIBKA-CNN-BiLSTM model for detecting false information. The single models included CNN, RNN, GRU, LSTM, and BiLSTM. The comparison of experimental results is shown in Table 7. All experiments are conducted under consistent operating conditions and parameter settings to ensure the reliability and reproducibility of the results.
BiLSTM achieves the best performance among all single models, with an 80.28% accuracy and the highest precision, recall, and F1-score. Compared to LSTM, BiLSTM improves accuracy by 1.57%, primarily due to its bidirectional structure. This design allows it to capture both forward and backward contextual information within the text sequence. This ability is crucial for identifying fake information, particularly when it involves implicit logic or misleading context. GRU performs slightly worse than LSTM but better than RNN, suggesting that its gated mechanism enhanced sequence modeling. RNN exhibits the weakest performance, with an accuracy of only 74.07%. This lower accuracy is mainly due to its susceptibility to gradient vanishing and limited capability in modeling long-term dependencies in text. CNN achieves an accuracy of 75.60%, outperforming RNN but falling behind LSTM and BiLSTM. This result suggests that CNN is relatively effective in extracting local semantic features and is particularly skilled at recognizing patterns at the phrase level. However, due to its limited structure, CNN struggled to model long-distance dependencies, resulting in performance degradation when handling longer and structurally complex text.
While CNN is limited when used independently, it is essential for extracting local features. Integrating it with BiLSTM is beneficial due to the complementary strengths of both models. CNN excels at identifying important phrases and structures in static text, while BiLSTM is designed to comprehend the sequence and context of words within the text. Although CNN may not be the highest-performing single model, it plays a crucial role in the hybrid structure by addressing BiLSTM’s limitations in local feature modeling and enhancing the overall discriminative capability of the model.

4.4.2. Comparison of Hybrid Models

Several hybrid models based on single architectures are constructed using fusion optimization algorithms to verify the performance improvement effects of structural enhancement and parameter optimization. The hybrid models include CNN-BiLSTM, GWO-CNN-BiLSTM, WOA-CNN-BiLSTM, BWO-CNN-BiLSTM, and BKA-CNN-BiLSTM. A comparison of the experimental results is shown in Table 8. All the experimental conditions and settings are the same as the single modal comparison experiment.
Compared to the BiLSTM structure alone, the basic hybrid model CNN-BiLSTM has improved the accuracy by 1.55%, and all other metrics have also shown improvement. This result highlights the complementary nature of the CNN and BiLSTM structures. The CNN enhances the local perception ability of the input sequence, while the BiLSTM effectively models the semantic flow between contexts. Together, they provide a more comprehensive understanding of semantics.
The model’s performance improves significantly when the swarm intelligence optimization algorithm is used for adaptive hyperparameter adjustments in the CNN-BiLSTM model. The accuracy rates for the GWO-CNN-BiLSTM, WOA-CNN-BiLSTM, BWO-CNN-BiLSTM, and BKA-CNN-BiLSTM models exceed 83%, showing a marked improvement compared to the unoptimized CNN-BiLSTM. The BKA-CNN-BiLSTM achieves an accuracy of 84.94%, showcasing its robust global search capabilities and effective parameter configuration. Among all the models evaluated, the proposed model performs the best, with an accuracy rate of 88.05% and an F1-score of 86.71%. This improvement can be attributed to several factors. The circle mapping introduced by MIBKA during the population initialization stage enhances diversity. Additionally, a differential mutation mechanism improves the population’s exploration capabilities. Moreover, the logarithmic spiral reverse learning strategy is used in the position update process to prevent the model from getting trapped in local optima. Integrating these three strategies enables the model to maintain stability while finding the optimal hyperparameter combination, substantially enhancing overall detection performance.

4.4.3. Comparison with Baseline Models

To better assess the model’s generalization ability across different datasets, we conduct comparative experiments using the Weibo21 dataset. Developed by Nan et al. in 2021, Weibo21 is a publicly available dataset for detecting fake information across various domains. It contains 4488 fake news items and 4640 real news items, all labeled with domain tags from nine distinct domains. We choose six baseline models to compare with the proposed model. The experimental results are presented in Table 9.
GRU-Attention [53]: This model uses a standard GRU with an attention mechanism to highlight important timesteps. However, despite incorporating the attention module, the GRU’s limited capacity and shallow structure hinder its ability to capture long-range dependencies. This limitation reduces its performance in complex reasoning tasks like fake information detection.
HSA-BiLSTM 54]: This model combines BiLSTM with a hierarchical self-attention mechanism to improve text structure and semantics modeling. Although it effectively extracts layered semantics, it struggles to capture complex semantic relationships and multi-paragraph content that may be misleading or fake. This limitation arises from its sensitivity to training data and a lack of specific tuning for the task at hand.
BERT-BiGRU [55]: This model mixes BERT’s deep contextual embeddings with BiGRU for sequential modeling. It benefits from BERT’s pretraining and BiGRU’s context modeling. However, GRU’s limited expressive power in handling long-text or deep-reasoning scenarios still leads to occasional misclassifications.
DE-BiLSTM [56]: This model uses DE to optimize the hyperparameters of the BiLSTM network. DE effectively addresses the challenge of local optima that often arises in a grid search, enabling a better capture of semantic and sequential features. However, DE’s performance is still limited by its initialization sensitivity and reduced efficiency in high-dimensional spaces.
BERT-CNN [57]: A lightweight architecture that integrates BERT and CNN to harness their strengths effectively. BERT provides rich semantic embeddings, while CNN captures local phrase-level features. Although this combination improves micro-level pattern recognition, the limited receptive field of CNN restricts its ability to model global dependencies. This limitation can affect performance in cases involving contextual jumps or logical confusion.
RoBERTa-WOA-CNN 58]: The model utilizes RoBERTa’s deep language modeling capabilities and incorporates the Whale Optimization Algorithm (WOA) for global hyperparameter tuning. Moreover, CNN layers are used for classification. Despite improvements in robustness and generalization, the model’s performance is still limited by WOA’s slower convergence and the CNN’s restricted ability to represent semantics in cross-domain scenarios.
MIBKA-CNN-BiLSTM: This model integrates CNN for local feature extraction, BiLSTM for modeling bidirectional dependency, and an improved BKA for hyperparameter optimization. As a result, it achieves precise tuning and optimal parameter combinations. Its lightweight structure ensures computational efficiency while maintaining high detection accuracy across domains.
The experimental results show that the model proposed in this paper outperforms all the baseline models every performance metric. Compared with the average of the six baseline models, MIBKA-CNN-BiLSTM improved the accuracy by 1.52% and the F1-score by 1.71%.

4.4.4. t-SNE Analysis

t-Distributed Stochastic Neighbor Embedding (t-SNE) is a powerful nonlinear dimensionality reduction technique that effectively maps high-dimensional data into a two-dimensional space. This mapping helps intuitively understand the distribution of the data. In this study, t-SNE is employed to visualize and analyze the detection performance of the proposed model on both the self-constructed dataset and the public Weibo21 dataset. The results are shown in Figure 5.
In the self-constructed dataset, there are 2817 samples of real information and 3605 samples of fake information, where the model achieves an accuracy of 88.05%. The t-SNE plot clearly shows a significant separation between the samples of real and fake information in the two-dimensional space. This separation strongly indicates that the model successfully identifies the distinguishing features of both categories, enabling reliable classification. Nonetheless, some misclassified points scatter between the clusters of blue and orange dots. These misclassifications primarily result from two factors: the inherent ambiguity in the features of specific samples complicating the model’s judgment and the model’s partial failure to learn some fine-grained distinctions, leading to errors in classification. Meanwhile, the blue and orange dots display a certain degree of intra-class aggregation, clearly reflecting the high similarity among samples within the same category in the feature space. However, the presence of overlapping regions between the two types of data points further highlights that, in these regions, real and fake information share highly similar features, thus increasing the classification difficulty.
The Weibo21 dataset contains 4488 fake and 4640 real information samples. The model achieves an accuracy of 86.72%. The t-SNE visualization shows a separation trend between real and fake samples. However, the mixing degree between the two categories is more prominent than in the self-built dataset. This observation suggests that the Weibo21 dataset presents more complex data characteristics and blurrier class boundaries. Many misclassified points are scattered across the space, confirming the increased classification difficulty of this dataset. Numerous interfering factors and intricate feature relations hinder the model’s ability to distinguish between real and fake content accurately. Although both categories exhibit localized aggregation in certain regions, the level of clustering is lower than that observed in the self-constructed dataset. The significant overlap between the two classes indicates that the feature differences in Weibo21 are less distinct. Therefore, the model must learn more discriminative feature representations to improve classification performance.
A comparative analysis of both datasets reveals that the model performs better on the self-constructed dataset than on Weibo21. This improvement is primarily due to the feature distribution of the self-constructed dataset aligning more closely with the model’s assumptions and learning capabilities. In contrast, the complexity of the Weibo21 dataset exceeds some of the model’s expectations. Future work will optimize the model architecture to strengthen its ability to extract intricate features and enhance its performance on more complex datasets. Alternatively, introducing advanced feature engineering techniques can help identify more distinctive features, which may reduce misclassifications and significantly improve the model’s generalization capability.

5. Conclusions and Future Work

5.1. Conclusions

This paper proposes an improved MIBKA-CNN-BiLSTM hybrid model for detecting fake information. The model enhances the Black Kite Optimization Algorithm by implementing a triple strategy and optimizes the dual-channel deep learning architecture. A high-quality dataset has also been constructed and validated, providing a reliable foundation for model performance evaluation.
Firstly, the MIBKA achieves breakthrough improvements through three key strategies. The first strategy involves the reconstruction of population initialization using circle chaotic mapping, effectively addressing the uneven distribution caused by traditional random initialization. This approach significantly accelerates convergence on the CEC2017 benchmark tests. The second strategy introduces a random-to-elite differential mutation (DE/rand-to-best/1), which replaces the original BKA attack phase’s random perturbation mechanism, establishing a dynamic balance between global exploration and local exploitation. Experiments demonstrate its effectiveness in avoiding local optima in complex optimization tasks. The third strategy designs an LSOBL mechanism that guides the active exploration of the reverse solution space through spiral phase angles and dynamic boundaries, substantially improving search efficiency and robustness. These improvements collectively endow MIBKA with superior parameter optimization capabilities.
Secondly, this study constructs an efficient CNN-BiLSTM dual-channel feature extraction network and employs MIBKA for intelligent hyperparameter optimization. The CNN branch utilizes multi-scale convolutional kernels to precisely capture local anomalous patterns in text, such as modifier stacking and abnormal referencing. Meanwhile, the BiLSTM branch models long-range contextual logical dependencies, such as causal breaks and concept shifts, through bidirectional state propagation. The MIBKA jointly optimizes key hybrid parameters, including the number of convolutional kernels and LSTM units, enabling the model structure to meet task requirements adaptively. Experiments on the self-built dataset show that the optimized model achieves an accuracy of 88.05%, improving by 6.22% over the unoptimized baseline CNN-BiLSTM and by 3.11% compared to the best-performing alternative, validating the effective synergy of architecture and optimization strategies.
Moreover, the proposed model compares with other baseline models on the publicly available Weibo21 dataset, achieving an accuracy of 86.72%, which is 0.41% higher than the best baseline model. The result indicates that the proposed model performs exceptionally well on the self-built dataset and exhibits strong generalization capability on a cross-domain, multi-source heterogeneous public dataset. The benefit primarily arises from MIBKA’s integration of global search and local adaptive mechanisms within complex parameter spaces. This integration enables the CNN-BiLSTM framework to maintain significant expressive power and robust classification across various textual styles and semantic differences. t-SNE analyses further illustrate clear class separability on both datasets, providing solid empirical support for the experimental results.
In summary, this research thoroughly investigates the synergistic application value of swarm intelligence optimization algorithms and deep learning models in fake information detection. Experimental results demonstrate that MIBKA-CNN-BiLSTM significantly outperforms mainstream single models, hybrid models, and pre-trained models on both the self-built and Weibo21 datasets, highlighting its comprehensive advantages in feature extraction, parameter optimization, and cross-scenario generalization. The constructed dataset provides excellent resources for domain research, significantly enhancing the rigor and reliability of model validation.

5.2. Suggestion

The proposed MIBKA-CNN-BiLSTM model demonstrates significant practical application potential in detecting and preventing false information. The efficient, accurate, and adaptable model enables effective performance in real-world scenarios. One application is real-time content review. Due to its lightweight architecture, the model can be integrated into the social media platform’s backend system to scan posts, comments, and shared content in real time. As a result, the system can quickly mark potential false information and shorten its diffusion window period. Another application is trend prediction for dissemination. By combining with social network analysis, the model can identify new clusters of false information. The ability of BiLSTM to capture semantic evolution can predict the changing trends of false information during the dissemination process. This capability enables the platform to limit the spread of high-risk topics actively.

5.3. Limitations and Future Work

Although MIBKA-CNN-BiLSTM excels at detecting fake information, it still has limitations and areas that require improvement. First, the model’s ability to discern ambiguous features or complex texts is inadequate due to its architecture’s lack of depth in understanding deep semantic contradictions and covert narrative strategies. Future research could explore incorporating external knowledge graphs or fact verification modules to enhance the model’s recognition of implicit logical fallacies. Second, although the model exhibits better generalization than baseline models in cross-domain scenarios, there remains room for improvement. Future work should investigate domain adaptation techniques to mitigate performance degradation caused by data distribution shifts. Finally, this study focuses on textual modality. Future work could expand to multimodal fake information detection, exploring cross-modal consistency modeling and joint optimization frameworks.

Author Contributions

Conceptualization, S.Z. and G.M.; methodology, S.Z.; software, X.L.; validation, G.M. and J.M.; formal analysis, S.Z.; investigation, G.M. and J.M.; resources, S.Z. and G.M.; data curation, G.M.; writing—original draft preparation, S.Z.; writing—review and editing, G.M. and J.M.; visualization, X.L.; supervision, S.Z.; project administration, G.M.; funding acquisition, G.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Social Science Fund of China under Grant No. 19BJY246, the Natural Science Fund Project of the Science and Technology Department of Jilin Province under Grant No. 20240101361JC, and the Pioneering Project of Jilin University of Finance and Economics under Grant 2024LH009.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original code and data presented in the study are openly available in GitHub at https://github.com/Kcoroo/MIBKA (accessed on 7 July 2025).

Acknowledgments

The authors are grateful for the financial support from the National Social Science Fund of China under Grant No.19BJY246, the Natural Science Fund Project of the Science and Technology Department of Jilin Province under Grant No.20240101361JC, and the Pioneering Project of Jilin University of Finance and Economics under Grant 2024LH009.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Almansoori, L.; Al-Katheeri, R.; Al-Kfairy, M. Users’ Adoption of Social Media Platforms for Government Services: The Role of Perceived Privacy, Perceived Security, Trust, and Social Influence. In Proceedings of the European Conference on Social Media, Brighton, UK, 30–31 May 2024; Volume 11. [Google Scholar] [CrossRef]
  2. Muhire, J.; Happy, J.B.A.; Mugisha, E.; Majyambere, S.; Ntakirutimana, T.; Shomuyiwa, D.O. Health Information Adoption Behaviour among Users of Social Media Platforms, Kigali-Rwanda: A Cross-Sectional Study; Wiley: Hoboken, NJ, USA, 2025. [Google Scholar] [CrossRef]
  3. Gaozhao, D. Flagging Fake News on Social Media: An Experimental Study of Media Consumers’ Identification of Fake News. Gov. Inf. Q. 2021, 38, 101591. [Google Scholar] [CrossRef]
  4. Kharlamov, A.; Raskhodchikov, A.; Pilgun, M. Information Dissemination and Perception by Social Media Users: Urban Planning Conflicts. In Proceedings of the 3rd International Symposium on Automation, Information and Computing; SCITEPRESS—Science and Technology Publications, Beijing, China, 9–11 December 2022; pp. 125–130. [Google Scholar]
  5. Wu, L.; Morstatter, F.; Carley, K.M.; Liu, H. Misinformation in Social Media; ACM SIGKDD Explorations Newsletter; ACM: New York, NY, USA, 2019; pp. 80–90. [Google Scholar] [CrossRef]
  6. Lazer, D.M.; Baum, M.A.; Benkler, Y.; Berinsky, A.J.; Greenhill, K.M.; Menczer, F.; Metzger, M.J.; Nyhan, B.; Pennycook, G.; Rothschild, D.; et al. The science of fake news. Science 2018, 359, 1094–1096. [Google Scholar] [CrossRef] [PubMed]
  7. Demandsage. Fake News Statistics 2025 (Latest Worldwide Data). Available online: https://www.demandsage.com/fake-news-statistics/ (accessed on 2 July 2025).
  8. Seth, R.; Sharaff, A. Sentiment-Aware Detection Method of Fake News Based on Linguistic Fuzzy Bi-LSTM. In Proceedings of the 2023 OITS International Conference on Information Technology (OCIT), Raipur, India, 13–15 December 2023; IEEE: New York, NY, USA, 2023; pp. 628–633. [Google Scholar]
  9. Sinha, M. Detecting Fake Reviews in E-Commerce Using Linguistic and Sentiment Analysis Features. J. Bus. Res. 2025, 167, 114143. [Google Scholar] [CrossRef]
  10. ResearchGraph. A Deep Dive Into Knowledge Graph Enhanced Pre-trained Language Models. Available online: https://hub.researchgraph.org/a-deep-dive-into-knowledge-graph-enhanced-pre-trained-language-models/ (accessed on 20 August 2025).
  11. Mu, G.; Chen, C.; Li, X.; Chen, Y.; Dai, J.; Li, J. CLAAF: Multimodal Fake Information Detection Based on Contrastive Learning and Adaptive Agg-Modality Fusion. PLoS ONE 2025, 20, e0322556. [Google Scholar] [CrossRef] [PubMed]
  12. Mu, G.; Ju, X.; Yan, H.; Li, J.; Gao, H.; Li, X. An Enhanced Misinformation Detection Model Based on an Improved Beluga Whale Optimization Algorithm and Cross-Modal Feature Fusion. Biomimetics 2025, 10, 128. [Google Scholar] [CrossRef]
  13. Güler, G.; Gündüz, S. Deep Learning Based Fake News Detection on Social Media. Int. J. Inf. Secur. Sci. 2023, 12, 1–21. [Google Scholar] [CrossRef]
  14. Alghamdi, J.; Lin, Y.; Luo, S. A Comparative Study of Machine Learning and Deep Learning Techniques for Fake News Detection. Information 2022, 13, 576. [Google Scholar] [CrossRef]
  15. Pidishetti, R.V.; Chinta, U. Breaking the Fake: A CNN-BiLSTM Approach for Fake News Detection on Social Media. In Proceedings of the 2025 IEEE 15th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, LA, USA, 6–8 January 2025; IEEE: New York, NY, USA, 2025; pp. 1118–1124. [Google Scholar]
  16. Shahriar, A.; Pandit, D.; Rahman, M.S. XLNet-CNN: Combining Global Context Understanding of XLNet with Local Context Capture through Convolution for Improved Multi-Label Text Classification. In Proceedings of the 11th International Conference on Networking, Systems, and Security, Khulna, Bangladesh, 19–21 December 2024; ACM: New York, NY, USA, 2025; pp. 24–31. [Google Scholar]
  17. Du, J.; Zhao, H.; Ye, Z.; Li, M. Bert-BiLSTM Model for Sentiment Analysis Using Contextual Embeddings and Bidirectional Dependencies. In Proceedings of the 2024 International Symposium on Internet of Things and Smart Cities (ISITSC), Nanjing, China, 21–25 June 2024; IEEE: New York, NY, USA, 2024; pp. 88–93. [Google Scholar]
  18. Howard, J.; Ruder, S. Universal Language Model Fine-Tuning for Text Classification. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); Association for Computational Linguistics: Stroudsburg, PA, USA, 2018. [Google Scholar]
  19. Liu, J.; Ma, H.; Xie, X.; Cheng, J. Short Text Classification for Faults Information of Secondary Equipment Based on Convolutional Neural Networks. Energies 2022, 15, 2400. [Google Scholar] [CrossRef]
  20. Singh, A.K.; Karthikeyan, S. Configuration of Neural Network Hyperparameter Using Ant Colony Optimization Algorithm. In Proceedings of the 2024 IEEE 31st International Conference on High Performance Computing, Data and Analytics Workshop (HiPCW), Bangalore, India, 18–21 December 2024; IEEE: New York, NY, USA, 2024; pp. 199–200. [Google Scholar]
  21. Sebastjan, P.; Kuś, W. Method for Parameter Tuning of Hybrid Optimization Algorithms for Problems with High Computational Costs of Objective Function Evaluations. Appl. Sci. 2023, 13, 6307. [Google Scholar] [CrossRef]
  22. Larwuy, L. Optimasi Parameter Artificial Neural Network (ANN) Menggunakan Particle Swarm Optimization (PSO) Untuk Pengkategorian Nasabah Bank. J. Mat. Komput. Stat. 2024, 3, 506–511. [Google Scholar] [CrossRef]
  23. Nia Andriani, L. Optimasi Parameter Support Vector Regression (SVR) Menggunakan Algoritma Grey Wolf Optimizer (GWO). J. Ilm. Mat. 2025, 11, 94–103. [Google Scholar] [CrossRef]
  24. Sun, H.; Yang, S. Range-Free Localization Algorithm Based on Modified Distance and Improved Black-Winged Kite Algorithm. Comput. Netw. 2025, 259, 111091. [Google Scholar] [CrossRef]
  25. Binay, A.; Binay, A.; Register, J. Fake News Detection: Traditional vs. Contemporary Machine Learning Approaches. J. Inf. Knowl. Manag. 2024, 23, 2450075. [Google Scholar] [CrossRef]
  26. Duma, R.A.; Niu, Z.; Nyamawe, A.S.; Tchaye-Kondi, J.; Jingili, N.; Yusuf, A.A.; Deve, A.F. Fake Review Detection Techniques, Issues, and Future Research Directions: A Literature Review. Knowl. Inf. Syst. 2024, 66, 5071–5112. [Google Scholar] [CrossRef]
  27. Fagundes, M.J.G.; Roman, N.T.; Digiampietri, L.A. The Use of Syntactic Information in Fake News Detection: A Systematic Review. SBC Rev. Comput. Sci. 2024, 4, 1–10. [Google Scholar] [CrossRef]
  28. Cao, H.; Guo, W.; Cheng, J. Aspect-Level Sentiment Analysis Based on Aspect Tree and Syntactic Matrix. In Proceedings of the 2024 IEEE 2nd International Conference on Image Processing and Computer Applications (ICIPCA), Shenyang, China, 28–30 June 2024; IEEE: New York, NY, USA, 2024; pp. 34–38. [Google Scholar]
  29. Deva Hema, D.; Rajeeth Jaison, T.; Santhanam, S. Text and Social Media Analytics for Fake News and Hate Speech Detection; Taylor & Francis: Oxfordshire, FL, USA, 2024; pp. 1–22. ISBN 978-100-340-951-9. [Google Scholar]
  30. Gogołek, W. Identification of Vortex Information. Detection of Fake News Eruption Time. Stud. Medioznawcze 2024, 25, 1–12. [Google Scholar] [CrossRef]
  31. Dhanuka, O.; Tiwari, S. Fake News Detection Using Machine Learning. Int. J. Res. Publ. Rev. 2024, 5, 2657–2664. [Google Scholar] [CrossRef]
  32. Hussain, A.; Sabel, B.; Thiel, M.; Nürnberger, A. Automated Detection of Fake Biomedical Papers: A Machine Learning Perspective. In Proceedings of the 27th International Conference on Enterprise Information Systems; SCITEPRESS—Science and Technology Publications, Porto, Portugal, 4–6 April 2025; pp. 662–670. [Google Scholar]
  33. Hadi, Z.; Utami, E.; Ariatmanto, D. Detect Fake Reviews Using Random Forest and Support Vector Machine. SinkrOn 2023, 8, 623–630. [Google Scholar] [CrossRef]
  34. Leung, J.; Vatsalan, D.; Arachchilage, N. Feature Analysis of Fake News: Improving Fake News Detection in Social Media. J. Cyber Secur. Technol. 2023, 7, 224–241. [Google Scholar] [CrossRef]
  35. Aslam, N.; Ullah Khan, I.; Alotaibi, F.S.; Aldaej, L.A.; Aldubaikil, A.K. Fake Detect: A Deep Learning Ensemble Model for Fake News Detection. Complexity 2021, 2021, 5557784. [Google Scholar] [CrossRef]
  36. Prathima, P. An Enhanced XGBoost Machine Learning Model to Detect Fake Social Media Accounts. J. Inf. Syst. Eng. Manag. 2025, 10, 372–383. [Google Scholar] [CrossRef]
  37. Alnabrisi, I.; Saad, M. Detect Arabic Fake News Through Deep Learning Models and Transformers. Expert Syst. Appl. 2023, 251, 123997. [Google Scholar] [CrossRef]
  38. Zhang, L.; Zhang, X.; Zhou, Z.; Zhang, X.; Li, C.; Yu, P.S. Knowledge-Aware Multimodal Pre-Training for Fake News Detection. Inf. Fusion 2024, 114, 102715. [Google Scholar] [CrossRef]
  39. Collen, D.R.; Nyandoro, L.K.; Zvarevashe, K. Fake News Detection Using 5L-CNN. In Proceedings of the 2022 1st Zimbabwe Conference of Information and Communication Technologies (ZCICT), Harare, Zimbabwe, 9–10 November 2022; IEEE: New York, NY, USA, 2022; pp. 1–7. [Google Scholar]
  40. Kekan, S.; Kharate, A.; Gajare, P.; Bhosale, S.; Dashrath, P. Fake Bank Name Detection Using LSTM and RNN; Springer: Berlin/Heidelberg, Germany, 2025. [Google Scholar] [CrossRef]
  41. Rajeev, A.; Raviraj, P. Multi Cascaded Face Artefact Detection with Xception Convoluted LSTM Network for Deep Fake Detection. J. Inf. Syst. Eng. Manag. 2025, 10, 687–701. [Google Scholar] [CrossRef]
  42. Sahi, A.; Albdair, M.; Diykh, M.; Abdulla, S.; Alghayab, H.; Aljebur, K.; Alkhafaji, S.K.D. Sgdm-Gru: Spectral Graph Deep Learning Based Gated Recurrent Unit Model for Accurate Fake News Detection. Expert Syst. Appl. 2025, 281, 127572. [Google Scholar] [CrossRef]
  43. Abbas, Q.; Zeshan, M.U.; Asif, M. A CNN-RNN Based Fake News Detection Model Using Deep Learning. In Proceedings of the 2022 International Seminar on Computer Science and Engineering Technology (SCSET), Indianapolis, IN, USA, 8–9 January 2022; IEEE: New York, NY, USA, 2022; pp. 40–45. [Google Scholar]
  44. Bharal, R.; Jonnakuti, S. Social Media Sentiment Analysis Using CNN-BiLSTM. Int. J. Sci. Res. 2021, 10, 656–661. [Google Scholar] [CrossRef]
  45. Chavhan, S.; Dharmik, R.C.; Jain, S. Evaluation of CNN-BiGRU and CNN-BiLSTM Model for Fake Job Post Detection: A Deep Learning Approach. In Proceedings of the 2024 2nd International Conference on Emerging Trends in Engineering and Medical Sciences (ICETEMS), Nagpur, India, 22–23 November 2024; IEEE: New York, NY, USA, 2024; pp. 1–8. [Google Scholar]
  46. Maurya, J.P.; Richhariya, V.; Gour, B.; Kumar, V. Implementation of Weight Adjusting GNN With Differentiable Pooling for User Preference-Aware Fake News Detection. Res. Sq. 2024. [Google Scholar] [CrossRef]
  47. Gupta, P.; Gupta, P.K. Performance Analysis of GCN, GNN, and GAT Models with Differentiable Pooling for Detection of Fake News. In Proceedings of the 2024 3rd Edition of IEEE Delhi Section Flagship Conference (DELCON), New Delhi, India, 21–23 November 2024; IEEE: New York, NY, USA, 2024; pp. 1–6. [Google Scholar]
  48. Qin, S.; Zhang, M. Boosting Generalization of Fine-Tuning BERT for Fake News Detection. Inf. Process. Manag. 2024, 61, 103745. [Google Scholar] [CrossRef]
  49. Pavlov, T.; Mirceva, G. COVID-19 Fake News Detection by Using BERT and RoBERTa Models. In Proceedings of the 2022 45th Jubilee International Convention on Information, Communication and Electronic Technology (MIPRO), Opatija, Croatia, 23–27 May 2022; IEEE: New York, NY, USA, 2022; pp. 312–316. [Google Scholar]
  50. Prabu, M.; Thinesh, L.B.; Sreekumar, M. Enhanced Fake News Detection Leveraging a Hybrid BERT-BiLSTM Model. In Proceedings of the 2024 Global Conference on Communications and Information Technologies (GCCIT), Bangalore, India, 25–26 October 2024; IEEE: New York, NY, USA, 2024; pp. 1–6. [Google Scholar]
  51. Jonnakuti, S. Enabling Scalable GPU Clusters for Distributed Deep Learning in the Cloud. Int. J. Sci. Technol. 2018, 9. [Google Scholar] [CrossRef]
  52. Yao, L. A Knowledge Enhanced Pre-Training Model for Chinese Weibo Sentiment Analysis. In Proceedings of the 2024 9th International Conference on Cloud Computing and Big Data Analytics (ICCCBDA), Chengdu, China, 25–27 April 2024; IEEE: New York, NY, USA, 2024; pp. 33–39. [Google Scholar]
  53. Zhang, X.; Yu, L.; Tian, S. BGAT: Aspect-Based Sentiment Analysis Based on Bidirectional GRU and Graph Attention Network. J. Intell. Fuzzy Syst. 2023, 44, 3115–3126. [Google Scholar] [CrossRef]
  54. Guo, Z.; Yang, J. Rumor Detection on Twitter with Hierarchical Attention Neural Networks. In Proceedings of the 2018 IEEE 9th International Conference on Software Engineering and Service Science (ICSESS), Beijing, China, 23–25 November 2018; IEEE: New York, NY, USA, 2018; pp. 783–787. [Google Scholar]
  55. Sr, S.M.; Ahmad, S. BERT Based Blended Approach for Fake News Detection. J. Big Data Artif. Intell. 2024, 2, 7–15. [Google Scholar] [CrossRef]
  56. Ganesh, V.; Kamarasan, M. Parameter Tuned Bi-Directional Long Short Term Memory Based Emotion with Intensity Sentiment Classification Model Using Twitter Data. In Proceedings of the 2020 International Conference on System, Computation, Automation and Networking (ICSCAN), Puducherry, India, 27–28 March 2020; IEEE: New York, NY, USA, 2020; pp. 1–6. [Google Scholar]
  57. Verma, P.K.; Agrawal, P.; Madaan, V.; Prodan, R. MCred: Multi-Modal Message Credibility for Fake News Detection Using BERT and CNN. J. Ambient Intell. Humaniz. Comput. 2022, 14, 10617–10629. [Google Scholar] [CrossRef] [PubMed]
  58. Pawestri, S.; Murinto, M.; Auzan, M. Sarcasm Detection: A Comparative Analysis of RoBERTa-CNN vs RoBERTa-RNN Architectures. Innov. Res. Inform. (Innov.) 2024, 6, 118–125. [Google Scholar] [CrossRef]
Figure 1. Architecture of the proposed MIBKA-CNN-BiLSTM model.
Figure 1. Architecture of the proposed MIBKA-CNN-BiLSTM model.
Biomimetics 10 00562 g001
Figure 2. Best Fitness of BKA improved by different chaotic maps.
Figure 2. Best Fitness of BKA improved by different chaotic maps.
Biomimetics 10 00562 g002
Figure 3. Best Fitness of BKA improved by different mutation strategies.
Figure 3. Best Fitness of BKA improved by different mutation strategies.
Biomimetics 10 00562 g003
Figure 4. Best Fitness of BKA improved by different OBL strategies.
Figure 4. Best Fitness of BKA improved by different OBL strategies.
Biomimetics 10 00562 g004
Figure 5. t-SNE visualization results.
Figure 5. t-SNE visualization results.
Biomimetics 10 00562 g005
Table 1. Hybrid Models in Fake Information Detection.
Table 1. Hybrid Models in Fake Information Detection.
ModelKey FeaturesAdvantagesLimitations
CNN-RNNlocal features + temporal dependenciesFuses local-global features; suited for fake newsRNN weak in long-distance dependency modeling
CNN-BiLSTMlocal features + bidirectional temporal modelingCaptures bidirectional context; robust to semanticsPerformance relies on hyperparameters
CNN-BiGRU-Attlocal features + simplified gating + attention mechanismFaster convergence; attention focuses on key semantic regions, boosting detection accuracyAttention adds computational overhead; may over-focus on noisy features
GNNweight-adjusted + differentiable poolingIntegrates user factors; enhances graph aggregationRelies on quality of user preference data
GCN/GNN/GATdiverse graph models + poolingExplores varied graph structures; improves down samplingPerformance varies by graph model
Table 2. Classification of chaotic maps.
Table 2. Classification of chaotic maps.
TypeIncluded MapsMathematical Characteristics
ContinuousCircle, Sine, Logistic, Gauss, Sinusoidal, Cubic, Henon, Fuch, Chebyshev, ICMIC, Singer, SPMGenerate smooth, ergodic sequences via continuous nonlinear equations, suitable for continuous parameters
Discrete-segmentBernoulli, PiecewiseGenerate jump sequences via piecewise functions, suitable for discrete parameters
Hybrid-modalLogistic-Tent-Cosine, Tent-Logistic-Cosine, Logistic-Sine-Cosine, Sine-Tent-Cosine, Iterative, Tent, KentCombine continuous and discrete features, adaptable to mixed parameter spaces
Table 3. Classification of mutation strategies.
Table 3. Classification of mutation strategies.
CategoryMutation StrategyCore Principle
Differential EvolutionDE/rand-to-best/1,
DE/best/1, DE/rand/2, DE/best/2
Generate new solutions via difference vectors between random and elite individuals
Probability-DrivenGuass1, Guass2, Cauchy1, Cauchy2, HUse stochastic perturbations from Gaussian or Cauchy distributions to adjust step size
Adaptive Regulationt, Self-t, PeriodicDynamically adjust mutation intensity according to iteration progress
Hybrid HeuristicsCloudCombine fuzzy logic with stochastic cloud models for uncertain perturbation
Table 4. Classification of OBL strategies.
Table 4. Classification of OBL strategies.
CategoryStrategiesCore Characteristics
Static MappingBasic OBL (OBL), Improved OBL (IOBL)Generate opposite solutions using fixed formulas with constant boundaries
Adaptive RegulationAdaptive OBL (AOBL), Dynamic OBL (DOBL)Adjust opposite boundaries or perturbation strength dynamically based on iteration progress
Hybrid HeuristicsChaotic OBL (COBL), Spiral OBL (SOBL) Logarithmic Spiral OBL (LSOBL)Use nonlinear mechanisms like chaos and spiral functions to enhance diversity
Elite-DrivenElite OBL (EOBL), Quasi OBL (QOBL), Weighted OBL (WOBL)Utilize elite or best-performing individuals to guide opposite solution generation
Probability-BasedBeta OBL (BOBL), Fast Random OBL (FROBL)Generate opposite solutions using Beta distribution or random sampling to balance exploration and exploitation
Table 5. Experimental parameter settings.
Table 5. Experimental parameter settings.
Module/ProcessParameter NameValue/RangeTypeDescription
MIBKAPopulation Size N 50FixedNumber of individuals in the initial population
Maximum Iterations T m a x 100FixedStopping criterion for hyperparameter tuning
Scaling Factor F 0.8FixedMutation factor in DE/rand-to-best/1
Spiral Tightness k 0.2FixedControls LSOBL search scope
CNN-BiLSTMInput Sequence Length256FixedTexts padded/truncated to uniform length
Word Embedding Dimension300FixedGloVe pre-trained embedding dimension
Number of Conv Layers L c o n v {1, 2, 3}DiscreteInteger range for layer depth
Number of Kernels K {32, 64, 128}DiscreteNumber of kernels per convolutional layer
BiLSTM Hidden Units H l s t m {64, 128, 256}DiscreteNumber of hidden units in BiLSTM
Dropout Rate β [0.2, 0.6]ContinuousUniform sampling from interval
Training & TestingOptimizerAdamFixedLearning rate set to 0.0001
L2 Regularization λ [1 × 10−5, 1 × 10−3]ContinuousLog-uniform sampling
Batch Size B {32, 64, 128}DiscreteNumber of samples per batch
Epochs100FixedEarly stopping enabled (patience = 10)
Table 6. Confusion matrix for fake information detection.
Table 6. Confusion matrix for fake information detection.
Prediction/ActualRealFake
RealTrue Real (TR)False Real (FR)
FakeFalse Fake (FF)True Fake (TF)
Table 7. Experimental results of the single models.
Table 7. Experimental results of the single models.
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
CNN75.6074.5571.1972.81
RNN74.0772.7769.4971.10
GRU77.1576.3372.8874.58
LSTM78.7178.1074.5876.29
BiLSTM 80.28 79.87 76.27 78.07
Table 8. Experimental results of the hybrid models.
Table 8. Experimental results of the hybrid models.
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
CNN-BiLSTM81.8381.6577.9779.69
GWO-CNN-BiLSTM84.3683.9281.0882.48
WOA-CNN-BiLSTM84.1684.4980.4182.38
BWO-CNN-BiLSTM83.4183.4179.6681.51
BKA-CNN-BiLSTM84.9485.2181.3683.21
MIBKA-CNN-BiLSTM 88.05 88.74 84.75 86.71
Table 9. Baseline models comparison on the Weibo21 dataset.
Table 9. Baseline models comparison on the Weibo21 dataset.
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
GRU-Attention [53]83.6484.0782.8883.47
HAS-BiLSTM [54]84.3284.1183.9384.02
BERT-BiGRU [55]85.2785.1985.0385.11
DE-BiLSTM [56]85.7485.8985.3385.61
BERT-CNN [57]85.9185.7685.6385.69
RoBERTa-WOA-CNN [58]86.3186.1886.0286.10
MIBKA-CNN-BiLSTM 86.72 86.89 86.54 86.71
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, S.; Mu, G.; Ma, J.; Li, X. An Enhanced MIBKA-CNN-BiLSTM Model for Fake Information Detection. Biomimetics 2025, 10, 562. https://doi.org/10.3390/biomimetics10090562

AMA Style

Zhu S, Mu G, Ma J, Li X. An Enhanced MIBKA-CNN-BiLSTM Model for Fake Information Detection. Biomimetics. 2025; 10(9):562. https://doi.org/10.3390/biomimetics10090562

Chicago/Turabian Style

Zhu, Sining, Guangyu Mu, Jie Ma, and Xiurong Li. 2025. "An Enhanced MIBKA-CNN-BiLSTM Model for Fake Information Detection" Biomimetics 10, no. 9: 562. https://doi.org/10.3390/biomimetics10090562

APA Style

Zhu, S., Mu, G., Ma, J., & Li, X. (2025). An Enhanced MIBKA-CNN-BiLSTM Model for Fake Information Detection. Biomimetics, 10(9), 562. https://doi.org/10.3390/biomimetics10090562

Article Metrics

Back to TopTop