Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (699)

Search Parameters:
Keywords = naive Bayes classification

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 2526 KB  
Article
An Explainable Deep Learning Framework with Adaptive Feature Selection for Smart Lemon Disease Classification in Agriculture
by Naeem Ullah, Michelina Ruocco, Antonio Della Cioppa, Ivanoe De Falco and Giovanna Sannino
Electronics 2025, 14(19), 3928; https://doi.org/10.3390/electronics14193928 - 2 Oct 2025
Abstract
Early and accurate detection of lemon disease is necessary for effective citrus crop management. Traditional approaches often lack refined diagnosis, necessitating more powerful solutions. The article introduces adaptive PSO-LemonNetX, a novel framework integrating a novel deep learning model, adaptive Particle Swarm Optimization (PSO)-based [...] Read more.
Early and accurate detection of lemon disease is necessary for effective citrus crop management. Traditional approaches often lack refined diagnosis, necessitating more powerful solutions. The article introduces adaptive PSO-LemonNetX, a novel framework integrating a novel deep learning model, adaptive Particle Swarm Optimization (PSO)-based feature selection, and explainable AI (XAI) using LIME. The approach improves the accuracy of classification while also enhancing the explainability of the model. Our end-to-end model obtained 97.01% testing and 98.55% validation accuracy. Performance was enhanced further with adaptive PSO and conventional classifiers—100% validation accuracy using Naive Bayes and 98.8% testing accuracy using Naive Bayes and an SVM. The suggested PSO-based feature selection performed better than ReliefF, Kruskal–Wallis, and Chi-squared approaches. Due to its lightweight design and good performance, this approach can be adapted for edge devices in IoT-enabled smart farms, contributing to sustainable and automated disease detection systems. These results show the potential of integrating deep learning, PSO, grid search, and XAI into smart agriculture workflows for enhancing agricultural disease detection and decision-making. Full article
(This article belongs to the Special Issue Image Processing and Pattern Recognition)
Show Figures

Figure 1

32 pages, 2827 KB  
Article
Understanding Post-COVID-19 Household Vehicle Ownership Dynamics Through Explainable Machine Learning
by Mahbub Hassan, Saikat Sarkar Shraban, Ferdoushi Ahmed, Mohammad Bin Amin and Zoltán Nagy
Future Transp. 2025, 5(4), 136; https://doi.org/10.3390/futuretransp5040136 - 2 Oct 2025
Abstract
Understanding household vehicle ownership dynamics in the post-COVID-19 era is critical for designing equitable, resilient, and sustainable transportation policies. This study employs an interpretable machine learning framework to model household vehicle ownership using data from the 2022 National Household Travel Survey (NHTS)—the first [...] Read more.
Understanding household vehicle ownership dynamics in the post-COVID-19 era is critical for designing equitable, resilient, and sustainable transportation policies. This study employs an interpretable machine learning framework to model household vehicle ownership using data from the 2022 National Household Travel Survey (NHTS)—the first nationally representative U.S. dataset collected after the onset of the pandemic. A binary classification task distinguishes between single- and multi-vehicle households, applying an ensemble of algorithms, including Random Forest, XGBoost, Support Vector Machines (SVM), and Naïve Bayes. The Random Forest model achieved the highest predictive accuracy (86.9%). To address the interpretability limitations of conventional machine learning approaches, SHapley Additive exPlanations (SHAP) were applied to extract global feature importance and directionality. Results indicate that the number of drivers, household income, and vehicle age are the most influential predictors of multi-vehicle ownership, while contextual factors such as housing tenure, urbanicity, and household lifecycle stage also exert substantial influence highlighting the spatial and demographic heterogeneity in ownership behavior. Policy implications include the design of equity-sensitive strategies such as targeted mobility subsidies, vehicle scrappage incentives, and rural transit innovations. By integrating explainable artificial intelligence into national-scale transportation modeling, this research bridges the gap between predictive accuracy and interpretability, contributing to adaptive mobility strategies aligned with the United Nations Sustainable Development Goals (SDGs), particularly SDG 11 (Sustainable Cities), SDG 10 (Reduced Inequalities), and SDG 13 (Climate Action). Full article
Show Figures

Figure 1

27 pages, 4168 KB  
Article
Electromyographic Diaphragm and Electrocardiographic Signal Analysis for Weaning Outcome Classification in Mechanically Ventilated Patients
by Alejandro Arboleda, Manuel Franco, Francisco Naranjo and Beatriz Fabiola Giraldo
Sensors 2025, 25(19), 6000; https://doi.org/10.3390/s25196000 - 29 Sep 2025
Abstract
Early prediction of weaning outcomes in mechanically ventilated patients has significant potential to influence the duration of treatment as well as associated morbidity and mortality. This study aimed to investigate the utility of signal analysis using electromyographic diaphragm (EMG) and electrocardiography (ECG) signals [...] Read more.
Early prediction of weaning outcomes in mechanically ventilated patients has significant potential to influence the duration of treatment as well as associated morbidity and mortality. This study aimed to investigate the utility of signal analysis using electromyographic diaphragm (EMG) and electrocardiography (ECG) signals to classify the success or failure of weaning in mechanically ventilated patients. Electromyographic signals of 40 subjects were recorded using 5-channel surface electrodes placed around the diaphragm muscle, along with an ECG recording through a 3-lead Holter system during extubation. EMG and ECG signals were recorded from mechanically ventilated patients undergoing weaning trials. Linear and nonlinear signal analysis techniques were used to assess the interaction between diaphragm muscle activity and cardiac activity. Supervised machine learning algorithms were then used to classify the weaning outcomes. The study revealed clear differences in diaphragmatic and cardiac patterns between patients who succeeded and failed in the weaning trials. Successful weaning was characterised by a higher ECG-derived respiration amplitude, whereas failed weaning was characterised by an elevated EMG amplitude. Furthermore, successful weaning exhibited greater oscillations in diaphragmatic muscle activity. Spectral analysis and parameter extraction identified 320 parameters, of which 43 were significant predictors of weaning outcomes. Using seven of these parameters, the Naive Bayes classifier demonstrated high accuracy in classifying weaning outcomes. Surface electromyographic and electrocardiographic signal analyses can predict weaning outcomes in mechanically ventilated patients. This approach could facilitate the early identification of patients at risk of weaning failure, allowing for improved clinical management. Full article
(This article belongs to the Section Biomedical Sensors)
Show Figures

Figure 1

7 pages, 333 KB  
Proceeding Paper
Predictive Analysis of Chronic Kidney Disease in Machine Learning
by Husnain Ali Haider, Manzoor Hussain and Ivana Lucia Kharisma
Eng. Proc. 2025, 107(1), 118; https://doi.org/10.3390/engproc2025107118 - 29 Sep 2025
Abstract
Chronic kidney disease is a systemic disease of multiple factors and slow progression, and is now becoming a rapidly changing global pathological problem affecting healthcare systems. Anyone who can go through diagnosis before getting to stage 5 Chronic Kidney Disease (CKD) or end [...] Read more.
Chronic kidney disease is a systemic disease of multiple factors and slow progression, and is now becoming a rapidly changing global pathological problem affecting healthcare systems. Anyone who can go through diagnosis before getting to stage 5 Chronic Kidney Disease (CKD) or end stage renal failure has a better shot at the result. This work involves 1659 patient records and dependent variables include demographics, lifestyle, and clinical biochemistry of CKD. Based on the supervised techniques of machine learning which are Random Forest, K Nearest Neighbors (KNN), Logistic Regression, and Naïve Bayes, it was agreed that the performance of the model metrics such as accuracy, precision and recall would need to be used. These models were applied, evaluated by means of more or less simple effectiveness parameters including, for instance, accuracy, precision, or recall. Out of these [best algorithm] achieved [accuracy value] % of predictive accuracy in CKD, and so can be used for diagnosis of CKD in its early stage. This work offers the Framework and results in the development of data-integrated approaches in healthcare and improves the disease control and management. Full article
Show Figures

Figure 1

35 pages, 3558 KB  
Article
Realistic Performance Assessment of Machine Learning Algorithms for 6G Network Slicing: A Dual-Methodology Approach with Explainable AI Integration
by Sümeye Nur Karahan, Merve Güllü, Deniz Karhan, Sedat Çimen, Mustafa Serdar Osmanca and Necaattin Barışçı
Electronics 2025, 14(19), 3841; https://doi.org/10.3390/electronics14193841 - 27 Sep 2025
Abstract
As 6G networks become increasingly complex and heterogeneous, effective classification of network slicing is essential for optimizing resources and managing quality of service. While recent advances demonstrate high accuracy under controlled laboratory conditions, a critical gap exists between algorithm performance evaluation under idealized [...] Read more.
As 6G networks become increasingly complex and heterogeneous, effective classification of network slicing is essential for optimizing resources and managing quality of service. While recent advances demonstrate high accuracy under controlled laboratory conditions, a critical gap exists between algorithm performance evaluation under idealized conditions and their actual effectiveness in realistic deployment scenarios. This study presents a comprehensive comparative analysis of two distinct preprocessing methodologies for 6G network slicing classification: Pure Raw Data Analysis (PRDA) and Literature-Validated Realistic Transformations (LVRTs). We evaluate the impact of these strategies on algorithm performance, resilience characteristics, and practical deployment feasibility to bridge the laboratory–reality gap in 6G network optimization. Our experimental methodology involved testing eleven machine learning algorithms—including traditional ML, ensemble methods, and deep learning approaches—on a dataset comprising 10,000 network slicing samples (expanded to 21,033 through realistic transformations) across five network slice types. The LVRT methodology incorporates realistic operational impairments including market-driven class imbalance (9:1 ratio), multi-layer interference patterns, and systematic missing data reflecting authentic 6G deployment challenges. The experimental results revealed significant differences in algorithm behavior between the two preprocessing approaches. Under PRDA conditions, deep learning models achieved perfect accuracy (100% for CNN and FNN), while traditional algorithms ranged from 60.9% to 89.0%. However, LVRT results exposed dramatic performance variations, with accuracies spanning from 58.0% to 81.2%. Most significantly, we discovered that algorithms achieving excellent laboratory performance experience substantial degradation under realistic conditions, with CNNs showing an 18.8% accuracy loss (dropping from 100% to 81.2%), FNNs experiencing an 18.9% loss (declining from 100% to 81.1%), and Naive Bayes models suffering a 34.8% loss (falling from 89% to 58%). Conversely, SVM (RBF) and Logistic Regression demonstrated counter-intuitive resilience, improving by 14.1 and 10.3 percentage points, respectively, under operational stress, demonstrating superior adaptability to realistic network conditions. This study establishes a resilience-based classification framework enabling informed algorithm selection for diverse 6G deployment scenarios. Additionally, we introduce a comprehensive explainable artificial intelligence (XAI) framework using SHAP analysis to provide interpretable insights into algorithm decision-making processes. The XAI analysis reveals that Packet Loss Budget emerges as the dominant feature across all algorithms, while Slice Jitter and Slice Latency constitute secondary importance features. Cross-scenario interpretability consistency analysis demonstrates that CNN, LSTM, and Naive Bayes achieve perfect or near-perfect consistency scores (0.998–1.000), while SVM and Logistic Regression maintain high consistency (0.988–0.997), making them suitable for regulatory compliance scenarios. In contrast, XGBoost shows low consistency (0.106) despite high accuracy, requiring intensive monitoring for deployment. This research contributes essential insights for bridging the critical gap between algorithm development and deployment success in next-generation wireless networks, providing evidence-based guidelines for algorithm selection based on accuracy, resilience, and interpretability requirements. Our findings establish quantitative resilience boundaries: algorithms achieving >99% laboratory accuracy exhibit 58–81% performance under realistic conditions, with CNN and FNN maintaining the highest absolute accuracy (81.2% and 81.1%, respectively) despite experiencing significant degradation from laboratory conditions. Full article
Show Figures

Figure 1

21 pages, 2807 KB  
Article
Discrimination of Multiple Foliar Diseases in Wheat Using Novel Feature Selection and Machine Learning
by Sen Zhuang, Yujuan Huang, Jie Zhu, Qingluo Yang, Wei Li, Yangyang Gu, Tongjie Li, Hengbiao Zheng, Chongya Jiang, Tao Cheng, Yongchao Tian, Yan Zhu, Weixing Cao and Xia Yao
Remote Sens. 2025, 17(19), 3304; https://doi.org/10.3390/rs17193304 - 26 Sep 2025
Abstract
Wheat, a globally vital food crop, faces severe threats from numerous foliar diseases, which often infect agricultural fields, significantly compromising yield and quality. Rapid and accurate identification of the specific disease is crucial for ensuring food security. Although progress has been made in [...] Read more.
Wheat, a globally vital food crop, faces severe threats from numerous foliar diseases, which often infect agricultural fields, significantly compromising yield and quality. Rapid and accurate identification of the specific disease is crucial for ensuring food security. Although progress has been made in wheat foliar disease detection using RGB imaging and spectroscopy, most prior studies have focused on identifying the presence of a single disease, without considering the need to operationalize such methods, and it will be necessary to differentiate between multiple diseases. In this study, we systematically investigate the differentiation of three wheat foliar diseases (e.g., powdery mildew, stripe rust, and leaf rust) and evaluate feature selection strategies and machine learning models for disease identification. Based on field experiments conducted from 2017 to 2024 employing artificial inoculation, we established a standardized hyperspectral database of wheat foliar diseases classified by disease severity. Four feature selection methods were employed to extract spectral features prior to classification: continuous wavelet projection algorithm (CWPA), continuous wavelet analysis (CWA), successive projections algorithm (SPA), and Relief-F. The selected features (which are derived by CWPA, CWA, SPA, and Relief-F algorithm) were then used as predictors for three disease-identification machine learning models: random forest (RF), k-nearest neighbors (KNN), and naïve Bayes (BAYES). Results showed that CWPA outperformed other feature selection methods. The combination of CWPA and KNN for discriminating disease-infected (powdery mildew, stripe rust, leaf rust) and healthy leaves by using only two key features (i.e., 668 nm at wavelet scale 5 and 894 nm at wavelet scale 7), achieved an overall accuracy (OA) of 77% and a map-level image classification efficacy (MICE) of 0.63. This combination of feature selection and machine learning model provides an efficient and precise procedure for discriminating between multiple foliar diseases in agricultural fields, thus offering technical support for precision agriculture. Full article
Show Figures

Figure 1

25 pages, 2375 KB  
Article
Evaluating the Effectiveness of Large Language Models (LLMs) Versus Machine Learning (ML) in Identifying and Detecting Phishing Email Attempts
by Saed Tarapiah, Linda Abbas, Oula Mardawi, Shadi Atalla, Yassine Himeur and Wathiq Mansoor
Algorithms 2025, 18(10), 599; https://doi.org/10.3390/a18100599 - 25 Sep 2025
Abstract
Phishing emails remain a significant concern and a growing cybersecurity threat in online communication. They often bypass traditional filters due to their increasing sophistication. This study presents a comparative evaluation of machine learning (ML) models and transformer-based large language models (LLMs) for phishing [...] Read more.
Phishing emails remain a significant concern and a growing cybersecurity threat in online communication. They often bypass traditional filters due to their increasing sophistication. This study presents a comparative evaluation of machine learning (ML) models and transformer-based large language models (LLMs) for phishing email detection, with embedded URL analysis. This study assessed ML training and LLM fine-tuning on both balanced and imbalanced datasets. We evaluated multiple ML models, including Random Forest, Logistic Regression, Support Vector Machine, Naïve Bayes, Gradient Boosting, Decision Tree, and K-Nearest Neighbors, alongside transformer-based LLMs DistilBERT, ALBERT, BERT-Tiny, ELECTRA, MiniLM, and RoBERTa. To further enhance realism, phishing emails generated by LLMs were included in the evaluation. Across all configurations, both the ML models and the fine-tuned LLMs demonstrated robust performance. Random Forest achieved over 98% accuracy in both email detection and URL classification. DistilBERT obtained almost as high scores on emails and URLs. Balancing the dataset led to slight accuracy gains in ML models but minor decreases in LLMs, likely due to their sensitivity to majority class reductions during training. Overall, LLMs are highly effective at capturing complex language patterns, while traditional ML models remain efficient and require low computational resources. Combining both approaches through a hybrid or ensemble method could enhance phishing detection effectiveness. Full article
(This article belongs to the Section Evolutionary Algorithms and Machine Learning)
Show Figures

Figure 1

22 pages, 3646 KB  
Article
Machine Learning in the Classification of RGB Images of Maize (Zea mays L.) Using Texture Attributes and Different Doses of Nitrogen
by Thiago Lima da Silva, Fernanda de Fátima da Silva Devechio, Marcos Silva Tavares, Jamile Raquel Regazzo, Edson José de Souza Sardinha, Liliane Maria Romualdo Altão, Gabriel Pagin, Adriano Rogério Bruno Tech and Murilo Mesquita Baesso
AgriEngineering 2025, 7(10), 317; https://doi.org/10.3390/agriengineering7100317 - 23 Sep 2025
Viewed by 130
Abstract
Nitrogen fertilization is decisive for maize productivity, fertilizer use efficiency, and sustainability, which calls for fast and nondestructive nutritional diagnosis. This study evaluated the classification of maize plant nutritional status from red, green, and blue (RGB) leaf images using texture attributes. A greenhouse [...] Read more.
Nitrogen fertilization is decisive for maize productivity, fertilizer use efficiency, and sustainability, which calls for fast and nondestructive nutritional diagnosis. This study evaluated the classification of maize plant nutritional status from red, green, and blue (RGB) leaf images using texture attributes. A greenhouse experiment was conducted under a completely randomized factorial design with four nitrogen doses, one maize hybrid Pioneer 30F35, and four replicates, at two sampling times corresponding to distinct phenological stages, totaling thirty-two experimental units. Images were processed with the gray-level cooccurrence matrix computed at three distances 1, 3, and 5 pixels and four orientations 0°, 45°, 90°, and 135°, yielding eight texture descriptors that served as inputs to five supervised classifiers: an artificial neural network, a support vector machine, k nearest neighbors, a decision tree, and Naive Bayes. The results indicated that texture descriptors discriminated nitrogen doses with good performance and moderate computational cost, and that homogeneity, dissimilarity, and contrast were the most informative attributes. The artificial neural network showed the most stable performance at both stages, followed by the support vector machine and k nearest neighbors, whereas the decision tree and Naive Bayes were less suitable. Confusion matrices and receiver operating characteristic curves indicated greater separability for omission and excess classes, with D1 standing out, and the patterns were consistent with the chemical analysis. Future work should include field validation, multiple seasons and genotypes, integration with spectral indices and multisensor data, application of model explainability techniques, and assessment of latency and scalability in operational scenarios. Full article
Show Figures

Figure 1

17 pages, 2969 KB  
Article
Multi-Domain CoP Feature Analysis of Functional Mobility for Parkinson’s Disease Detection Using Wearable Pressure Insoles
by Thathsara Nanayakkara, H. M. K. K. M. B. Herath, Hadi Sedigh Malekroodi, Nuwan Madusanka, Myunggi Yi and Byeong-il Lee
Sensors 2025, 25(18), 5859; https://doi.org/10.3390/s25185859 - 19 Sep 2025
Viewed by 318
Abstract
Parkinson’s disease (PD) impairs balance and gait through neuromotor dysfunction, yet conventional assessments often overlook subtle postural deficits during dynamic tasks. This study evaluated the diagnostic utility of center-of-pressure (CoP) features captured by pressure-sensing insoles during the Timed Up and Go (TUG) test. [...] Read more.
Parkinson’s disease (PD) impairs balance and gait through neuromotor dysfunction, yet conventional assessments often overlook subtle postural deficits during dynamic tasks. This study evaluated the diagnostic utility of center-of-pressure (CoP) features captured by pressure-sensing insoles during the Timed Up and Go (TUG) test. Using 39 PD and 38 control participants from the recently released open-access WearGait-PD dataset, the authors extracted 144 CoP features spanning positional, dynamic, frequency, and stochastic domains, including per-foot averages and asymmetry indices. Two scenarios were analyzed: the complete TUG and its 3 m walking segment. Model development followed a fixed protocol with a single participant-level 80/20 split; sequential forward selection with five-fold cross-validation optimized the number of features within the training set. Five classifiers were evaluated: SVM-RBF, logistic regression (LR), random forest (RF), k-nearest neighbors (k-NN), and Gaussian naïve Bayes (NB). LR performed best on the held-out test set (accuracy = 0.875, precision = 1.000, recall = 0.750, F1 = 0.857, ROC-AUC = 0.921) using a 23-feature subset. RF and SVM-RBF each achieved 0.812 accuracy. In contrast, applying the identical pipeline to the 3 m walking segment yielded lower performance (best model: k-NN, accuracy = 0.688, F1 = 0.615, ROC–AUC = 0.734), indicating that the multi-phase TUG task captures PD-related balance deficits more effectively than straight walking. All four feature families contributed to classification performance. Dynamic and frequency-domain descriptors, often appearing in both average and asymmetry form, were most consistently selected. These features provided robust magnitude indicators and offered complementary insights into reduced control complexity in PD. Full article
(This article belongs to the Section Wearables)
Show Figures

Figure 1

17 pages, 6828 KB  
Article
Precision Mapping of Fodder Maize Cultivation in Peri-Urban Areas Using Machine Learning and Google Earth Engine
by Sasikarn Plaiklang, Pharkpoom Meengoen, Wittaya Montre and Supattra Puttinaovarat
AgriEngineering 2025, 7(9), 302; https://doi.org/10.3390/agriengineering7090302 - 16 Sep 2025
Viewed by 326
Abstract
Fodder maize constitutes a key economic crop in Thailand, particularly in the northeastern region, where it significantly contributes to livestock feed production and local economic development. Nevertheless, the planning and management of cultivation areas remain a major challenge, especially in urban and peri-urban [...] Read more.
Fodder maize constitutes a key economic crop in Thailand, particularly in the northeastern region, where it significantly contributes to livestock feed production and local economic development. Nevertheless, the planning and management of cultivation areas remain a major challenge, especially in urban and peri-urban agricultural zones, due to the limited availability of spatial data and suitable analytical frameworks. These difficulties are exacerbated in urban settings, where the complexity of land use patterns and high spectral similarity among land cover types hinder accurate classification. The Google Earth Engine (GEE) platform provides an efficient and scalable solution for geospatial data processing, enabling rapid land use classification and spatiotemporal analysis. This study aims to enhance the classification accuracy of fodder maize cultivation areas in Mueang District, Nakhon Ratchasima Province, Thailand—an area characterized by a heterogeneous mix of urban development and agricultural land use. The research integrates GEE with four machine learning algorithms: Random Forest (RF), Support Vector Machine (SVM), Naïve Bayes (NB), and Classification and Regression Trees (CART). Eleven datasets were developed using Sentinel-2 imagery and a combination of biophysical variables, including elevation, slope, Normalized Difference Vegetation Index (NDVI), Normalized Difference Built-up Index (NDBI), and Modified Normalized Difference Water Index (MNDWI), to classify land use into six categories: fodder maize cultivation, urban and built-up areas, forest, water bodies, paddy fields, and other field crops. Among the 44 classification scenarios evaluated, the highest performance was achieved using Dataset 11—which integrated all spectral and biophysical variables—with the SVM classifier. This model attained an overall accuracy of 97.41% and a Kappa coefficient of 96.97%. Specifically, fodder maize was classified with 100% accuracy in both Producer’s and User’s metrics, as well as a Conditional Kappa of 100%. These findings demonstrate the effectiveness of integrating GEE with machine learning techniques for precise agricultural land classification. This approach also facilitates timely monitoring of land use changes and supports sustainable land management through informed planning, optimized resource allocation, and mitigation of land degradation in urban and peri-urban agricultural landscapes. Full article
(This article belongs to the Section Remote Sensing in Agriculture)
Show Figures

Figure 1

18 pages, 3374 KB  
Article
Evaluation of Apical Closure in Panoramic Radiographs Using Vision Transformer Architectures ViT-Based Apical Closure Classification
by Sümeyye Coşgun Baybars, Merve Daldal, Merve Parlak Baydoğan and Seda Arslan Tuncer
Diagnostics 2025, 15(18), 2350; https://doi.org/10.3390/diagnostics15182350 - 16 Sep 2025
Viewed by 276
Abstract
Objective: To evaluate the performance of vision transformer (ViT)-based deep learning models in the classification of open apex on panoramic radiographs (orthopantomograms (OPGs)) and compare their diagnostic accuracy with conventional convolutional neural network (CNN) architectures. Materials and Methods: OPGs were retrospectively [...] Read more.
Objective: To evaluate the performance of vision transformer (ViT)-based deep learning models in the classification of open apex on panoramic radiographs (orthopantomograms (OPGs)) and compare their diagnostic accuracy with conventional convolutional neural network (CNN) architectures. Materials and Methods: OPGs were retrospectively collected and labeled by two observers based on apex closure status. Two ViT models (Base Patch16 and Patch32) and three CNN models (ResNet50, VGG19, and EfficientNetB0) were evaluated using eight classifiers (support vector machine (SVM), random forest (RF), XGBoost, logistic regression (LR), K-nearest neighbors (KNN), naïve Bayes (NB), decision tree (DT), and multi-layer perceptron (MLP)). Performance metrics (accuracy, precision, recall, F1 score, and area under the curve (AUC)) were computed. Results: ViT Base Patch16 384 with MLP achieved the highest accuracy (0.8462 ± 0.0330) and AUC (0.914 ± 0.032). Although CNN models like EfficientNetB0 + MLP performed competitively (0.8334 ± 0.0479 accuracy), ViT models demonstrated more balanced and robust performance. Conclusions: ViT models outperformed CNNs in classifying open apex, suggesting their integration into dental radiologic decision support systems. Future studies should focus on multi-center and multimodal data to improve generalizability. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

11 pages, 833 KB  
Proceeding Paper
Heart Failure Prediction Through a Comparative Study of Machine Learning and Deep Learning Models
by Mohid Qadeer, Rizwan Ayaz and Muhammad Ikhsan Thohir
Eng. Proc. 2025, 107(1), 61; https://doi.org/10.3390/engproc2025107061 - 4 Sep 2025
Viewed by 4533
Abstract
The heart is essential to human life, so it is important to protect it and understand any kind of damage it can have. All the diseases related to hearts leads to heart failure. To help address this, a tool for predicting survival is [...] Read more.
The heart is essential to human life, so it is important to protect it and understand any kind of damage it can have. All the diseases related to hearts leads to heart failure. To help address this, a tool for predicting survival is needed. This study explores the use of several classification models for forecasting heart failure outcomes using the Heart Failure Clinical Records dataset. The outcome contrasts a deep learning (DL) model known as the Convolutional Neural Network (CNN) with many machine learning models, including Random Forest (RF), K-Nearest Neighbors (KNN), Decision Tree (DT), and Naïve Bayes (NB). Various data processing techniques, like standard scaling and Synthetic Minority Oversampling Technique (SMOTE), are used to improve prediction accuracy. The CNN model performs best by achieving 99%. In comparison, the best-performing ML model, Naïve Bayes, reaches 92.57%. This shows that deep learning provides better predictions of heart failure, making it a useful tool for early detection and better patient care. Full article
Show Figures

Figure 1

29 pages, 434 KB  
Article
Comparative Analysis of Natural Language Processing Techniques in the Classification of Press Articles
by Kacper Piasta and Rafał Kotas
Appl. Sci. 2025, 15(17), 9559; https://doi.org/10.3390/app15179559 - 30 Aug 2025
Viewed by 471
Abstract
The study undertook a comprehensive review and comparative analysis of natural language processing techniques for news article classification, with a particular focus on Java language libraries. The dataset comprised an excess of 200,000 items of news metadata sourced from The Huffington Post. The [...] Read more.
The study undertook a comprehensive review and comparative analysis of natural language processing techniques for news article classification, with a particular focus on Java language libraries. The dataset comprised an excess of 200,000 items of news metadata sourced from The Huffington Post. The traditional algorithms based on mathematical statistics and deep machine learning were evaluated. The libraries chosen for tests were Apache OpenNLP, Stanford CoreNLP, Waikato Weka, and the Huggingface ecosystem with the Pytorch backend. The efficacy of the trained models in forecasting specific topics was evaluated, and diverse methodologies for the feature extraction and analysis of word-vector representations were explored. The study considered aspects such as hardware resource management, implementation simplicity, learning time, and the quality of the resulting model in terms of detection, and it examined a range of techniques for attribute selection, feature filtering, vector representation, and the handling of imbalanced datasets. Advanced techniques for word selection and named entity recognition were employed. The study compared different models and configurations in terms of their performance and the resources they consumed. Furthermore, it addressed the difficulties encountered when processing lengthy texts with transformer neural networks, and it presented potential solutions such as sequence truncation and segment analysis. The elevated computational cost inherent to Java-based languages may present challenges in machine learning tasks. OpenNLP model achieved 84% accuracy, Weka and CoreNLP attained 86% and 88%, respectively, and DistilBERT emerged as the top performer, with an accuracy rate of 92%. Deep learning models demonstrated superior performance, training time, and ease of implementation compared to conventional statistical algorithms. Full article
(This article belongs to the Special Issue Natural Language Processing (NLP) and Applications—2nd Edition)
Show Figures

Figure 1

26 pages, 3398 KB  
Article
Hybrid Mamba and Attention-Enhanced Bi-LSTM for Obesity Classification and Key Determinant Identification
by Chongyang Fu, Mohd Shahril Nizam Bin Shaharom and Syed Kamaruzaman Bin Syed Ali
Electronics 2025, 14(17), 3445; https://doi.org/10.3390/electronics14173445 - 29 Aug 2025
Viewed by 558
Abstract
Obesity is a major public health challenge linked to increased risks of chronic diseases. Effective prevention and intervention strategies require accurate classification and identification of key determinants. This study aims to develop a robust deep learning framework to enhance the accuracy and interpretability [...] Read more.
Obesity is a major public health challenge linked to increased risks of chronic diseases. Effective prevention and intervention strategies require accurate classification and identification of key determinants. This study aims to develop a robust deep learning framework to enhance the accuracy and interpretability of obesity classification using comprehensive datasets, and to compare its performance with both traditional and state-of-the-art deep learning models. We propose a hybrid deep learning framework that combines an improved Mamba model with an attention-enhanced bidirectional LSTM (ABi-LSTM). The framework utilizes the Obesity and CDC datasets. A feature tokenizer is integrated into the Mamba model to improve scalability and representation learning. Channel-independent processing is employed to prevent overfitting through independent feature analysis. The ABi-LSTM component is used to capture complex temporal dependencies in the data, thereby enhancing classification performance. The proposed framework achieved an accuracy of 93.42%, surpassing existing methods such as ID3 (91.87%), J48 (89.98%), Naïve Bayes (90.31%), Bayesian Network (89.23%), as well as deep learning-based approaches such as VAE (92.12%) and LightCNN (92.50%). Additionally, the model improved sensitivity to 91.11% and specificity to 92.34%. The hybrid model demonstrates superior performance in obesity classification and determinant identification compared to both traditional and advanced deep learning methods. These results underscore the potential of deep learning in enabling data-driven personalized healthcare and targeted obesity interventions. Full article
(This article belongs to the Special Issue Knowledge Representation and Reasoning in Artificial Intelligence)
Show Figures

Figure 1

31 pages, 1856 KB  
Article
Optimizing Chatbots to Improve Customer Experience and Satisfaction: Research on Personalization, Empathy, and Feedback Analysis
by Shimon Uzan, David Freud and Amir Elalouf
Appl. Sci. 2025, 15(17), 9439; https://doi.org/10.3390/app15179439 - 28 Aug 2025
Viewed by 1611
Abstract
This study addresses the ongoing challenge of optimizing chatbot interactions to significantly enhance customer experience and satisfaction through personalized, empathetic responses. Using advanced NLP tools and strong statistical methodologies, we developed and evaluated a multi-layered analytical framework to accurately identify user intents, assess [...] Read more.
This study addresses the ongoing challenge of optimizing chatbot interactions to significantly enhance customer experience and satisfaction through personalized, empathetic responses. Using advanced NLP tools and strong statistical methodologies, we developed and evaluated a multi-layered analytical framework to accurately identify user intents, assess customer feedback, and generate emotionally intelligent interactions. With over 270,000 customer chatbot interaction records in our dataset, we employed spaCy-based NER and clustering algorithms (HDBSCAN and K-Means) to categorize customer queries precisely. Text classification was performed using random forest, logistic regression, and SVM, achieving near-perfect accuracy. Sentiment analysis was conducted using VADER, Naive Bayes, and TextBlob, complemented by semantic analysis via LDA. Statistical tests, including Chi-square, Kruskal–Wallis, Dunn’s test, ANOVA, and logistic regression, confirmed the significant impact of tailored, empathetic response strategies on customer satisfaction. Correlation analysis indicated that traditional measures like sentiment polarity and text length insufficiently capture customer satisfaction nuances. The results underscore the critical role of context-specific adjustments and emotional responsiveness, paving the way for future research into chatbot personalization and customer-centric system optimization. Full article
Show Figures

Figure 1

Back to TopTop