Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (4,506)

Search Parameters:
Keywords = XGBoost model

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
33 pages, 6401 KB  
Article
An Explainable Machine Learning Framework for Flood Damage Mapping Using Remote Sensing and Ground-Based Data: Application to the Basilicata Ionian Coast (Italy)
by Silvano Fortunato Dal Sasso, Maríca Rondinone, Htay Htay Aung and Vito Telesca
Remote Sens. 2026, 18(8), 1257; https://doi.org/10.3390/rs18081257 (registering DOI) - 21 Apr 2026
Abstract
Flood damage assessment remains challenging, as conventional flood risk management mainly relies on hydraulic hazard maps that do not explicitly reproduce observed damage patterns. Recent advances in remote sensing and machine learning (ML) enable the integration of environmental and socio-economic data with historical [...] Read more.
Flood damage assessment remains challenging, as conventional flood risk management mainly relies on hydraulic hazard maps that do not explicitly reproduce observed damage patterns. Recent advances in remote sensing and machine learning (ML) enable the integration of environmental and socio-economic data with historical impact information to improve flood damage modeling. This study proposes an explainable machine learning framework for flood damage susceptibility mapping, using observed institutional damage records from the 2011 and 2013 flood events combined with 17 geospatial flood risk factors (FRFs) representing hazard, exposure, and vulnerability. This approach enables the capture of non-linear relationships between flood damage and FRFs. For comparison purposes, the same framework was also applied using hydraulically modeled flood extents corresponding to return periods of 30, 200, and 500 years. The framework was tested along the Basilicata Ionian coast in southern Italy, a Mediterranean region characterized by complex geomorphology, intense rainfall events, and recurrent flood impacts. An eXtreme Gradient Boosting (XGBoost) model was trained using 17 FRFs related to hazard, exposure, and vulnerability at a spatial resolution of 20 m. The model achieved high performance with an accuracy of 0.988, an F1-score for the minority class of 0.860, and an ROC-AUC (test) of 0.996. High to very high flood damage probability was predicted in approximately 4.1% of the study area, mainly in low-lying floodplains near river corridors and infrastructure. SHAP-based explainability analysis revealed that damage susceptibility was predominantly driven by hazard and exposure factors: Drainage density (17.10%), Railway distance (16.33%), and Elevation (15.42%), extreme precipitation (Max rainfall, 10.66%) and Street distance (7.51%), with socio-economic vulnerability contributing less than 4%. The observed damage target exhibited clear threshold-like patterns (e.g., sharp risk increases below ~25/35 m elevation or within ~150/200 m of road infrastructure), contrasting with the smoother, continuous gradients produced by hydraulic scenarios. This analysis identified the most influential predictors and their response ranges. The proposed framework complements hydraulic hazard mapping by explicitly modeling observed flood damage, supporting flood risk assessment in flood-prone coastal regions. Full article
Show Figures

Figure 1

21 pages, 1071 KB  
Article
A Tiered Classification Framework for Detecting and Diagnosing Man-in-the-Middle Attacks in Smart Grid Protocols
by Hassan N. Noura, Zaid Allal, Ola Salman and Khaled Chahine
Future Internet 2026, 18(4), 220; https://doi.org/10.3390/fi18040220 (registering DOI) - 21 Apr 2026
Abstract
The increasing reliance on smart grid communication systems has significantly raised the demand for robust cybersecurity measures to defend against advanced threats. This paper proposes a two-tier classification framework to enhance the detection and diagnosis of man-in-the-middle attacks within smart grid communication protocols. [...] Read more.
The increasing reliance on smart grid communication systems has significantly raised the demand for robust cybersecurity measures to defend against advanced threats. This paper proposes a two-tier classification framework to enhance the detection and diagnosis of man-in-the-middle attacks within smart grid communication protocols. Initially, the model detects the presence of an attack and then identifies the specific type of man-in-the-middle attack through subsequent inferences. To achieve this, the “Man-in-the-Middle Attacks Targeting Modbus TCP/IP and MMS Protocols in the Smart Grid” dataset was carefully preprocessed and analyzed to better understand the underlying hidden characteristics. This understanding, coupled with existing works on fault detection and diagnosis, facilitated the engineering of new features from the original dataset. Four classifiers were employed in each tier: Random Forest, XGBoost, LightGBM, and CatBoost. The first tier exhibited exceptional performance, with the CatBoost framework achieving 99.6% accuracy. The second tier also demonstrated strong results, with the same model achieving 99.1% accuracy. Systematic model explainability was conducted using SHapley Additive exPlanations for both tiers and revealed that the highest accuracy was achieved using five features for the first and six for the second. The average inference time was approximately 4.76 milliseconds. The proposed framework is accurate, fast, interpretable, lightweight, and well-optimized for direct implementation in smart grid systems to detect and diagnose man-in-the-middle attacks. Full article
(This article belongs to the Special Issue Artificial Intelligence in Smart Grids)
19 pages, 21540 KB  
Article
XGBoost for Multi-Fault Diagnosis and Prediction in Permanent Magnet Synchronous Machines
by Yacine Maanani, Chuan Pham, Qingsong Wang, Kim Khoa Nguyen and Kamal Al-Haddad
Electronics 2026, 15(8), 1759; https://doi.org/10.3390/electronics15081759 (registering DOI) - 21 Apr 2026
Abstract
In this study, we propose a data-driven diagnostic system that uses Extreme Gradient Boosting (XGBoost) to detect, classify, and assess the severity of multiple faults in permanent magnet synchronous motors (PMSMs). The three main fault categories that are the focus of the suggested [...] Read more.
In this study, we propose a data-driven diagnostic system that uses Extreme Gradient Boosting (XGBoost) to detect, classify, and assess the severity of multiple faults in permanent magnet synchronous motors (PMSMs). The three main fault categories that are the focus of the suggested method are inter-turn short-circuit (ITSC) faults, stator open-circuit faults, and permanent magnet demagnetization. To capture fault-specific characteristics and their development with severity, discriminative electrical features are retrieved from stator currents, flux linkage, and dq-axis values. Next, using the chosen electrical indications, an aggregated diagnostic index is created to facilitate defect diagnosis and severity quantification in a single learning process. The XGBoost-based model has been shown to produce excellent diagnostic accuracy and robust separation between various fault causes via extensive assessment. It also maintains dependable performance under previously unknown operating or fault situations. These findings show that an XGBoost-only approach offers a scalable and efficient way to monitor advanced PMSM conditions in industrial and safety-critical applications. Full article
(This article belongs to the Special Issue Design and Control of Drives and Electrical Machines)
Show Figures

Figure 1

27 pages, 2004 KB  
Review
Machine Learning in Personalized Medication Regimen Design for the Geriatric Population: Integrating Pharmacokinetic and Pharmacodynamic Modeling with Clinical Decision-Making
by Ahmad R. Alsayed, Mohanad Al-Darraji, Mohannad Al-Qaiseiah, Anas Samara and Mustafa Al-Bayati
Technologies 2026, 14(4), 241; https://doi.org/10.3390/technologies14040241 (registering DOI) - 21 Apr 2026
Abstract
Geriatric pharmacotherapy is usually challenged by physiological senescence. For instance, progressive declines in organ function and alterations in body composition can complicate drug disposition. However, conventional pharmacometrics models commonly have limited capacity to map these high-dimensional, nonlinear relationships. In this review, we are [...] Read more.
Geriatric pharmacotherapy is usually challenged by physiological senescence. For instance, progressive declines in organ function and alterations in body composition can complicate drug disposition. However, conventional pharmacometrics models commonly have limited capacity to map these high-dimensional, nonlinear relationships. In this review, we are examining the recent shift toward integrating machine learning (ML) with mechanistic pharmacokinetic (PK)/pharmacodynamic (PD) models to improve the accuracy and precision of dosing. Machine learning approaches like Random Forest and XGBoost consistently provided more accurate exposure predictions and significantly more efficient computational workflows than conventional methods. Nevertheless, concerns such as “black box” transparency and the potential of algorithmic bias toward specific patient demographics are challenging. It is important to incorporate explainability tools like SHAP, and adopting FAIR data principles is crucial for achieving professional trust and ensuring site-specific generalizability. Full article
(This article belongs to the Special Issue Technological Advances in Science, Medicine, and Engineering 2025)
Show Figures

Figure 1

24 pages, 27840 KB  
Article
Decoding Public Perception of Brownfield-Transformed Urban Parks: An Interpretable Machine Learning Framework Integrating XGBoost–SHAP
by Xiaomin Wang, Xiangru Chen, Chao Yang, Zhongyuan Zhao and Xinling Chen
Buildings 2026, 16(8), 1632; https://doi.org/10.3390/buildings16081632 (registering DOI) - 21 Apr 2026
Abstract
Brownfield-transformed urban parks, particularly those derived from industrial heritage, play a critical role in both cultural preservation and public-space provision. However, existing studies often rely on linear models and general urban contexts, limiting their ability to capture nonlinear, interaction-driven perception and translate analytical [...] Read more.
Brownfield-transformed urban parks, particularly those derived from industrial heritage, play a critical role in both cultural preservation and public-space provision. However, existing studies often rely on linear models and general urban contexts, limiting their ability to capture nonlinear, interaction-driven perception and translate analytical results into design-oriented insights. To address this gap, this study develops an interpretable data-driven framework integrating NLP (natural language processing) with explainable machine learning. Using social media reviews from Shougang Park in Beijing, built environmental elements are identified and structured into four dimensions—Accessibility, Safety, Comfort, and Enjoyment. An XGBoost model combined with SHAP analysis is employed to examine variable importance, nonlinear relationships, and interaction effects. The results reveal that visitor satisfaction is governed by heterogeneous and nonlinear relationships rather than independent additive effects. Several variables exhibit threshold-like, diminishing, and inverted-U-shaped patterns, indicating sensitivity to intensity ranges. More importantly, spatial perception emerges from the nonlinear coupling of multiple elements, forming four representative interaction types: compensatory, inverted-U-shaped, context-dependent, and threshold-like relationships. Key interactions are concentrated around industrial landscape, leisure activities, and supporting facilities. Building on these findings, the study translates interactions into design-oriented strategies, emphasizing synergistic configuration, functional balance, moderated development intensity, and context- sensitive programming. By linking interpretable machine learning with spatial design, this research advances an interaction-oriented paradigm and provides a transferable framework for satisfaction-informed evaluation and optimization of brownfields. Full article
(This article belongs to the Section Architectural Design, Urban Science, and Real Estate)
Show Figures

Figure 1

18 pages, 1499 KB  
Article
Toward Personalized Rotator Cuff Physical Therapy Dosage Using a Machine Learning-Based Pilot Study with EMG
by AmirHossein MajidiRad, Iram Azam, Japp Adhikari and Mehrnoosh Damircheli
Bioengineering 2026, 13(4), 483; https://doi.org/10.3390/bioengineering13040483 (registering DOI) - 21 Apr 2026
Abstract
Rotator cuff injuries are among the most common musculoskeletal conditions that affect shoulder function and can ultimately impact quality of life. While physical therapy is essential in the care of rotator cuff injuries, the ideal dose of therapeutic exercises continues to be a [...] Read more.
Rotator cuff injuries are among the most common musculoskeletal conditions that affect shoulder function and can ultimately impact quality of life. While physical therapy is essential in the care of rotator cuff injuries, the ideal dose of therapeutic exercises continues to be a significant clinical dilemma because of the generalized nature of rehabilitation protocols. This pilot study proposes a machine learning approach to personalize rehabilitation using surface electromyography (sEMG) data collected from eight healthy individuals by testing four key shoulder movements: scaption, internal rotation, external rotation, and external rotation at 90° abduction. In this research, the XGBoost algorithm was used to model muscle activation patterns by achieving a high predictive accuracy (R2 = 0.5325; MSE = 0.0084 μV2). Because sEMG reliably measures superficial muscle activity, a linear programming model was used to divide a 60 min therapy session in a way that increases activation of superficial muscles (such as deltoid and trapezius) while reducing strain on deep muscles (such as supraspinatus and infraspinatus). Three optimization scenarios were tested by reflecting a different clinical goal: prioritizing superficial muscles, minimizing deep muscle strain, or balancing both. Optimized time allocations assigned more time to external rotation at 90° abduction and scaption. This research demonstrates the potential for data-driven methods to transform rotator cuff rehabilitation through personalized and evidence-based treatment plans. The results enhance clinical practice by enabling adaptive rehabilitation planning and show that machine learning can support decision-making in complex muscle activation analysis with strong performance and low latency. Full article
(This article belongs to the Special Issue Advances in Physical Therapy and Rehabilitation, 2nd Edition)
Show Figures

Figure 1

19 pages, 1412 KB  
Article
A Micro-Manifold Identity-Preserving Spatiotemporal Graph Neural Network for Financial Risk Early Warning
by Jin Kuang, Fusheng Chen, Te Guo and Chiawei Chu
Mathematics 2026, 14(8), 1388; https://doi.org/10.3390/math14081388 (registering DOI) - 21 Apr 2026
Abstract
Traditional financial early warning models often rely on the independent and identically distributed (IID) assumption, failing to adequately capture cross-sectional spatial contagion effects and temporal dynamic mutations, and are susceptible to the over-smoothing problem when processing highly imbalanced graph networks. To address these [...] Read more.
Traditional financial early warning models often rely on the independent and identically distributed (IID) assumption, failing to adequately capture cross-sectional spatial contagion effects and temporal dynamic mutations, and are susceptible to the over-smoothing problem when processing highly imbalanced graph networks. To address these limitations, this study proposes a micro-manifold-based identity-preserving spatiotemporal graph neural network framework (Micro-STAGNN). In the spatial dimension, an identity-preserving graph convolutional operator (IP-GCN) is constructed. By hard-coding a self-preservation coefficient (λ=0.8), it quantifies peer risk spillover while mitigating feature dilution, ensuring the transmission of heterogeneous default signals. In the temporal dimension, Long Short-Term Memory networks are cascaded with a temporal attention mechanism to capture the nonlinear temporal inflection points that trigger financial distress. The empirical study utilizes a sample of China’s A-share market from 2015 to 2025, evaluating the model using an Out-of-Time Validation protocol and Focal Loss. Results indicate that under a highly imbalanced distribution with a positive-to-negative sample ratio of approximately 1:50, Micro-STAGNN achieves an OOT ROC-AUC of 0.9095, a minority class default recall of 89%, and reduces the missed detection rate to 11%, outperforming traditional nonlinear cross-sectional models such as XGBoost. Furthermore, temporal attention weights provide explainable support for the early warning results. Full article
(This article belongs to the Special Issue Mathematical Methods for Economics, Finance and Actuarial Sciences)
Show Figures

Figure 1

31 pages, 10033 KB  
Article
Prediction Model for the Local Bearing Capacity of Stirrup-Confined Concrete Based on the PSO-BP Neural Network
by Tianming Miao, Junwu Dai, Tao Jiang, Yongjian Ding, Ruchen Qie, Yingqi Liu and Ying Zhou
Infrastructures 2026, 11(4), 143; https://doi.org/10.3390/infrastructures11040143 - 20 Apr 2026
Abstract
The calculation for the local bearing capacity of stirrup-confined concrete is an important issue in structural design. Due to the coupling effects of multiple factors, there is no unified calculation method recognized by scholars. The improved backpropagation neural network model based on the [...] Read more.
The calculation for the local bearing capacity of stirrup-confined concrete is an important issue in structural design. Due to the coupling effects of multiple factors, there is no unified calculation method recognized by scholars. The improved backpropagation neural network model based on the particle swarm optimization algorithm (PSO-BPNN) is used in this research to conduct a systematic analysis. The results of 40 stirrup-confined concrete specimens from the tests conducted by ourselves and an additional 92 similar test data points from references were combined; the calculation efficiency and accuracy of the PSO-BPNN model were verified. Compared with the BPNN model, the training iterations of the PSO-BPNN model were reduced by 74.23% with the condition of same training effect. The mean squared error (MSE) is reduced by 33.9%, and the coefficient of determination (R2) is increased by 5.5% with the condition of the same number training iterations. In addition, compared with the calculation stability and accuracy of Random Forest Regression (RFR), Support Vector Regression (SVR), and Extreme Gradient Boosting (XGBoost) models, the PSO-BPNN model also shows better results. Within the applicable range of the codes, the average ratio of the predicted values to the calculated values for GB50010-2010, MC2020 and ACI318-25 are 1.988, 1.719, and 5.387, respectively. A higher evaluation for the contribution of stirrup is considered in the MC2020 code; the predicted values of some specimens are lower than the calculated values when Acor/Al is less than 1.35. The brittleness effect is not adequately considered: the predicted values of some specimens are also lower than the calculated values with the active powder concrete (RPC) is used. The sensitivity ranking of the model with coupling effect for parameters is Al, Ab, fc,k, s, d, dcor, and fy,k. It is slightly different from the sensitivity ranking obtained by analyzing individual parameters, but the calculation logic is consistent. The research results can provide a theoretical basis for practical engineering. Full article
(This article belongs to the Section Infrastructures and Structural Engineering)
17 pages, 3376 KB  
Article
Design and Feasibility Assessment of a Compact Emergency Unit in Rural and Remote Areas: A Multicenter Analysis of KTAS-Based Triage Data
by Kyungman Cha, Youngjin Kim, Sohee Lee, Jaekwang Shin and Jee Yong Lim
Healthcare 2026, 14(8), 1099; https://doi.org/10.3390/healthcare14081099 - 20 Apr 2026
Abstract
Background/Objectives: Emergency department (ED) overcrowding burdens rural and remote areas where geographic isolation limits timely care. The Compact Emergency Unit (CEU)—a 24 h facility with remote physician oversight—has been proposed but lacks an empirical foundation. We aimed to (1) quantify CEU-eligible (final KTAS [...] Read more.
Background/Objectives: Emergency department (ED) overcrowding burdens rural and remote areas where geographic isolation limits timely care. The Compact Emergency Unit (CEU)—a 24 h facility with remote physician oversight—has been proposed but lacks an empirical foundation. We aimed to (1) quantify CEU-eligible (final KTAS 4–5) patients in a multicenter ED cohort; (2) compare their operational metrics with non-eligible patients; (3) characterize hourly demand for facility planning; and (4) develop machine-learning models for non-discharge prediction within this low-acuity stratum. Methods: Retrospective analysis of 12 months (January–December 2025) of NEDIS data from two Korean university-affiliated EDs. Effect sizes (Cliff’s δ, Cramér’s V) were reported alongside p-values. Three classifiers (logistic regression, random forest, and XGBoost) were developed with patient-level cross-validation, comparing a 16-feature baseline and a 22-feature set augmented with arrival vital signs. Calibration and decision curve analysis were performed. Results: Of 34,544 valid triage visits (27,743 unique patients), 9871 (28.6%) were CEU-eligible. They had shorter LOS (92 vs. 171 min; Cliff’s δ = −0.51), 98.8% symptomatic home discharge, and a median of 0 specialty consultations. Nighttime visits comprised 43.7% of CEU-eligible encounters, peaking at 20:00 (1.76 visits/h/day). The non-discharge rate was 1.20% (118/9871). The vital-augmented random forest reached AUROC 0.794 (95% CI 0.758–0.829); XGBoost calibration was near-perfect (ECE 0.020). A combined ML-or-vital-sign screening rule raised non-discharge sensitivity to 94.1%. Conclusions: Approximately 29% of ED visits could be CEU-suitable. Single-modality machine learning is insufficient for safety-critical triage, but a layered ML-plus-vitals screening approach achieves operationally relevant sensitivity. Prospective implementation studies are required before clinical deployment. Full article
Show Figures

Figure 1

44 pages, 2312 KB  
Article
Classification Model of Emotional Tone in Hate Speech and Its Relationship with Inequality and Gender Stereotypes, Using NLP and Machine Learning Algorithms
by Aymé Escobar Díaz, Ricardo Rivadeneira, Walter Fuertes and Washington Loza
Future Internet 2026, 18(4), 218; https://doi.org/10.3390/fi18040218 - 20 Apr 2026
Abstract
Hate speech on social media reproduces norms of inequality and gender stereotypes, disproportionately affecting women. This study proposes a hybrid approach that integrates emotional tone classification with explicit hostility detection to strengthen preventive moderation. We constructed a corpus from three open data sets [...] Read more.
Hate speech on social media reproduces norms of inequality and gender stereotypes, disproportionately affecting women. This study proposes a hybrid approach that integrates emotional tone classification with explicit hostility detection to strengthen preventive moderation. We constructed a corpus from three open data sets (1,236,371 records; 1,003,991 after ETL) and represented the text using TF-IDF and contextual RoBERTa embeddings. We trained individual models (RoBERTa fine-tuned, Random Forest, and XGBoost) and a stacking metamodel (Gradient Boosting) that combines their probabilities. On the test set, the ensemble outperformed the base classifiers, achieving accuracy of 0.93 in hate detection and 0.90 in emotion classification, with an AUC of 0.98 for emotion classification. We implemented a RESTful API and a web client to validate the moderation flow before publication, along with an administration panel for auditing. Performance tests in a prototype deployment (Google Colab exposed through an Ngrok tunnel) provided proof-of-concept validation, revealing concurrency limitations from around 300 users due to infrastructure constraints. In general, the results indicate that incorporating emotional tone analysis improves the model’s ability to identify implicit hostility and offers a practical way to promote safer digital environments. The probabilistic outputs produced by the ensemble model were subsequently analyzed using the Bayesian Calibration and Optimal Design under Asymmetric Risk (BACON-AR) framework, which serves as a mathematical post hoc decision layer for evaluating classification behaviour under unequal error costs. Rather than modifying the trained architecture or improving its predictive performance, the framework identifies a cost-sensitive operating threshold that minimizes the total expected risk under the selected asymmetric cost configuration. The experiments were conducted using an English-language data set; therefore, the findings of this study are limited to hate speech detection in English. Full article
(This article belongs to the Section Techno-Social Smart Systems)
27 pages, 1901 KB  
Article
Comparative Forecasting and Misclassification Analysis Using Health Survey Data
by Ermioni Traka, George Papageorgiou, Georgios Mantzavinis and Christos Tjortjis
AI 2026, 7(4), 148; https://doi.org/10.3390/ai7040148 - 20 Apr 2026
Abstract
Background: Accurate mortality prediction remains a major challenge in public health due to the complex interactions among demographic, socioeconomic, behavioral, and medical factors. This problem is particularly relevant for identifying high-risk groups and improving preventive healthcare strategies. While existing studies demonstrate strong predictive [...] Read more.
Background: Accurate mortality prediction remains a major challenge in public health due to the complex interactions among demographic, socioeconomic, behavioral, and medical factors. This problem is particularly relevant for identifying high-risk groups and improving preventive healthcare strategies. While existing studies demonstrate strong predictive performance, they mainly rely on clinically structured data and focus on model performance. Challenges such as misclassification and atypical cases remain less explored. Methods: Using the Integrated Public Use Microdata Series National Health Interview Survey (IPUMS-NHIS) 2010 and 2015 datasets (193,765 records, 104 features), this study investigates mortality prediction through comparative Machine Learning. Data preprocessing included feature engineering, categorical encoding, and removal of missing entries. Class imbalance was addressed using SMOTE and SMOTE-ENN resampling, followed by hyperparameter tuning. Three models—Logistic Regression, Random Forest, and XGBoost—were trained to classify mortality, with recall prioritized to ensure accurate identification of deceased cases. Results: Results showed that XGBoost achieved the best performance (Recall = 69%, F1 = 0.39, AUC = 0.92), outperforming other models in balancing sensitivity and specificity. Feature importance and permutation analyses highlighted age, employment status, self-reported health, and lifestyle indicators as key predictors. Misclassification analysis combined with Isolation Forest revealed atypical profiles not captured by standard models. Conclusions: The findings underscore XGBoost’s effectiveness and demonstrate the value of integrating anomaly detection with classification to improve mortality prediction and inform public health planning. Full article
Show Figures

Figure 1

31 pages, 4910 KB  
Article
Comparative Evaluation of Machine Learning and Deep Learning Models for Tropical Cyclone Track and Intensity Forecasting in the North Atlantic Basin
by Henry A. Ogu, Liping Liu and Yuh-Lang Lin
Atmosphere 2026, 17(4), 418; https://doi.org/10.3390/atmos17040418 - 20 Apr 2026
Abstract
Accurate forecasts of tropical cyclone (TC) track and intensity with a sufficient lead time are critical for disaster preparedness and risk mitigation. Traditional numerical weather prediction models, while fundamental to operational forecasting, often exhibit systematic errors due to limitations in observations, physical parameterizations, [...] Read more.
Accurate forecasts of tropical cyclone (TC) track and intensity with a sufficient lead time are critical for disaster preparedness and risk mitigation. Traditional numerical weather prediction models, while fundamental to operational forecasting, often exhibit systematic errors due to limitations in observations, physical parameterizations, and model resolution. In recent years, machine learning (ML) and deep learning (DL) approaches have emerged as promising data-driven alternatives for improving TC forecasts. This study presents a comparative evaluation of six ML and DL models—Random Forest (RF), Extreme Gradient Boosting (XGBoost), Light Gradient Boosting Machine (LightGBM), Categorical Boosting (CatBoost), Artificial Neural Network (ANN), and Convolutional Neural Network (CNN)—for forecasting TC track and intensity in the North Atlantic basin. The models are trained using the National Hurricane Center’s (NHC) HURDAT2 best-track dataset for storms from 1990 to 2019 and evaluated on an independent test set from the 2020 season. Model performance is compared across all models and benchmarked against the 2020 mean Decay-SHIFOR5 intensity error, CLIPER5 track errors, and the NHC official forecast (OFCL) errors. Forecast skill is assessed using mean absolute error (MAE) with 95% bootstrap confidence intervals and the coefficient of determination (R2) across lead times of 6, 12, 18, 24, 48, and 72 h. The results show that: (1) several ML and DL models achieve intensity forecast performance that is broadly comparable in magnitude to the 2020 mean OFCL benchmarks, with an average error reduction of 5–11% at the 24 h lead time; (2) among the ML models, XGBoost and CatBoost slightly outperform LightGBM and RF in accuracy, while LightGBM demonstrates the highest computational efficiency; and (3) among the DL models, CNNs outperform ANNs in predictive accuracy and intensity forecasting efficiency, while ANNs exhibit lower computational cost for track forecast. Bootstrap confidence intervals indicate relatively low variability in model errors, supporting the statistical stability of the results within the 2020 season. However, these results reflect within-season variability and do not necessarily generalize across different years or climatological conditions. Overall, the findings demonstrate the potential of ML/DL-based approaches to complement existing operational forecast systems and enhance TC track and intensity forecasting in the North Atlantic basin. Full article
(This article belongs to the Special Issue Machine Learning for Atmospheric and Remote Sensing Research)
35 pages, 2051 KB  
Article
Leakage-Controlled Horizon-Specific Model Selection for Daily Equity Forecasting: An Automated Multi-Model Pipeline
by Francisco Augusto Nuñez Perez, Francisco Javier Aguilar Mosqueda, Adrian Ramos Cuevas, Jaqueline Muñoz Beltran and Jose Cruz Nuñez Perez
Forecasting 2026, 8(2), 34; https://doi.org/10.3390/forecast8020034 - 20 Apr 2026
Abstract
Short-horizon equity forecasting remains challenging because daily prices are noisy, heavy-tailed, and subject to structural breaks and regime shifts. We develop a fully automated, reproducible, and leakage-controlled multi-model pipeline for daily forecasting with horizon-specific configuration selection. The task is formulated as predicting cumulative [...] Read more.
Short-horizon equity forecasting remains challenging because daily prices are noisy, heavy-tailed, and subject to structural breaks and regime shifts. We develop a fully automated, reproducible, and leakage-controlled multi-model pipeline for daily forecasting with horizon-specific configuration selection. The task is formulated as predicting cumulative H-day log-returns from OHLCV-derived information and converting them to implied price forecasts. All model families share a homologated design: causal feature construction, a strictly chronological split with an explicit purging rule to prevent label-window overlap for multi-day targets, training-only robustification (winsorization and adaptive clipping), and a unified metric suite computed consistently in return and price spaces. The framework benchmarks transparent baselines (zero- and mean-return), gradient-boosted trees (XGBoost), and deep temporal models (LSTM and CNN/TCN). Lookback length L{60,180,500} is selected via an internal walk-forward procedure on the pre-evaluation block, and final performance is reported on an external hold-out segment (last 15% of instances). Experiments on daily data for MT, DELL, and the S&P 500 index (through 3 February 2026) show that all families achieve similarly strong price-level fit at H=1, largely driven by persistence in the price process, while separation across families becomes more visible at H=5. However, predictive performance in return space remains weak, with R2 close to zero or negative, and Diebold–Mariano tests do not provide consistent evidence of statistical superiority over naive benchmarks. Under an operational rule that minimizes hold-out RMSE on the price scale, selected models are asset- and horizon-dependent, supporting horizon-wise selection rather than a single global architecture. Overall, the primary contribution lies in the proposed leakage-controlled evaluation and benchmarking framework rather than in demonstrating consistent predictive gains in financial time series forecasting. Full article
33 pages, 2947 KB  
Article
A Reproducible Hybrid Architecture of Fuzzy Logic and XGBoost for Explainable Tabular Classification of Territorial Vulnerability
by Aiman Akynbekova, Ayagoz Mukhanova, Raikhan Muratkhan, Lunara Diyarova, Saya Baigubenova, Gulden Murzabekova, Gulaim Orazymbetova, Aliya Satybaldieva and Zhanat Abdikadyr
Computers 2026, 15(4), 259; https://doi.org/10.3390/computers15040259 - 20 Apr 2026
Abstract
This study proposes a reproducible hybrid computational model for the explainable classification of territorial vulnerability using heterogeneous tabular data. The approach integrates fuzzy logic and extreme gradient boosting in a two-stage architecture that balances interpretability and predictive performance. First, a fuzzy transformation is [...] Read more.
This study proposes a reproducible hybrid computational model for the explainable classification of territorial vulnerability using heterogeneous tabular data. The approach integrates fuzzy logic and extreme gradient boosting in a two-stage architecture that balances interpretability and predictive performance. First, a fuzzy transformation is applied to construct interpretable risk and resilience indicators based on multi-source administrative indicators. The analytical dataset was formed by integrating 11 heterogeneous administrative sources into a single matrix of 166 territorial units and 76 features. The model was evaluated on a stratified 75/25 split of the training and test sets using the F1 score, ROC-AUC, precision, recall, and integrated quality criterion. Experimental results show that the proposed Fuzzy-XGBoost framework achieved an F1 score of 0.7333 on the test dataset, an ROC-AUC of 0.8291, and an Integrated Score of 0.768, outperforming the strongest baseline and improving recall in highly vulnerable areas. Furthermore, probabilistic threshold optimization identified an operating point at τ = 0.35, reducing the number of missed high-risk cases while maintaining acceptable specificity. The results demonstrate that fuzzy feature expansion combined with gradient boosting provides an efficient and interpretable solution for tabular risk classification and decision support problems under heterogeneity and uncertainty. Full article
21 pages, 9107 KB  
Article
Experimental and ML Modeling of Drying Shrinkage and Water Loss in Low-Heat Cement Concrete Under Extreme Plateau Curing
by Guohui Zhang, Zhipeng Yang, Rongheng Duan, Zhuang Yan and Gongfei Wang
Buildings 2026, 16(8), 1616; https://doi.org/10.3390/buildings16081616 - 20 Apr 2026
Abstract
To investigate concrete drying shrinkage in high-altitude environments, moisture evaporation and shrinkage rates were examined under combined curing regimes of four temperatures (40 °C, 20 °C, 0 °C, −10 °C) and three relative humidities (RH40%, RH60%, RH80%). Curing temperature and humidity primarily regulate [...] Read more.
To investigate concrete drying shrinkage in high-altitude environments, moisture evaporation and shrinkage rates were examined under combined curing regimes of four temperatures (40 °C, 20 °C, 0 °C, −10 °C) and three relative humidities (RH40%, RH60%, RH80%). Curing temperature and humidity primarily regulate shrinkage deformation by altering the internal moisture evaporation rate. Both evaporation and shrinkage rates exhibited a rapid initial increase, followed by deceleration, and finally stabilization with increasing age. A strong positive correlation was observed between these two parameters. The high-temperature and low-humidity condition (40 °C, RH40%) induced the most severe shrinkage. Four machine learning algorithms (XGBoost, RF, ANN, and KNN) were used to construct prediction models. After hyperparameter optimization and cross-validation, the RF models exhibited superior generalization and robustness (test set R2 > 0.94). The model accurately captures the complex non-linear relationship between environmental parameters and shrinkage. SHAP analysis on the optimal models identified the moisture evaporation rate as the primary driving factor. The analysis quantified the non-linear contributions of temperature and age, alongside the inhibitory effect of humidity. The study verified the consistency between data-driven models and physical mechanisms. This study elucidates the shrinkage mechanism under extreme conditions. It provides a reliable reference for crack control and life prediction in high-altitude engineering. Full article
(This article belongs to the Special Issue Geopolymers and Low Carbon Building Materials for Infrastructures)
Show Figures

Figure 1

Back to TopTop