Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (685)

Search Parameters:
Keywords = spatiotemporal convolutional neural network

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 20258 KB  
Article
Mining Scene Classification and Semantic Segmentation Using 3D Convolutional Neural Networks
by André Estevam Costa Oliveira, Matheus Corrêa Domingos, Valdivino Alexandre de Santiago Júnior and Maria Isabel Sobral Escada
Remote Sens. 2026, 18(8), 1112; https://doi.org/10.3390/rs18081112 - 8 Apr 2026
Viewed by 164
Abstract
High spatio-temporal resolution satellite imagery has become increasingly accessible thanks to advancements in the aerospace industry which, combined with a growing computational power, has enabled the spring of novel techniques regarding recognition in remote sensing (RS) images. However, there is still a lack [...] Read more.
High spatio-temporal resolution satellite imagery has become increasingly accessible thanks to advancements in the aerospace industry which, combined with a growing computational power, has enabled the spring of novel techniques regarding recognition in remote sensing (RS) images. However, there is still a lack of studies around 3D convolutions for spatio-temporal data applied to classification problems in RS. Hence, this study investigates the feasibility of 3D convolutional neural networks (3DCNNs) within a spatio-temporal perspective for scene classification and semantic segmentation in RS images, focusing on the identification of mining sites. We firstly developed a dataset covering several parts of Brazil based on MapBiomas products and Planet imagery, then we evaluated the effectiveness of 3DCNNs in capturing temporal information from a sequence of monthly captured images. Moreover, not only for scene classification but also for semantic segmentation, we compared 3D and 2D approaches. As for scene classification, a 3DCNN was better than the corresponding 2D model, while a 2D U-Net was better than a U-Net3D for semantic segmentation. The main explanation for this lies in the fact that a less costly annotation and training time strategy was adopted, but this may have harmed spatio-temporal approaches for semantic segmentation but not for scene classification. However, U-Net3D presented the highest Precision of all models, meaning that it is highly accurate when it predicts a positive. Moreover, 3DCNN (U-Net3D) presented significantly better performance with respect to semantic segmentation compared to other spatio-temporal approaches like ConvLSTM+U-Net and TempCNN. Sensitivity analysis revealed that the near-infrared (NIR) band played a decisive role in distinguishing mining areas, emphasizing its importance in highlighting subtle spectral variations associated with land-cover disturbances. Full article
(This article belongs to the Section Environmental Remote Sensing)
Show Figures

Figure 1

19 pages, 4124 KB  
Article
Prediction of Maximum Usable Frequency Based on a New Hybrid Deep Learning Model
by Yuyang Li, Zhigang Zhang and Jian Shen
Electronics 2026, 15(7), 1539; https://doi.org/10.3390/electronics15071539 - 7 Apr 2026
Viewed by 188
Abstract
The reliability of high-frequency (HF) frequency selection technology relies on the prediction accuracy of the Maximum Usable Frequency of the ionospheric F2 layer (MUF-F2). To improve its short-term prediction performance, a novel hybrid deep learning prediction model is proposed, which achieves accurate modeling [...] Read more.
The reliability of high-frequency (HF) frequency selection technology relies on the prediction accuracy of the Maximum Usable Frequency of the ionospheric F2 layer (MUF-F2). To improve its short-term prediction performance, a novel hybrid deep learning prediction model is proposed, which achieves accurate modeling of the complex spatiotemporal variation patterns of MUF-F2 by integrating a feature enhancement mechanism, a dual-branch feature extraction structure, and a bidirectional temporal dependency capture network. The hybrid prediction model integrates the Channel Attention mechanism (CA), Dual-Branch Convolutional Neural Network (DCNN), and Bidirectional Long Short-Term Memory network (BiLSTM). The model is trained and validated using MUF-F2 data from 5 communication links over China during geomagnetically quiet periods and 4 during geomagnetic storm periods, with the difference in the number of links attributed to experimental constraints and the disruptive effects of geomagnetic storms. Its performance is evaluated via multiple metrics, and a comparative analysis is conducted with commonly used prediction models such as the Long Short-Term Memory (LSTM) network. Experimental results show that during geomagnetically quiet periods, the proposed model achieves lower prediction errors (Root Mean Square Error (RMSE) < 1.1 MHz, Mean Absolute Percentage Error (MAPE) < 3.8%) and a higher goodness of fit (coefficient of determination (R2) > 0.94), with the average error reduction across all links ranging 8 from 6.2% to 46.9% compared with the baseline model. Under geomagnetic storm disturbance conditions, the model still maintains robust prediction performance, with R2 > 0.89 for all communication links, as well as RMSE < 0.6 MHz, Mean Absolute Error (MAE) < 0.4 MHz, and MAPE < 3.3%. The study demonstrates that the proposed CA-DCNN-BiLSTM model exhibits excellent prediction accuracy and anti-interference capability under different geomagnetic activity conditions, which can effectively improve the short-term prediction accuracy of MUF-F2 and provide more reliable technical support for HF communication frequency decision-making. Full article
Show Figures

Figure 1

21 pages, 1281 KB  
Article
A Lightweight Multi-Classification Model for Identifying Network Application Traffic Using Knowledge Distillation
by Zhiyuan Li and Yonghao Feng
Future Internet 2026, 18(4), 197; https://doi.org/10.3390/fi18040197 - 7 Apr 2026
Viewed by 176
Abstract
To address the limitations of insufficient feature representation, large model size, and high deployment cost in network traffic classification, a lightweight classification framework based on multi-teacher knowledge distillation is proposed. The framework consists of two heterogeneous teacher networks and a compact student network [...] Read more.
To address the limitations of insufficient feature representation, large model size, and high deployment cost in network traffic classification, a lightweight classification framework based on multi-teacher knowledge distillation is proposed. The framework consists of two heterogeneous teacher networks and a compact student network to enable end-to-end traffic classification under constrained computational resources. The teacher networks incorporate complementary spatio-temporal modeling strategies, including a bidirectional temporal convolutional network (BiTCN) enhanced with attention mechanisms and convolutional neural network (CNN), and a parallel spatio-temporal fusion architecture integrating bidirectional long short-term memory (BiLSTM) and CNN. Knowledge from the teacher ensemble is distilled into a lightweight CNN-based student network through soft-target supervision, leading to improved generalization capability with significantly reduced model complexity. Experimental results demonstrate that effective knowledge transfer is achieved while reducing model parameters by more than 80%, and performance gains of about 1–3% are obtained compared with baseline methods. These results indicate strong potential for practical deployment in resource-constrained network environments. Full article
Show Figures

Graphical abstract

17 pages, 2174 KB  
Article
RadarSSM: A Lightweight Spatiotemporal State Space Network for Efficient Radar-Based Human Activity Recognition
by Rubin Zhao, Fucheng Miao and Yuanjian Liu
Sensors 2026, 26(7), 2259; https://doi.org/10.3390/s26072259 - 6 Apr 2026
Viewed by 293
Abstract
Millimeter-wave radar has gradually gained popularity as a sensor mode for Human Activity Recognition (HAR) in recent years because it preserves the privacy of individuals and is resistant to environmental conditions. Nevertheless, the fast inference of high-dimensional and sparse 4D radar data is [...] Read more.
Millimeter-wave radar has gradually gained popularity as a sensor mode for Human Activity Recognition (HAR) in recent years because it preserves the privacy of individuals and is resistant to environmental conditions. Nevertheless, the fast inference of high-dimensional and sparse 4D radar data is still difficult to perform on low-resource edge devices. Current models, including 3D Convolutional Neural Networks and Transformer-based models, are frequently plagued by extensive parameter overhead or quadratic computational complexity, which restricts their applicability to edge applications. The present paper attempts to resolve these issues by introducing RadarSSM as a lightweight spatiotemporal hybrid network in the context of radar-based HAR. The explicit separation of spatial feature extraction and temporal dependency modeling helps RadarSSM decrease the overall complexity of computation significantly. Specifically, a spatial encoder based on depthwise separable 3D convolutions is designed to efficiently capture fine-grained geometric and motion features from voxelized radar data. For temporal modeling, a bidirectional State Space Model is introduced to capture long-range temporal dependencies with linear time complexity O(T), thereby avoiding the quadratic cost associated with self-attention mechanisms. Extensive experiments conducted on public radar HAR datasets demonstrate that RadarSSM achieves accuracy competitive with state-of-the-art methods while substantially reducing parameter count and computational cost relative to representative convolutional baselines. These results validate the effectiveness of RadarSSM and highlight its suitability for efficient radar sensing on edge hardware. Full article
(This article belongs to the Special Issue Radar and Multimodal Sensing for Ambient Assisted Living)
Show Figures

Figure 1

28 pages, 5258 KB  
Article
Dual-View Entropy-Driven AIS–Sonar Fusion for Surface and Underwater Target Discrimination
by Xiaoshuang Zhang, Jiayi Che, Xiaodan Xiong, Yucheng Zhang, Xinbo He, Mengsha Deng and Dezhi Wang
J. Mar. Sci. Eng. 2026, 14(7), 675; https://doi.org/10.3390/jmse14070675 - 4 Apr 2026
Viewed by 261
Abstract
Distinguishing surfaces from underwater targets in complex marine environments is challenging when relying solely on physical sonar features. To address the high uncertainty inherent in single-modal features and the conflicts arising from heterogeneous data, we propose a Dual-View Entropy-Driven Negation Dempster–Shafer (DVE-NDS) fusion [...] Read more.
Distinguishing surfaces from underwater targets in complex marine environments is challenging when relying solely on physical sonar features. To address the high uncertainty inherent in single-modal features and the conflicts arising from heterogeneous data, we propose a Dual-View Entropy-Driven Negation Dempster–Shafer (DVE-NDS) fusion method that integrates AIS kinematic priors with passive sonar signals. First, a heterogeneous recognition framework is constructed. LOFAR and DEMON features are extracted via convolutional neural networks (CNNs), while a Negation Basic Probability Assignment (Negation BPA) strategy is introduced to transform AIS spatiotemporal mismatches into effective "negation support" for non-cooperative underwater targets. Instead of relying on a single conflict coefficient, the proposed method jointly considers evidence self-information and inter-source consistency. Evidence quality is quantified using improved Deng entropy and negation belief entropy, while mutual trust is evaluated via the Jousselme distance. Heterogeneous evidence is weighted and corrected by generated coupling weights, effectively suppressing low-quality evidence and sharpening decision boundaries. Simulation results confirm that DVE-NDS improves macro-F1 over classical fusion, indicating the framework’s potential for handling conflicting evidence, though the current validation remains simulation-based and should be regarded as a methodological proof-of-concept. Full article
(This article belongs to the Special Issue Emerging Computational Methods in Intelligent Marine Vehicles)
Show Figures

Figure 1

24 pages, 3090 KB  
Article
A Convolutional Neural Network Framework for Opportunistic GNSS-R Wind Speed Retrieval over Inland Lakes
by Yanan Ni, Jiajia Chen, Jiajia Jia and Xinnian Guo
Electronics 2026, 15(7), 1501; https://doi.org/10.3390/electronics15071501 - 3 Apr 2026
Viewed by 227
Abstract
Global Navigation Satellite System Reflectometry (GNSS-R) provides a promising approach for wind speed retrieval over inland waters, with relevance to wind energy assessment and lake–atmosphere exchange studies. Existing GNSS-R wind retrieval methods are well established for open oceans but face major challenges over [...] Read more.
Global Navigation Satellite System Reflectometry (GNSS-R) provides a promising approach for wind speed retrieval over inland waters, with relevance to wind energy assessment and lake–atmosphere exchange studies. Existing GNSS-R wind retrieval methods are well established for open oceans but face major challenges over inland waters, where coherent scattering dominates and traditional ocean models produce large systematic biases. Unlike open oceans, inland waters are dominated by coherent scattering due to limited fetch, resulting in Delay-Doppler Maps (DDM) with highly concentrated energy and minimal spreading. These characteristics render conventional ocean-based retrieval models—built on incoherent scattering assumptions—often inadequate. To overcome this, we develop a lightweight convolutional neural network (CNN) tailored to the coherent regime, using raw CYGNSS DDM as input for end-to-end wind speed regression. Cross-seasonal validation over Lake Victoria and Lake Hongze shows that the model robustly captures wind-driven spatiotemporal patterns aligned with ERA5. Notably, ERA5 reanalysis winds exhibit uncertainties over inland waters, with a root mean square error (RMSE) of 1.5–2.5 m/s against in situ buoys. The model yields a low RMSE (<0.7 m/s) in reconstructing ERA5-resolved wind patterns. This work extends GNSS-R to inland waters, offering a lightweight, deployable remote sensing solution for wind energy and lake–atmosphere research. Full article
Show Figures

Figure 1

21 pages, 13827 KB  
Article
An Integrated Model Based on CNN-Transformer and PLUS for Urban Expansion Simulation in the Yangtze River Delta, China
by Linyu Ma, Jue Xiao, Gan Teng, Ting Zhang and Longqian Chen
Remote Sens. 2026, 18(7), 1071; https://doi.org/10.3390/rs18071071 - 2 Apr 2026
Viewed by 325
Abstract
Land use changes within urban agglomerations exhibit significant spatiotemporal heterogeneity and regional diversity. In urban agglomeration land simulation, traditional models often struggle to systematically capture these variations. We introduce the GCTP, a novel framework that integrates guided Geographical zoning, Convolutional Neural Networks (CNN)-Transformer, [...] Read more.
Land use changes within urban agglomerations exhibit significant spatiotemporal heterogeneity and regional diversity. In urban agglomeration land simulation, traditional models often struggle to systematically capture these variations. We introduce the GCTP, a novel framework that integrates guided Geographical zoning, Convolutional Neural Networks (CNN)-Transformer, and the Patch-generating Land Use Simulation (PLUS) model. Initially, guided K-means clustering was employed for geographic zoning to characterize regional spatial non-stationarity. Then, a CNN-Transformer network leveraged self-attention mechanisms to capture multi-scale spatial correlations, obtaining pixel-level development probabilities. Finally, these probabilities were fused with PLUS- Land Expansion Analysis Strategy (LEAS) outputs to drive PLUS- Cellular Automata with multi-type Random Seeds (CARS) for patch-level simulation. The results demonstrate the following: (1) The embedding of guided zoning enabled the model to achieve an Overall Accuracy (OA) of 0.941, effectively mitigating global simulation bias. (2) The optimal simulation performance occurred at a fusion weight of 0.81, yielding a Kappa of 0.8917 and an Figure of Merit (FoM) of 0.3830, significantly exceeding a single model. (3) The 2030 simulation indicates that the GCTP model effectively reduces isolated pixels at urban fringes. The GCTP generates neighborhood patterns with high spatial compactness and geographic consistency. This study highlights the significant advantages of integrating long-range spatial perception with geographical heterogeneity constraints in the land expansion simulation of urban agglomerations. The findings support more precise territorial spatial planning practices. Full article
(This article belongs to the Special Issue Machine Learning of Remote Sensing Imagery for Land Cover Mapping)
Show Figures

Figure 1

26 pages, 4761 KB  
Article
A CNN–LSTM Framework for Player-Specific Baseball Pitch Type Prediction from Video Sequences
by Chin-Chih Chang, Chi-Hung Wei, Hao-Chen Li and Sean Hsiao
Appl. Syst. Innov. 2026, 9(4), 75; https://doi.org/10.3390/asi9040075 - 30 Mar 2026
Viewed by 405
Abstract
The performance of the pitcher is the cornerstone of baseball, often determining the flow and ultimate outcome of a game. Given this centrality, understanding the mechanics of an elite pitcher and decoding their strategies are paramount for both internal optimization and competitive scouting. [...] Read more.
The performance of the pitcher is the cornerstone of baseball, often determining the flow and ultimate outcome of a game. Given this centrality, understanding the mechanics of an elite pitcher and decoding their strategies are paramount for both internal optimization and competitive scouting. This study proposes an end-to-end deep learning pipeline for automatically classifying five distinct pitch types from raw broadcast footage of MLB pitcher Max Scherzer between 2015 and 2020. By formulating pitch delivery as a time-series classification problem tailored to the unique biomechanics of an elite athlete, the proposed CNN–LSTM framework integrates per-frame spatial feature extraction using an advanced CNN backbone (YOLOv8s-cls) with a two-layer long short-term memory (LSTM) network to capture subtle biomechanical cues across a standardized 20-frame delivery sequence. While skeletal pose estimation primarily focuses on tracking major joints to analyze standard pitching mechanics, the proposed pixel-based method preserves fine-grained visual cues—such as finger grip and wrist rotation—that are critical for distinguishing pitch variations. The proposed framework achieved an accuracy of 91.8% under a standard Random Split and, importantly, 84.5% under a strict Chronological Split across different seasons, validating the feasibility of automated pitch “tell” detection from broadcast video. The resulting system provides coaches and analysts with an objective, data-driven tool for generating personalized scouting reports, identifying mechanical inconsistencies, and refining pitching strategies. Full article
Show Figures

Figure 1

28 pages, 7867 KB  
Article
A CEEMDAN-CNN-BiLSTM-SDQN Framework for Photovoltaic Power Forecasting: Integrating Multi-Scale Decomposition with Adaptive Reinforcement Learning Compensation
by Weijie Jia, Keying Liu, Jinghui Xu and Yapeng Zhu
Energies 2026, 19(7), 1649; https://doi.org/10.3390/en19071649 - 27 Mar 2026
Viewed by 329
Abstract
Accurate photovoltaic (PV) power forecasting is crucial for grid stability and the integration of renewable energy. To address the multiscale, nonlinear characteristics of PV power series and the limitations of traditional methods in dynamic error compensation, a novel hybrid forecasting framework is proposed, [...] Read more.
Accurate photovoltaic (PV) power forecasting is crucial for grid stability and the integration of renewable energy. To address the multiscale, nonlinear characteristics of PV power series and the limitations of traditional methods in dynamic error compensation, a novel hybrid forecasting framework is proposed, integrating Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (CEEMDAN), Convolutional Neural Network-Bidirectional Long Short-Term Memory (CNN-BiLSTM), and a Simplified Deep Q-Network (SDQN). The framework first decomposes the power series into subcomponents across different frequency bands via CEEMDAN. Subsequently, dedicated CNN-BiLSTM sub-models are employed in parallel to extract spatiotemporal features from each component. Finally, an SDQN agent is introduced to perform real-time error compensation. Validation based on operational data from a PV plant in Ningxia, China, demonstrates that the proposed framework achieves RMSE, MAE, MAPE, and R2 values of 0.4463, 0.1256, 1.2814%, and 92.58%, respectively, significantly outperforming benchmark models. Specifically, the CEEMDAN decomposition effectively mitigates mode mixing. The CNN-BiLSTM as the base predictor reduces RMSE by 25.04–65.68% compared to mainstream models. Furthermore, the SDQN compensation mechanism delivers an additional 24.5% reduction in prediction error. The proposed approach thus constitutes a high-precision, adaptive solution for PV power forecasting. Full article
(This article belongs to the Section A2: Solar Energy and Photovoltaic Systems)
Show Figures

Figure 1

23 pages, 129074 KB  
Article
High-Resolution Air Temperature Estimation Using the Full Landsat Spectral Range and Information-Based Machine Learning
by Daniel Eitan, Asher Holder, Zohar Yakhini and Alexandra Chudnovsky
Remote Sens. 2026, 18(6), 954; https://doi.org/10.3390/rs18060954 - 22 Mar 2026
Viewed by 353
Abstract
Accurate mapping of near-surface air temperature (Tair) at the fine spatial resolution is required for city-scale monitoring and remains a critical challenge in Earth Observation (EO). Reliance on ground-based measurements is constrained by their sparse spatial coverage and high operational [...] Read more.
Accurate mapping of near-surface air temperature (Tair) at the fine spatial resolution is required for city-scale monitoring and remains a critical challenge in Earth Observation (EO). Reliance on ground-based measurements is constrained by their sparse spatial coverage and high operational costs. We present a novel, scalable machine learning framework designed to overcome this limitation. Our method utilizes interpretable Convolutional Neural Networks (CNNs) to fuse high-resolution Landsat data, integrating both thermal and reflective spectral bands, with contextual spatiotemporal metadata. This approach allows for inference, at 30 m resolution, of Tair fields without relying on dense, localized ground monitoring networks. Our hybrid CNN architecture is optimized for spatial generalization, maintaining strong and transferable performance (station-wise R20.88) across diverse environments from humid coasts (R20.89) to arid interiors (R20.84). Although focused on a specific geographical region, our results suggest a robust and reproducible pathway for generating spatially consistent temperature fields from globally available EO archives, directly supporting urban heat island mitigation, climate policy development, and high-resolution public health assessment worldwide. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

16 pages, 5787 KB  
Article
USTGCN: A Unified Spatio-Temporal Graph Convolutional Network for Stock-Ranking Prediction
by Wenjie Yao, Lele Gao, Xiangzhou Zhang, Haotao Chen, Mingzhe Liu and Yong Hu
Electronics 2026, 15(6), 1317; https://doi.org/10.3390/electronics15061317 - 21 Mar 2026
Viewed by 283
Abstract
Stock-ranking prediction is an important task in quantitative finance because it directly influences portfolio construction and alpha generation. Recent Graph Neural Network (GNN) models provide a promising way to describe inter-stock dependencies, but many existing methods still have difficulty balancing rapidly changing market [...] Read more.
Stock-ranking prediction is an important task in quantitative finance because it directly influences portfolio construction and alpha generation. Recent Graph Neural Network (GNN) models provide a promising way to describe inter-stock dependencies, but many existing methods still have difficulty balancing rapidly changing market interactions with relatively stable structural relationships. They are also easily affected by financial micro-structure noise. To address these issues, this paper proposes USTGCN, a Unified Spatio-Temporal Graph Convolutional Network for stock-ranking prediction. USTGCN adopts a dual-stream temporal encoder based on ALSTM and GRU to capture short-term dynamic patterns and longer-horizon structural information, respectively. We further introduce a rolling-window correlation smoothing strategy to build a more stable dynamic graph, and then integrate the dynamic and structural graph views through a shared fusion layer. Skip connections are used to preserve original temporal information during spatial aggregation. Experiments on the CSI100 and CSI300 benchmark datasets show that USTGCN achieves IC values of 0.141 and 0.154, respectively, and exhibits improved drawdown control during stressed market periods, indicating its practical value for quantitative trading. Full article
Show Figures

Figure 1

32 pages, 7914 KB  
Article
UAV Target Detection and Tracking Integrating a Dynamic Brain–Computer Interface
by Jun Wang, Zanyang Li, Lirong Yan, Muhammad Imtiaz, Hang Li, Muhammad Usman Shoukat, Jianatihan Jinsihan, Benjun Feng, Yi Yang, Fuwu Yan, Shumo He and Yibo Wu
Drones 2026, 10(3), 222; https://doi.org/10.3390/drones10030222 - 21 Mar 2026
Viewed by 589
Abstract
To address the inherent limitations in the robustness of fully autonomous unmanned aerial vehicle (UAV) visual perception and the high cognitive workload associated with manual control, this paper proposes a human-in-the-loop brain–computer interface (BCI) control framework. The system integrates steady-state visual evoked potential [...] Read more.
To address the inherent limitations in the robustness of fully autonomous unmanned aerial vehicle (UAV) visual perception and the high cognitive workload associated with manual control, this paper proposes a human-in-the-loop brain–computer interface (BCI) control framework. The system integrates steady-state visual evoked potential (SSVEP) with deep learning techniques to create a spatio-temporally dynamic interaction paradigm, enabling real-time alignment between visual targets and frequency stimuli. At the perception level, an enhanced YOLOv11 network incorporating partial convolution (PConv) and shape intersection over union (Shape-IoU) loss is developed and coupled with the DeepSort multi-object tracking algorithm. This configuration ensures high-speed execution on edge computing platforms while maintaining stable stimulus coverage over dynamic targets, thus providing a robust visual induction environment for EEG decoding. At the neural decoding level, an enhanced task-discriminant component analysis (TDCA-V) algorithm is introduced to improve signal detection stability within non-stationary flight conditions. Experimental results demonstrate that within the predefined fixation task window, the system achieves 100% success in maintaining target identity (ID). The BCI system achieved an average command recognition accuracy of 91.48% within a 1.0 s time window, with the TDCA-V algorithm significantly outperforming traditional spatial filtering methods in dynamic scenarios. These findings demonstrate the system’s effectiveness in decoupling human cognitive intent from machine execution, providing a robust solution for human–machine collaborative control. Full article
(This article belongs to the Section Artificial Intelligence in Drones (AID))
Show Figures

Figure 1

28 pages, 4748 KB  
Article
ProMix-DGNet: A Process-Aware Spatiotemporal Network for Sintering System Prediction
by Zhili Zhang, Yuxin Wan, Liya Wang and Jie Li
Sensors 2026, 26(6), 1953; https://doi.org/10.3390/s26061953 - 20 Mar 2026
Viewed by 421
Abstract
Multistep-ahead prediction of critical states in the iron ore sintering process is essential for maintaining production stability, enhancing energy efficiency, and reducing industrial emissions. However, large time delays, strong coupling, and condition drifts challenge existing spatiotemporal graph neural networks (STGNNs). This paper proposes [...] Read more.
Multistep-ahead prediction of critical states in the iron ore sintering process is essential for maintaining production stability, enhancing energy efficiency, and reducing industrial emissions. However, large time delays, strong coupling, and condition drifts challenge existing spatiotemporal graph neural networks (STGNNs). This paper proposes Process-aware Mixed Dynamic Graph Network (ProMix-DGNet), which integrates a Decoupled Two-Stream Topology Learning mechanism—fusing Adaptive Static Graph with a Radial Basis Function (RBF)-driven Dynamic Graph Constructor—to ensure robust spatial modeling under high-noise conditions. Furthermore, Process-View Global Mixer explicitly captures long-range process coupling across the entire sintering strand, overcoming the receptive field limitations of traditional graph convolutions. In the decoding phase, a future control-informed module utilizes a bidirectional Long Short-Term Memory (BiLSTM) and a global mixer to align known future control setpoints with the system’s spatial topology. These features are integrated via a gated residual mechanism that dynamically modulates the interaction between control intents and historical representations. Extensive experiments conducted on two real-world industrial datasets, Sinter-A and Sinter-B, demonstrate that ProMix-DGNet consistently outperforms mainstream baselines across multiple metrics, including Mean Absolute Error (MAE) and Root Mean Square Error (RMSE). The results verify the model’s higher accuracy and robustness in complex large-time-delay systems, offering a reliable framework for the intelligent monitoring and closed-loop optimization of sintering process. Full article
Show Figures

Figure 1

19 pages, 1361 KB  
Article
A New Method for Optimizing Low-Earth-Orbit Satellite Communication Links Based on Deep Reinforcement Learning
by He Yu, Shengli Li, Junchao Wu, Yanhong Sun and Limin Wang
Aerospace 2026, 13(3), 285; https://doi.org/10.3390/aerospace13030285 - 18 Mar 2026
Viewed by 281
Abstract
In low-Earth-orbit (LEO) satellite networks, the need for intelligent parameter-adjustment strategies has become increasingly critical due to the presence of highly dynamic channel conditions, limited spectrum resources, and complex interference environments. In this paper, a method for optimizing LEO satellite communication links based [...] Read more.
In low-Earth-orbit (LEO) satellite networks, the need for intelligent parameter-adjustment strategies has become increasingly critical due to the presence of highly dynamic channel conditions, limited spectrum resources, and complex interference environments. In this paper, a method for optimizing LEO satellite communication links based on deep reinforcement learning (DRL) is proposed. Through the optimization of the transmit power, the modulation and coding scheme (MCS), the beamforming parameters, and the retransmission mechanisms, adaptive link control is achieved in dynamic operational scenarios. A multidimensional state space is constructed, within which the channel state information, the interference environment, and the historical performance metrics are integrated. The spatio-temporal characteristics of the channel are extracted by means of a hybrid neural architecture that incorporates a convolutional neural network (CNN) and a long short-term memory (LSTM) network. To effectively accommodate both continuous and discrete action spaces, a hybrid DRL framework that combines proximal policy optimization (PPO) with a deep Q-network (DQN) is employed, thereby enabling cross-layer optimization of the physical-layer and link-layer parameters. The results demonstrate that substantial improvements in throughput, bit error rate (BER), and transmit-power efficiency are achieved under severely time-varying channel conditions, which provides a new idea for resource management and dynamic-environment adaptation in satellite communication systems. Full article
(This article belongs to the Special Issue Advanced Spacecraft/Satellite Technologies (2nd Edition))
Show Figures

Figure 1

23 pages, 5091 KB  
Article
Multiclass Anomaly Detection in Bridge Health Monitoring Data via Attention Enhancement and Class Imbalance Mitigation
by Wenda Ma, Qizhi Tang, Lei Huang and Shihao Zhang
Buildings 2026, 16(6), 1181; https://doi.org/10.3390/buildings16061181 - 17 Mar 2026
Viewed by 308
Abstract
Bridge structural health monitoring (BSHM) systems are essential for assessing the operational performance and safety of long-span bridges. However, monitoring data are often affected by factors such as sensor malfunctions, environmental disturbances, or power interruptions, leading to various anomalous data. Moreover, the multiclass [...] Read more.
Bridge structural health monitoring (BSHM) systems are essential for assessing the operational performance and safety of long-span bridges. However, monitoring data are often affected by factors such as sensor malfunctions, environmental disturbances, or power interruptions, leading to various anomalous data. Moreover, the multiclass imbalance of the data presents a major challenge to traditional anomaly detection methods. To address this issue, a novel multiclass anomaly detection method based on an improved deep convolutional neural network is proposed. Specifically, a ResNet50 architecture integrated with the convolutional block attention module (CBAM) is developed to enhance the extraction of discriminative features. Additionally, the Focal Loss function is introduced to emphasize the loss weight of minority samples, reducing the influence of majority classes, thereby effectively overcoming the class imbalance issue in multiclass anomaly detection. The proposed method is trained and validated using measured acceleration data collected from a large-scale cable-stayed bridge. The experimental results indicate that the model achieves an overall accuracy of 98.28%, while effectively improving the classification performance of minority categories. The method further reproduces the spatiotemporal distribution of anomalies in full-month monitoring data, confirming its robustness and engineering applicability for large-scale automated anomaly diagnosis in BSHM systems. Full article
Show Figures

Figure 1

Back to TopTop