Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,155)

Search Parameters:
Keywords = multi-task fusion

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 2125 KB  
Article
A Load Forecasting Model Based on Spatiotemporal Partitioning and Cross-Regional Attention Collaboration
by Xun Dou, Ruiang Yang, Zhenlan Dou, Chunyan Zhang, Chen Xu and Jiacheng Li
Sustainability 2025, 17(18), 8162; https://doi.org/10.3390/su17188162 - 10 Sep 2025
Abstract
With the advancement of new power system construction, thermostatically controlled loads represented by regional air conditioning systems are being extensively integrated into the grid, leading to a surge in the number of user nodes. This large-scale integration of new loads creates challenges for [...] Read more.
With the advancement of new power system construction, thermostatically controlled loads represented by regional air conditioning systems are being extensively integrated into the grid, leading to a surge in the number of user nodes. This large-scale integration of new loads creates challenges for the grid, as the resulting load data exhibits strong periodicity and randomness over time. These characteristics are influenced by factors like temperature and user behavior. At the same time, spatially adjacent nodes show similarities and clustering in electricity usage. This creates complex spatiotemporal coupling features. These complex spatiotemporal characteristics challenge traditional forecasting methods. Their high model complexity and numerous parameters often lead to overfitting or the curse of dimensionality, which hinders both prediction accuracy and efficiency. To address this issue, this paper proposes a load forecasting method based on spatiotemporal partitioning and collaborative cross-regional attention. First, a spatiotemporal similarity matrix is constructed using the Shape Dynamic Time Warping (ShapeDTW) algorithm and an adaptive Gaussian kernel function based on the Haversine distance. Spectral clustering combined with the Gap Statistic criterion is then applied to adaptively determine the optimal number of partitions, dividing all load nodes in the power grid into several sub-regions with homogeneous spatiotemporal characteristics. Second, for each sub-region, a local Spatiotemporal Graph Convolutional Network (STGCN) model is built. By integrating gated temporal convolution with spatial feature extraction, the model accurately captures the spatiotemporal evolution patterns within each sub-region. On this basis, a cross-regional attention mechanism is designed to dynamically learn the correlation weights among sub-regions, enabling collaborative fusion of global features. Finally, the proposed method is evaluated on a multi-node load dataset. The effectiveness of the approach is validated through comparative experiments and ablation studies (that is, by removing key components of the model to evaluate their contribution to the overall performance). Experimental results demonstrate that the proposed method achieves excellent performance in short-term load forecasting tasks across multiple nodes. Full article
(This article belongs to the Special Issue Energy Conservation Towards a Low-Carbon and Sustainability Future)
30 pages, 5137 KB  
Article
High-Resolution Remote Sensing Imagery Water Body Extraction Using a U-Net with Cross-Layer Multi-Scale Attention Fusion
by Chunyan Huang, Mingyang Wang, Zichao Zhu and Yanling Li
Sensors 2025, 25(18), 5655; https://doi.org/10.3390/s25185655 - 10 Sep 2025
Abstract
The accurate extraction of water bodies from remote sensing imagery is crucial for water resource monitoring and flood disaster warning. However, this task faces significant challenges due to complex land cover, large variations in water body morphology and spatial scales, and spectral similarities [...] Read more.
The accurate extraction of water bodies from remote sensing imagery is crucial for water resource monitoring and flood disaster warning. However, this task faces significant challenges due to complex land cover, large variations in water body morphology and spatial scales, and spectral similarities between water and non-water features, leading to misclassification and low accuracy. While deep learning-based methods have become a research hotspot, traditional convolutional neural networks (CNNs) struggle to represent multi-scale features and capture global water body information effectively. To enhance water feature recognition and precisely delineate water boundaries, we propose the AMU-Net model. Initially, an improved residual connection module was embedded into the U-Net backbone to enhance complex feature learning. Subsequently, a multi-scale attention mechanism was introduced, combining grouped channel attention with multi-scale convolutional strategies for lightweight yet precise segmentation. Thereafter, a dual-attention gated modulation module dynamically fusing channel and spatial attention was employed to strengthen boundary localization. Furthermore, a cross-layer geometric attention fusion module, incorporating grouped projection convolution and a triple-level geometric attention mechanism, optimizes segmentation accuracy and boundary quality. Finally, a triple-constraint loss framework synergistically optimized global classification, regional overlap, and background specificity to boost segmentation performance. Evaluated on the GID and WHDLD datasets, AMU-Net achieved remarkable IoU scores of 93.6% and 95.02%, respectively, providing an effective new solution for remote sensing water body extraction. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

16 pages, 3435 KB  
Article
Incremental Spatio-Temporal Augmented Sampling for Power Grid Operation Behavior Recognition
by Lingwen Meng, Di He, Guobang Ban and Siqi Guo
Electronics 2025, 14(18), 3579; https://doi.org/10.3390/electronics14183579 - 9 Sep 2025
Abstract
Accurate recognition of power grid operation behaviors is crucial for ensuring both safety and operational efficiency in smart grid systems. However, this task presents significant challenges due to dynamic environmental variations, limited labeled training data availability, and the necessity for continuous model adaptation. [...] Read more.
Accurate recognition of power grid operation behaviors is crucial for ensuring both safety and operational efficiency in smart grid systems. However, this task presents significant challenges due to dynamic environmental variations, limited labeled training data availability, and the necessity for continuous model adaptation. To overcome these limitations, we propose an Incremental Spatio-temporal Augmented Sampling (ISAS) method for power grid operation behavior recognition. Specifically, we design a spatio-temporal Feature-Enhancement Fusion Module (FEFM) which employs multi-scale spatio-temporal augmented fusion combined with a cross-scale aggregation mechanism, enabling robust feature learning that is resilient to environmental interference. Furthermore, we introduce a Selective Replay Mechanism (SRM) that implements a dual-criteria sample selection strategy based on error variability and feature-space divergence metrics, ensuring optimal memory bank updates that simultaneously maximize information gain while minimizing feature redundancy. Experimental results on the power grid behavior dataset demonstrate significant advantages of the proposed method in recognition robustness and knowledge retention compared to other methods. For example, it achieves an accuracy of 89.80% on sunny days and maintains exceptional continual learning stability with merely 2.74% forgetting rate on three meteorological scenarios. Full article
(This article belongs to the Special Issue Applications and Challenges of Image Processing in Smart Environment)
Show Figures

Figure 1

21 pages, 5345 KB  
Article
Design and Development of an Intelligent Robotic Feeding Control System for Sheep
by Haina Jiang, Haijun Li and Guoxing Cai
Agriculture 2025, 15(18), 1912; https://doi.org/10.3390/agriculture15181912 - 9 Sep 2025
Abstract
With the widespread adoption of intelligent technologies in animal husbandry, traditional manual feeding methods can no longer meet the demands for precision and efficiency in modern sheep farming. To address this gap, we present an intelligent robotic feeding system designed to enhance feeding [...] Read more.
With the widespread adoption of intelligent technologies in animal husbandry, traditional manual feeding methods can no longer meet the demands for precision and efficiency in modern sheep farming. To address this gap, we present an intelligent robotic feeding system designed to enhance feeding efficiency, reduce labor intensity, and enable precise delivery of feed. This system, developed on the ROS platform, integrates LiDAR-based SLAM with point cloud rendering and an Octomap 3D grid map. It combines an improved bidirectional RRT* algorithm with Dynamic Window Approach (DWA) for efficient path planning and uses 3D LiDAR data along with the RANSAC algorithm for slope detection and navigation information extraction. The YOLOv8s model is utilized for precise sheep pen marker identification, while integration with weighing sensors and a farm management system ensures accurate feed distribution control. The main research contribution lies in the development of a comprehensive, multi-sensor fusion system capable of achieving autonomous feeding in dynamic and complex environments. Experimental results show that the system achieves centimeter-level accuracy in localization and attitude control, with FAST-LIO2 maintaining precision within 1° of attitude angle errors. Compared to baseline performance, the system reduces node count by 17.67%, shortens path length by 0.58 cm, and cuts computation time by 42.97%. At a speed of 0.8 m/s, the robot achieves a maximum longitudinal deviation of 7.5 cm and a maximum heading error of 5.6°, while straight-line deviation remains within ±2.2 cm. In a 30 kg feeding task, the system demonstrates zero feed wastage, highlighting its potential for intelligent feeding in modern sheep farming. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

16 pages, 846 KB  
Article
MMKT: Multimodal Sentiment Analysis Model Based on Knowledge-Enhanced and Text-Guided Learning
by Chengkai Shi and Yunhua Zhang
Appl. Sci. 2025, 15(17), 9815; https://doi.org/10.3390/app15179815 - 7 Sep 2025
Viewed by 283
Abstract
Multimodal Sentiment Analysis (MSA) aims to predict subjective human emotions by leveraging multimodal information. However, existing research inadequately utilizes explicit sentiment semantic information at the lexical level in text and overlooks noise interference from non-dominant modalities, such as irrelevant movements in visual modalities [...] Read more.
Multimodal Sentiment Analysis (MSA) aims to predict subjective human emotions by leveraging multimodal information. However, existing research inadequately utilizes explicit sentiment semantic information at the lexical level in text and overlooks noise interference from non-dominant modalities, such as irrelevant movements in visual modalities and background noise in audio modalities. To address this issue, we propose a multimodal sentiment analysis model based on knowledge enhancement and text-guided learning (MMKT). The model constructs a sentiment knowledge graph for the textual modality using the SenticNet knowledge base. This graph directly annotates word-level sentiment polarity, strengthening the model’s understanding of emotional vocabulary. Furthermore, global sentiment knowledge features are generated through graph embedding computations to enhance the multimodal fusion process. Simultaneously, a dynamic text-guided learning approach is introduced, which dynamically leverages multi-scale textual features to actively suppress redundant or conflicting information in visual and audio modalities, thereby generating purer cross-modal representations. Finally, concatenated textual features, cross-modal features, and knowledge features are utilized for sentiment prediction. Experimental results on the CMU-MOSEI and Twitter2019 dataset demonstrate the superior performance of the MMKT model. Full article
Show Figures

Figure 1

25 pages, 69171 KB  
Article
CrackNet-Weather: An Effective Pavement Crack Detection Method Under Adverse Weather Conditions
by Wei Wang, Xiaoru Yu, Bin Jing, Ziqi Tang, Wei Zhang, Shengyu Wang, Yao Xiao, Shu Li and Liping Yang
Sensors 2025, 25(17), 5587; https://doi.org/10.3390/s25175587 - 7 Sep 2025
Viewed by 307
Abstract
Accurate pavement crack detection under adverse weather conditions is essential for road safety and effective pavement maintenance. However, factors such as reduced visibility, background noise, and irregular crack morphology make this task particularly challenging in real-world environments. To address these challenges, we propose [...] Read more.
Accurate pavement crack detection under adverse weather conditions is essential for road safety and effective pavement maintenance. However, factors such as reduced visibility, background noise, and irregular crack morphology make this task particularly challenging in real-world environments. To address these challenges, we propose CrackNet-Weather, which is a robust and efficient detection method that systematically incorporates three key modules: a Haar Wavelet Downsampling Block (HWDB) for enhanced frequency information preservation, a Strip Pooling Bottleneck Block (SPBB) for multi-scale and context-aware feature fusion, and a Dynamic Sampling Upsampling Block (DSUB) for content-adaptive spatial feature reconstruction. Extensive experiments conducted on a challenging dataset containing both rainy and snowy weather demonstrate that CrackNet-Weather significantly outperforms mainstream baseline models, achieving notable improvements in mean Average Precision, especially for low-contrast, fine, and irregular cracks. Furthermore, our method maintains a favorable balance between detection accuracy and computational complexity, making it well suited for practical road inspection and large-scale deployment. These results confirm the effectiveness and practicality of CrackNet-Weather in addressing the challenges of real-world pavement crack detection under adverse weather conditions. Full article
Show Figures

Figure 1

24 pages, 5456 KB  
Article
Remaining Useful Life Prediction for Aero-Engines Based on Multi-Scale Dilated Fusion Attention Model
by Guosong Xiao, Chenfeng Jin and Jie Bai
Appl. Sci. 2025, 15(17), 9813; https://doi.org/10.3390/app15179813 - 7 Sep 2025
Viewed by 959
Abstract
To address the limitations of CNNs and RNNs in handling complex operating conditions, multi-scale degradation patterns, and long-term dependencies—with attention mechanisms often failing to highlight key degradation features—this paper proposes a remaining useful life (RUL) prediction framework based on a multi-scale dilated fusion [...] Read more.
To address the limitations of CNNs and RNNs in handling complex operating conditions, multi-scale degradation patterns, and long-term dependencies—with attention mechanisms often failing to highlight key degradation features—this paper proposes a remaining useful life (RUL) prediction framework based on a multi-scale dilated fusion attention (MDFA) module. The MDFA leverages parallel dilated convolutions with varying dilation rates to expand receptive fields, while a global-pooling branch captures sequence-level degradation trends. Additionally, integrated channel and spatial attention mechanisms enhance the model’s ability to emphasize informative features and suppress noise, thereby improving overall prediction robustness. The proposed method is evaluated on NASA’s C-MAPSS and N-CMAPSS datasets, achieving MAE values of 0.018–0.026, RMSE values of 0.021–0.032, and R2 scores above 0.987, demonstrating superior accuracy and stability compared to existing baselines. Furthermore, to verify generalization across domains, experiments on the PHM2012 bearing dataset show similar performance (MAE: 0.023–0.026, RMSE: 0.031–0.032, R2: 0.987–0.995), confirming the model’s effectiveness under diverse operating conditions and its adaptability to different degradation behaviors. This study provides a practical and interpretable deep-learning solution for RUL prediction, with broad applicability to aero-engine prognostics and other industrial health-monitoring tasks. Full article
(This article belongs to the Section Mechanical Engineering)
Show Figures

Figure 1

20 pages, 3675 KB  
Article
FDMNet: A Multi-Task Network for Joint Detection and Segmentation of Three Fish Diseases
by Zhuofu Liu, Zigan Yan and Gaohan Li
J. Imaging 2025, 11(9), 305; https://doi.org/10.3390/jimaging11090305 - 6 Sep 2025
Viewed by 151
Abstract
Fish diseases are one of the primary causes of economic losses in aquaculture. Existing deep learning models have progressed in fish disease detection and lesion segmentation. However, many models still have limitations, such as detecting only a single type of fish disease or [...] Read more.
Fish diseases are one of the primary causes of economic losses in aquaculture. Existing deep learning models have progressed in fish disease detection and lesion segmentation. However, many models still have limitations, such as detecting only a single type of fish disease or completing only a single task within fish disease detection. To address these limitations, we propose FDMNet, a multi-task learning network. Built upon the YOLOv8 framework, the network incorporates a semantic segmentation branch with a multi-scale perception mechanism. FDMNet performs detection and segmentation simultaneously. The detection and segmentation branches use the C2DF dynamic feature fusion module to address information loss during local feature fusion across scales. Additionally, we use uncertainty-based loss weighting together with PCGrad to mitigate conflicting gradients between tasks, improving the stability and overall performance of FDMNet. On a self-built image dataset containing three common fish diseases, FDMNet achieved 97.0% mAP50 for the detection task and 85.7% mIoU for the segmentation task. Relative to the multi-task YOLO-FD baseline, FDMNet’s detection mAP50 improved by 2.5% and its segmentation mIoU by 5.4%. On the dataset constructed in this study, FDMNet achieved competitive accuracy in both detection and segmentation. These results suggest potential practical utility. Full article
Show Figures

Figure 1

16 pages, 1705 KB  
Article
MLEDNet: Multi-Directional Learnable Edge Information-Assisted Dense Nested Network for Infrared Small Target Detection
by Yong Li, Wenjie Kang, Wei Zhao and Xuchong Liu
Electronics 2025, 14(17), 3547; https://doi.org/10.3390/electronics14173547 - 6 Sep 2025
Viewed by 357
Abstract
Infrared small target detection (IRSTD) remains a critical yet challenging task due to the inherent low signal-to-noise ratio, weak target features, and complex backgrounds prevalent in infrared images. Existing methods often struggle to effectively capture the subtle edge features of targets and suppress [...] Read more.
Infrared small target detection (IRSTD) remains a critical yet challenging task due to the inherent low signal-to-noise ratio, weak target features, and complex backgrounds prevalent in infrared images. Existing methods often struggle to effectively capture the subtle edge features of targets and suppress background clutter simultaneously. To address these limitations, this study proposed a novel Multi-directional Learnable Edge-assisted Dense Nested Attention Network (MLEDNet). Firstly, we propose a multi-directional learnable edge extraction module (MLEEM), which is designed to capture rich directional edge information. The extracted multi-directional edge features are hierarchically integrated into the dense nested attention module (DNAM) to significantly enhance the model’s capability in discerning the crucial edge features of infrared small targets. Then, we design a feature fusion module guided by residual channel spatial attention (ResCSAM-FFM). This module leverages spatio-channel contextual cues to intelligently fuse features across different levels output by the DNAM, effectively enhancing target representation while robustly suppressing complex background interferences. By combining the MLEEM and the ResCSAM-FFM within a dense nested attention framework, we present a new model named MLEDNet. Extensive experiments conducted on benchmark datasets NUDT-SIRST and NUAA-SIRST demonstrate that the proposed MLEDNet achieves superior performance compared to state-of-the-art methods. Full article
(This article belongs to the Special Issue Deep Learning-Based Object Detection/Classification)
Show Figures

Figure 1

21 pages, 922 KB  
Article
Research on Agricultural Meteorological Disaster Event Extraction Method Based on Character–Word Fusion
by Minghui Qiu, Lihua Jiang, Nengfu Xie, Huanping Wu, Ying Chen and Yonglei Li
Agronomy 2025, 15(9), 2135; https://doi.org/10.3390/agronomy15092135 - 5 Sep 2025
Viewed by 268
Abstract
Meteorological disasters are significant factors that impact agricultural production. Given the vast volume of online agrometeorological disaster information, accurately and automatically extracting essential details—such as the time, location, and extent of damage—is crucial for understanding disaster mechanisms and evolution, as well as for [...] Read more.
Meteorological disasters are significant factors that impact agricultural production. Given the vast volume of online agrometeorological disaster information, accurately and automatically extracting essential details—such as the time, location, and extent of damage—is crucial for understanding disaster mechanisms and evolution, as well as for enhancing disaster prevention capabilities. This paper constructs a comprehensive dataset of agrometeorological disasters in China, providing a robust data foundation and strong support for event extraction tasks. Additionally, we propose a novel model named character and word embedding fusion-based GCN network (CWEF-GCN). This integration of character- and word-level information enhances the model’s ability to better understand and represent text, effectively addressing the challenges of multi-events and argument overlaps in the event extraction process. The experimental results on the agrometeorological disaster dataset indicate that the F1 score of the proposed model is 81.66% for trigger classification and 63.31% for argument classification. Following the extraction of batch agricultural meteorological disaster events, this study analyzes the triggering mechanisms, damage patterns, and disaster response strategies across various disaster types using the extracted event. The findings offer actionable decision-making support for research on agricultural disaster prevention and mitigation. Full article
Show Figures

Figure 1

20 pages, 2226 KB  
Article
RST-Net: A Semantic Segmentation Network for Remote Sensing Images Based on a Dual-Branch Encoder Structure
by Na Yang, Chuanzhao Tian, Xingfa Gu, Yanting Zhang, Xuewen Li and Feng Zhang
Sensors 2025, 25(17), 5531; https://doi.org/10.3390/s25175531 - 5 Sep 2025
Viewed by 594
Abstract
High-resolution remote sensing images often suffer from inadequate fusion between global and local features, leading to the loss of long-range dependencies and blurred spatial details, while also exhibiting limited adaptability to multi-scale object segmentation. To overcome these limitations, this study proposes RST-Net, a [...] Read more.
High-resolution remote sensing images often suffer from inadequate fusion between global and local features, leading to the loss of long-range dependencies and blurred spatial details, while also exhibiting limited adaptability to multi-scale object segmentation. To overcome these limitations, this study proposes RST-Net, a semantic segmentation network featuring a dual-branch encoder structure. The encoder integrates a ResNeXt-50-based CNN branch for extracting local spatial features and a Shunted Transformer (ST) branch for capturing global contextual information. To further enhance multi-scale representation, the multi-scale feature enhancement module (MSFEM) is embedded in the CNN branch, leveraging atrous and depthwise separable convolutions to dynamically aggregate features. Additionally, the residual dynamic feature fusion (RDFF) module is incorporated into skip connections to improve interactions between encoder and decoder features. Experiments on the Vaihingen and Potsdam datasets show that RST-Net achieves promising performance, with MIoU scores of 77.04% and 79.56%, respectively, validating its effectiveness in semantic segmentation tasks. Full article
Show Figures

Figure 1

18 pages, 15698 KB  
Article
MDEM: A Multi-Scale Damage Enhancement MambaOut for Pavement Damage Classification
by Shizheng Zhang, Kunpeng Wang, Pu Li, Min Huang and Jianxiang Guo
Sensors 2025, 25(17), 5522; https://doi.org/10.3390/s25175522 - 4 Sep 2025
Viewed by 653
Abstract
Pavement damage classification is crucial for road maintenance and driving safety. However, restricted to the varying scales, irregular shapes, small area ratios, and frequent overlap with background noise, traditional methods struggle to achieve accurate recognition. To address these challenges, a novel pavement damage [...] Read more.
Pavement damage classification is crucial for road maintenance and driving safety. However, restricted to the varying scales, irregular shapes, small area ratios, and frequent overlap with background noise, traditional methods struggle to achieve accurate recognition. To address these challenges, a novel pavement damage classification model is designed based on the MambaOut named Multi-scale Damage Enhancement MambaOut (MDEM). The model incorporates two key modules to improve damage classification performance. The Multi-scale Dynamic Feature Fusion Block (MDFF) adaptively integrates multi-scale information to enhance feature extraction, effectively distinguishing visually similar cracks at different scales. The Damage Detail Enhancement Block (DDE) emphasizes fine structural details while suppressing background interference, thereby improving the representation of small-scale damage regions. Experiments were conducted on multiple datasets, including CQU-BPMDD, CQU-BPDD, and Crack500-PDD. On the CQU-BPMDD dataset, MDEM outperformed the baseline model with improvements of 2.01% in accuracy, 2.64% in precision, 2.7% in F1-score, and 4.2% in AUC. The extensive experimental results demonstrate that MDEM significantly surpasses MambaOut and other comparable methods in pavement damage classification tasks. It effectively addresses challenges such as varying scales, irregular shapes, small damage areas, and background noise, enhancing inspection accuracy in real-world road maintenance. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

20 pages, 8561 KB  
Article
LCW-YOLO: An Explainable Computer Vision Model for Small Object Detection in Drone Images
by Dan Liao, Rengui Bi, Yubi Zheng, Cheng Hua, Liangqing Huang, Xiaowen Tian and Bolin Liao
Appl. Sci. 2025, 15(17), 9730; https://doi.org/10.3390/app15179730 - 4 Sep 2025
Viewed by 624
Abstract
Small targets in drone imagery are often difficult to accurately locate and identify due to scale imbalance and limitations, such as pixel representation and dynamic environmental interference, and the balance between detection accuracy and resource consumption of the model also poses challenges. Therefore, [...] Read more.
Small targets in drone imagery are often difficult to accurately locate and identify due to scale imbalance and limitations, such as pixel representation and dynamic environmental interference, and the balance between detection accuracy and resource consumption of the model also poses challenges. Therefore, we propose an interpretable computer vision framework based on YOLOv12m, called LCW-YOLO. First, we adopt multi-scale heterogeneous convolutional kernels to improve the lightweight channel-level and spatial attention combined context (LA2C2f) structure, enhancing spatial perception capabilities while reducing model computational load. Second, to enhance feature fusion capabilities, we propose the Convolutional Attention Integration Module (CAIM), enabling the fusion of original features across channels, spatial dimensions, and layers, thereby strengthening contextual attention. Finally, the model incorporates Wise-IoU (WIoU) v3, which dynamically allocates loss weights for detected objects. This allows the model to adjust its focus on samples of average quality during training based on object difficulty, thereby improving the model’s generalization capabilities. According to experimental results, LCW-YOLO eliminates 0.4 M parameters and improves mAP@0.5 by 3.3% on the VisDrone2019 dataset when compared to YOLOv12m. And the model improves mAP@0.5 by 1.9% on the UAVVaste dataset. In the task of identifying small objects with drones, LCW-YOLO, as an explainable AI (XAI) model, provides visual detection results and effectively balances accuracy, lightweight design, and generalization capabilities. Full article
(This article belongs to the Special Issue Explainable Artificial Intelligence Technology and Its Applications)
Show Figures

Figure 1

22 pages, 11486 KB  
Article
RAP-Net: A Region Affinity Propagation-Guided Semantic Segmentation Network for Plateau Karst Landform Remote Sensing Imagery
by Dongsheng Zhong, Lingbo Cai, Shaoda Li, Wei Wang, Yijing Zhu, Yaning Liu and Ronghao Yang
Remote Sens. 2025, 17(17), 3082; https://doi.org/10.3390/rs17173082 - 4 Sep 2025
Viewed by 485
Abstract
Karst rocky desertification on the Qinghai–Tibet Plateau poses a severe threat to the region’s fragile ecosystem. Accordingly, the rapid and accurate delineation of plateau karst landforms is essential for monitoring ecological degradation and guiding restoration strategies. However, automatic recognition of these landforms in [...] Read more.
Karst rocky desertification on the Qinghai–Tibet Plateau poses a severe threat to the region’s fragile ecosystem. Accordingly, the rapid and accurate delineation of plateau karst landforms is essential for monitoring ecological degradation and guiding restoration strategies. However, automatic recognition of these landforms in remote sensing imagery is hindered by challenges such as blurred boundaries, fragmented targets, and poor intra-region consistency. To address these issues, we propose the Region Affinity Propagation Network (RAP-Net). This framework enhances intra-region consistency, edge sensitivity, and multi-scale context fusion through its core modules: Region Affinity Propagation (RAP), High-Frequency Multi-Scale Attention (HFMSA), and Global–Local Cross Attention (GLCA). In addition, we constructed the Plateau Karst Landform Dataset (PKLD), a high-resolution remote sensing dataset specifically tailored for this task, which provides a standardized benchmark for future studies. On the PKLD, RAP-Net surpasses eight state-of-the-art methods, achieving 3.69–10.31% higher IoU and 3.88–14.28% higher Recall, thereby demonstrating significant improvements in boundary delineation and structural completeness. Moreover, in a cross-regional generalization test on the Mount Genyen area, RAP-Net—trained solely on PKLD without fine-tuning—achieved 2.38% and 1.94% higher IoU and F1-scores, respectively, than the Swin Transformer, confirming its robustness and generalizability in complex, unseen environments. Full article
(This article belongs to the Section Remote Sensing in Geology, Geomorphology and Hydrology)
Show Figures

Figure 1

18 pages, 1767 KB  
Article
A Blind Few-Shot Learning for Multimodal-Biological Signals with Fractal Dimension Estimation
by Nadeem Ullah, Seung Gu Kim, Jung Soo Kim, Min Su Jeong and Kang Ryoung Park
Fractal Fract. 2025, 9(9), 585; https://doi.org/10.3390/fractalfract9090585 - 3 Sep 2025
Viewed by 297
Abstract
Improving the decoding accuracy of biological signals has been a research focus for decades to advance health, automation, and robotic industries. However, challenges like inter-subject variability, data scarcity, and multifunctional variability cause low decoding accuracy, thus hindering the practical deployment of biological signal [...] Read more.
Improving the decoding accuracy of biological signals has been a research focus for decades to advance health, automation, and robotic industries. However, challenges like inter-subject variability, data scarcity, and multifunctional variability cause low decoding accuracy, thus hindering the practical deployment of biological signal paradigms. This paper proposes a multifunctional biological signals network (Multi-BioSig-Net) that addresses the aforementioned issues by devising a novel blind few-shot learning (FSL) technique to quickly adapt to multiple target domains without needing a pre-trained model. Specifically, our proposed multimodal similarity extractor (MMSE) and self-multiple domain adaptation (SMDA) modules address data scarcity and inter-subject variability issues by exploiting and enhancing the similarity between multimodal samples and quickly adapting the target domains by adaptively adjusting the parameters’ weights and position, respectively. For multifunctional learning, we proposed inter-function discriminator (IFD) that discriminates the classes by extracting inter-class common features and then subtracts them from both classes to avoid false prediction of the proposed model due to overfitting on the common features. Furthermore, we proposed a holistic-local fusion (HLF) module that exploits contextual-detailed features to adapt the scale-varying features across multiple functions. In addition, fractal dimension estimation (FDE) was employed for the classification of left-hand motor imagery (LMI) and right-hand motor imagery (RMI), confirming that proposed method can effectively extract the discriminative features for this task. The effectiveness of our proposed algorithm was assessed quantitatively and statistically against competent state-of-the-art (SOTA) algorithms utilizing three public datasets, demonstrating that our proposed algorithm outperformed SOTA algorithms. Full article
Show Figures

Figure 1

Back to TopTop