Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

Search Results (323)

Search Parameters:
Keywords = three-channel architecture

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 6014 KB  
Article
Enhancing Instance Segmentation in Agriculture: An Optimized YOLOv8 Solution
by Qiaolong Wang, Dongshun Chen, Wenfei Feng, Liang Sun and Gaohong Yu
Sensors 2025, 25(17), 5506; https://doi.org/10.3390/s25175506 - 4 Sep 2025
Abstract
To address the limitations of traditional segmentation algorithms in processing complex agricultural scenes, this paper proposes an improved YOLOv8n-seg model. Building upon the original three detection layers, we introduce a dedicated layer for small object detection, which significantly enhances the detection accuracy of [...] Read more.
To address the limitations of traditional segmentation algorithms in processing complex agricultural scenes, this paper proposes an improved YOLOv8n-seg model. Building upon the original three detection layers, we introduce a dedicated layer for small object detection, which significantly enhances the detection accuracy of small targets (e.g., people) after processing images through fourfold downsampling. In the neck network, we replace the C2f module with our proposed C2f_CPCA module, which incorporates a channel prior attention mechanism (CPCA). This mechanism dynamically adjusts attention weights across channels and spatial dimensions to effectively capture relationships between different spatial scales, thereby improving feature extraction and recognition capabilities while maintaining low computational complexity. Finally, we propose a C3RFEM module based on the RFEM architecture and integrate it into the main network. This module combines dilated convolutions and weighted layers to enhance feature extraction capabilities across different receptive field ranges. Experimental results demonstrated that the improved model achieved 1.4% and 4.0% increases in precision and recall rates on private datasets, respectively, with mAP@0.5 and mAP@0.5:0.95 metrics improved by 3.0% and 3.5%, respectively. In comparative evaluations with instance segmentation algorithms such as the YOLOv5 series, YOLOv7, YOLOv8n, YOLOv9t, YOLOv10n, YOLOv10s, Mask R-CNN, and Mask2Former, our model achieved an optimal balance between computational efficiency and detection performance. This demonstrates its potential for the research and development of small intelligent precision operation technology and equipment. Full article
(This article belongs to the Section Smart Agriculture)
Show Figures

Figure 1

24 pages, 4455 KB  
Article
HDAMNet: Hierarchical Dilated Adaptive Mamba Network for Accurate Cloud Detection in Satellite Imagery
by Yongcong Wang, Yunxin Li, Xubing Yang, Rui Jiang and Li Zhang
Remote Sens. 2025, 17(17), 2992; https://doi.org/10.3390/rs17172992 - 28 Aug 2025
Viewed by 350
Abstract
Cloud detection is one of the primary challenges in preprocessing high-resolution remote sensing imagery, the accuracy of which is severely constrained by the multi-scale and complex morphological characteristics of clouds. Many approaches have been proposed to detect cloud. However, these methods still face [...] Read more.
Cloud detection is one of the primary challenges in preprocessing high-resolution remote sensing imagery, the accuracy of which is severely constrained by the multi-scale and complex morphological characteristics of clouds. Many approaches have been proposed to detect cloud. However, these methods still face significant challenges, particularly in handling the complexities of multi-scale cloud clusters and reliably distinguishing clouds from snow, ice and complex cloud shadows. To overcome these challenges, this paper proposes a novel cloud detection network based on the state space model (SSM), termed the Hierarchical Dilated Adaptive Mamba Network (HDAMNet). This network utilizes an encoder–decoder architecture, significantly expanding the receptive field and improving the capture of fine-grained cloud boundaries by introducing the Hierarchical Dilated Cross Scan (HDCS) mechanism in the encoder module. The multi-resolution adaptive feature extraction (MRAFE) integrates multi-scale semantic information to reduce channel confusion and emphasize essential features effectively. The Layer-wise Adaptive Attention (LAA) mechanism adaptively recalibrates features at skip connections, balancing fine-grained boundaries with global semantic information. On three public cloud detection datasets, HDAMNet achieves state-of-the-art performance across key evaluation metrics. Particularly noteworthy is its superior performance in identifying small-scale cloud clusters, delineating complex cloud–shadow boundaries, and mitigating interference from snow and ice. Full article
Show Figures

Figure 1

16 pages, 1205 KB  
Article
Design and Simulation of Cross-Medium Two-Hop Relaying Free-Space Optical Communication System Based on Multiple Diversity and Multiplexing Technologies
by Min Guo, Pengxiang Wang and Yan Wu
Photonics 2025, 12(9), 867; https://doi.org/10.3390/photonics12090867 - 28 Aug 2025
Viewed by 259
Abstract
To address the issues of link mismatch and channel impairment in wireless optical communication across atmospheric-oceanic media, this paper proposes a two-hop relay transmission architecture based on the multiple-input multiple-output (MIMO)-enhanced multi-level hybrid multiplexing. The system implements decode-and-forward operations via maritime buoy/ship relays, [...] Read more.
To address the issues of link mismatch and channel impairment in wireless optical communication across atmospheric-oceanic media, this paper proposes a two-hop relay transmission architecture based on the multiple-input multiple-output (MIMO)-enhanced multi-level hybrid multiplexing. The system implements decode-and-forward operations via maritime buoy/ship relays, achieving physical layer isolation between atmospheric and oceanic channels. The transmitter employs coherent orthogonal frequency division multiplexing technology with quadrature amplitude modulation to achieve frequency division multiplexing of baseband signals, combines with orthogonal polarization modulation to generate polarization-multiplexed signal beams, and finally realizes multi-dimensional signal transmission through MIMO spatial diversity. To cope with cross-medium environmental interference, a composite channel model is established, which includes atmospheric turbulence (Gamma–Gamma model), rain attenuation, and oceanic chlorophyll absorption and scattering effects. Simulation results show that the multi-level hybrid multiplexing method can significantly improve the data transmission rate of the system. Since the system adopts three channels of polarization-state data, the data transmission rate is increased by 200%; the two-hop relay method can effectively improve the communication performance of cross-medium optical communication and fundamentally solve the problem of light transmission in cross-medium planes; the use of MIMO technology has a compensating effect on the impacts of both atmospheric and marine environments, and as the number of light beams increases, the system performance can be further improved. This research provides technical implementation schemes and reference data for the design of high-capacity optical communication systems across air-sea media. Full article
(This article belongs to the Special Issue Emerging Technologies for 6G Space Optical Communication Networks)
Show Figures

Figure 1

18 pages, 6467 KB  
Article
State-Space Model Meets Linear Attention: A Hybrid Architecture for Internal Wave Segmentation
by Zhijie An, Zhao Li, Saheya Barintag, Hongyu Zhao, Yanqing Yao, Licheng Jiao and Maoguo Gong
Remote Sens. 2025, 17(17), 2969; https://doi.org/10.3390/rs17172969 - 27 Aug 2025
Viewed by 455
Abstract
Internal waves (IWs) play a crucial role in the transport of energy and matter within the ocean while also posing significant risks to marine engineering, navigation, and underwater communication systems. Consequently, effective segmentation methods are essential for mitigating their adverse impacts and minimizing [...] Read more.
Internal waves (IWs) play a crucial role in the transport of energy and matter within the ocean while also posing significant risks to marine engineering, navigation, and underwater communication systems. Consequently, effective segmentation methods are essential for mitigating their adverse impacts and minimizing associated hazards. A promising strategy involves applying remote sensing image segmentation techniques to accurately identify IWs, thereby enabling predictions of their propagation velocity and direction. However, current IWs segmentation models struggle to balance computational efficiency and segmentation accuracy, often resulting in either excessive computational costs or inadequate performance. Motivated by recent developments in the Mamba2 architecture, this paper introduces the state-space model meets linear attention (SMLA), a novel segmentation framework specifically designed for IWs. The proposed hybrid architecture effectively integrates three key components: a feature-aware serialization (FAS) block to efficiently convert spatial features into sequences; a state-space model with linear attention (SSM-LA) block that synergizes a state-space model with linear attention for comprehensive feature extraction; and a decoder driven by hierarchical fusion and upsampling, which performs channel alignment and scale unification across multi-level features to ensure high-fidelity spatial detail recovery. Experiments conducted on a dataset of 484 synthetic-aperture radar (SAR) images containing IWs from the South China Sea achieved a mean Intersection over Union (MIoU) of 74.3%, surpassing competing methods evaluated on the same dataset. These results demonstrate the superior effectiveness of SMLA in extracting features of IWs from SAR imagery. Full article
(This article belongs to the Special Issue Advancements of Vision-Language Models (VLMs) in Remote Sensing)
Show Figures

Figure 1

26 pages, 2700 KB  
Article
An Enhanced MIBKA-CNN-BiLSTM Model for Fake Information Detection
by Sining Zhu, Guangyu Mu, Jie Ma and Xiurong Li
Biomimetics 2025, 10(9), 562; https://doi.org/10.3390/biomimetics10090562 - 23 Aug 2025
Viewed by 312
Abstract
The complexity of fake information and the inefficiency of parameter optimization in detection models present dual challenges for current detection technologies. Therefore, this paper proposes a hybrid detection model named MIBKA-CNN-BiLSTM, which significantly improves detection accuracy and efficiency through a triple-strategy enhancement of [...] Read more.
The complexity of fake information and the inefficiency of parameter optimization in detection models present dual challenges for current detection technologies. Therefore, this paper proposes a hybrid detection model named MIBKA-CNN-BiLSTM, which significantly improves detection accuracy and efficiency through a triple-strategy enhancement of the Black Kite Optimization Algorithm (MIBKA) and an optimized dual-channel deep learning architecture. First, three improvements are introduced in the MIBKA. The population initialization process is restructured using circle chaotic mapping to enhance parameter space coverage. The conventional random perturbation is replaced by a random-to-elite differential mutation strategy (DE/rand-to-best/1) to balance global exploration and local exploitation. Moreover, a logarithmic spiral opposition-based learning (LSOBL) mechanism is integrated to dynamically explore the opposition solution space. Second, a CNN-BiLSTM dual-channel feature extraction network is constructed, with hyperparameters such as the number of convolutional kernels and LSTM units optimized by MIBKA to enable adaptive model structure alignment with task requirements. Finally, a high-quality fake information dataset is created based on social media platforms, including CCTV. The experimental results show that our model achieves the highest accuracy on the self-built dataset, which is 3.11% higher than the optimal hybrid model. Additionally, on the Weibo21 dataset, our model’s accuracy and F1-score increased by 1.52% and 1.71%, respectively, compared to the average values of all baseline models. These findings offer a practical and effective approach for detecting lightweight and robust false information. Full article
(This article belongs to the Special Issue Nature-Inspired Metaheuristic Optimization Algorithms 2025)
Show Figures

Figure 1

16 pages, 1786 KB  
Article
Enhanced SSVEP Bionic Spelling via xLSTM-Based Deep Learning with Spatial Attention and Filter Bank Techniques
by Liuyuan Dong, Chengzhi Xu, Ruizhen Xie, Xuyang Wang, Wanli Yang and Yimeng Li
Biomimetics 2025, 10(8), 554; https://doi.org/10.3390/biomimetics10080554 - 21 Aug 2025
Viewed by 380
Abstract
Steady-State Visual Evoked Potentials (SSVEPs) have emerged as an efficient means of interaction in brain–computer interfaces (BCIs), achieving bioinspired efficient language output for individuals with aphasia. Addressing the underutilization of frequency information of SSVEPs and redundant computation by existing transformer-based deep learning methods, [...] Read more.
Steady-State Visual Evoked Potentials (SSVEPs) have emerged as an efficient means of interaction in brain–computer interfaces (BCIs), achieving bioinspired efficient language output for individuals with aphasia. Addressing the underutilization of frequency information of SSVEPs and redundant computation by existing transformer-based deep learning methods, this paper analyzes signals from both the time and frequency domains, proposing a stacked encoder–decoder (SED) network architecture based on an xLSTM model and spatial attention mechanism, termed SED-xLSTM, which firstly applies xLSTM to the SSVEP speller field. This model takes the low-channel spectrogram as input and employs the filter bank technique to make full use of harmonic information. By leveraging a gating mechanism, SED-xLSTM effectively extracts and fuses high-dimensional spatial-channel semantic features from SSVEP signals. Experimental results on three public datasets demonstrate the superior performance of SED-xLSTM in terms of classification accuracy and information transfer rate, particularly outperforming existing methods under cross-validation across various temporal scales. Full article
(This article belongs to the Special Issue Exploration of Bioinspired Computer Vision and Pattern Recognition)
Show Figures

Figure 1

26 pages, 2611 KB  
Article
Multi-Channel Graph Convolutional Network for Evaluating Innovation Capability Toward Sustainable Seed Enterprises
by Shanshan Tang, Kaiyi Wang, Feng Yang and Shouhui Pan
Sustainability 2025, 17(16), 7522; https://doi.org/10.3390/su17167522 - 20 Aug 2025
Viewed by 416
Abstract
The innovation capability of seed enterprises reflects their core competitiveness and serves as a vital foundation for sustainable agricultural development and modernization. Therefore, evaluating this capability is of great importance. However, existing evaluation methods primarily focus on internal enterprise attributes, overlooking the complex [...] Read more.
The innovation capability of seed enterprises reflects their core competitiveness and serves as a vital foundation for sustainable agricultural development and modernization. Therefore, evaluating this capability is of great importance. However, existing evaluation methods primarily focus on internal enterprise attributes, overlooking the complex inter-enterprise relationships and lacking sufficient feature fusion capabilities to capture latent information. To address these limitations, this paper proposes a Multi-Channel Graph Convolutional Network (MGCN) model that integrates enterprise attributes with three types of relational graphs. The model adopts a multi-channel architecture for feature extraction and employs a gated attention mechanism for cross-graph feature fusion, jointly considering node features and relation information to improve prediction accuracy. Experimental results demonstrate that MGCN achieves an average accuracy of 83.59% under five-fold cross-validation, outperforming several mainstream models such as Random Forest and traditional GCN. Case studies further reveal that MGCN not only captures key features of individual enterprises but also leverages features and label distribution from neighboring enterprises, facilitating more context-aware classification decisions. In conclusion, the MGCN model provides an effective method for the intelligent evaluation of innovation capability in seed enterprises and supports the formulation of sustainable strategic plans at both the national and enterprise level. Full article
Show Figures

Figure 1

38 pages, 6706 KB  
Article
Intelligent Method for Generating Criminal Community Influence Risk Parameters Using Neural Networks and Regional Economic Analysis
by Serhii Vladov, Lyubomyr Chyrun, Eduard Muzychuk, Victoria Vysotska, Vasyl Lytvyn, Tetiana Rekunenko and Andriy Basko
Algorithms 2025, 18(8), 523; https://doi.org/10.3390/a18080523 - 18 Aug 2025
Viewed by 293
Abstract
This article develops an innovative and intelligent method for analysing the criminal community’s influence on risk-forming parameters based on an analysis of regional economic processes. The research motivation was the need to create an intelligent method for quantitative assessment and risk control arising [...] Read more.
This article develops an innovative and intelligent method for analysing the criminal community’s influence on risk-forming parameters based on an analysis of regional economic processes. The research motivation was the need to create an intelligent method for quantitative assessment and risk control arising from the interaction between regional economic processes and criminal activity. The method includes a three-level mathematical model in which the economic activity dynamics are described by a modified logistic equation, taking into account the criminal activity’s negative impact and feedback through the integral risk. The criminal activity itself is modelled by a similar logistic equation, taking into account the economic base. The risk parameter accumulates the direct impact and delayed effects through the memory core. To numerically solve the spatio-temporal optimal control problem, a neural network based on the convolutional architecture was developed: two successive convolutional layers (N1 with 3 × 3 filters and N2 with 3 × 3 filters) extract local features, after which two 1 × 1 convolutional layers (FC1 and FC2) form a three-channel output corresponding to the control actions UE, UC, and UI. The loss function combines the supervised component and the residual terms of the differential equations, which ensures the satisfaction of physical constraints. The computational experiment showed the high accuracy of the model: accuracy is 0.9907, precision is 0.9842, recall is 0.9983, and F1-score is 0.9912, with a minimum residual loss of 0.0093 and superiority over alternative architectures in key metrics (MSE is 0.0124, IoU is 0.74, and Dice is 0.83). Full article
(This article belongs to the Section Analysis of Algorithms and Complexity Theory)
Show Figures

Figure 1

32 pages, 7175 KB  
Article
VisFactory: Adaptive Multimodal Digital Twin with Integrated Visual-Haptic-Auditory Analytics for Industry 4.0 Engineering Education
by Tsung-Ching Lin, Cheng-Nan Chiu, Po-Tong Wang and Li-Der Fang
Multimedia 2025, 1(1), 3; https://doi.org/10.3390/multimedia1010003 - 18 Aug 2025
Viewed by 473
Abstract
Industry 4.0 has intensified the skills gap in industrial automation education, with graduates requiring extended on boarding periods and supplementary training investments averaging USD 11,500 per engineer. This paper introduces VisFactory, a multimedia learning system that extends the cognitive theory of multimedia learning [...] Read more.
Industry 4.0 has intensified the skills gap in industrial automation education, with graduates requiring extended on boarding periods and supplementary training investments averaging USD 11,500 per engineer. This paper introduces VisFactory, a multimedia learning system that extends the cognitive theory of multimedia learning by incorporating haptic feedback as a third processing channel alongside visual and auditory modalities. The system integrates a digital twin architecture with ultra-low latency synchronization (12.3 ms) across all sensory channels, a dynamic feedback orchestration algorithm that distributes information optimally across modalities, and a tripartite student model that continuously calibrates instruction parameters. We evaluated the system through a controlled experiment with 127 engineering students randomly assigned to experimental and control groups, with assessments conducted immediately and at three-month and six-month intervals. VisFactory significantly enhanced learning outcomes across multiple dimensions: 37% reduction in time to mastery (t(125) = 11.83, p < 0.001, d = 2.11), skill acquisition increased from 28% to 85% (ηp2=0.54), and 28% higher knowledge retention after six months. The multimodal approach demonstrated differential effectiveness across learning tasks, with haptic feedback providing the most significant benefit for procedural skills (52% error reduction) and visual–auditory integration proving most effective for conceptual understanding (49% improvement). The adaptive modality orchestration reduced cognitive load by 43% compared to unimodal interfaces. This research advances multimedia learning theory by validating tri-modal integration effectiveness and establishing quantitative benchmarks for sensory channel synchronization. The findings provide a theoretical framework and implementation guidelines for optimizing multimedia learning environments for complex skill development in technical domains. Full article
Show Figures

Figure 1

26 pages, 7726 KB  
Article
Multi-Branch Channel-Gated Swin Network for Wetland Hyperspectral Image Classification
by Ruopu Liu, Jie Zhao, Shufang Tian, Guohao Li and Jingshu Chen
Remote Sens. 2025, 17(16), 2862; https://doi.org/10.3390/rs17162862 - 17 Aug 2025
Viewed by 407
Abstract
Hyperspectral classification of wetland environments remains challenging due to high spectral similarity, class imbalance, and blurred boundaries. To address these issues, we propose a novel Multi-Branch Channel-Gated Swin Transformer network (MBCG-SwinNet). In contrast to previous CNN-based designs, our model introduces a Swin Transformer [...] Read more.
Hyperspectral classification of wetland environments remains challenging due to high spectral similarity, class imbalance, and blurred boundaries. To address these issues, we propose a novel Multi-Branch Channel-Gated Swin Transformer network (MBCG-SwinNet). In contrast to previous CNN-based designs, our model introduces a Swin Transformer spectral branch to enhance global contextual modeling, enabling improved spectral discrimination. To effectively fuse spatial and spectral features, we design a residual feature interaction chain comprising a Residual Spatial Fusion (RSF) module, a channel-wise gating mechanism, and a multi-scale feature fusion (MFF) module, which together enhance spatial adaptivity and feature integration. Additionally, a DenseCRF-based post-processing step is employed to refine classification boundaries and suppress salt-and-pepper noise. Experimental results on three UAV-based hyperspectral wetland datasets from the Yellow River Delta (Shandong, China)—NC12, NC13, and NC16—demonstrate that MBCG-SwinNet achieves superior classification performance, with overall accuracies of 97.62%, 82.37%, and 97.32%, respectively—surpassing state-of-the-art methods. The proposed architecture offers a robust and scalable solution for hyperspectral image classification in complex ecological settings. Full article
Show Figures

Figure 1

22 pages, 8647 KB  
Article
A High-Performance Ka-Band Cylindrical Conformal Transceiver Phased Array with Full-Azimuth Scanning Capability
by Weiwei Liu, Shiqiao Zhang, Anxue Zhang and Wenchao Chen
Appl. Sci. 2025, 15(16), 8982; https://doi.org/10.3390/app15168982 - 14 Aug 2025
Viewed by 249
Abstract
This paper presents a Ka-band cylindrical conformal transceiver active phased array (CCTAPA) with a full-azimuth scanning gain fluctuation of 0.8 dB and low power consumption. The array comprises 20 panels of 4 × 4 antenna elements, RF beam-control circuits, a Wilkinson power divider [...] Read more.
This paper presents a Ka-band cylindrical conformal transceiver active phased array (CCTAPA) with a full-azimuth scanning gain fluctuation of 0.8 dB and low power consumption. The array comprises 20 panels of 4 × 4 antenna elements, RF beam-control circuits, a Wilkinson power divider network, and frequency converters. The proposed three-subarray architecture enables ±9° beam scanning with minimal gain degradation. By dynamically switching subarrays and transceiver channels across azimuthal directions, the array achieves full 360° coverage with low gain fluctuation and power consumption. Fabrication and testing demonstrate a gain fluctuation of 0.8 dB, equivalent isotropically radiated power (EIRP) between 50.6 and 51.3 dBm, and a gain-to-noise-temperature ratio (G/T) ranging from −8 dB/K to −8.5 dB/K at 28.5 GHz. The RF power consumption remains below 8.73 W during full-azimuth scanning. This design is particularly suitable for airborne platforms requiring full-azimuth coverage with stringent power budgets. Full article
Show Figures

Figure 1

25 pages, 7900 KB  
Article
Multi-Label Disease Detection in Chest X-Ray Imaging Using a Fine-Tuned ConvNeXtV2 with a Customized Classifier
by Kangzhe Xiong, Yuyun Tu, Xinping Rao, Xiang Zou and Yingkui Du
Informatics 2025, 12(3), 80; https://doi.org/10.3390/informatics12030080 - 14 Aug 2025
Viewed by 656
Abstract
Deep-learning-based multiple label chest X-ray classification has achieved significant success, but existing models still have three main issues: fixed-scale convolutions fail to capture both large and small lesions, standard pooling is lacking in the lack of attention to important regions, and linear classification [...] Read more.
Deep-learning-based multiple label chest X-ray classification has achieved significant success, but existing models still have three main issues: fixed-scale convolutions fail to capture both large and small lesions, standard pooling is lacking in the lack of attention to important regions, and linear classification lacks the capacity to model complex dependency between features. To circumvent these obstacles, we propose CONVFCMAE, a lightweight yet powerful framework that is built on a backbone that is partially frozen (77.08 % of the initial layers are fixed) in order to preserve complex, multi-scale features while decreasing the number of trainable parameters. Our architecture adds (1) an intelligent global pooling module that is learnable, with 1×1 convolutions that are dynamically weighted by their spatial location, and (2) a multi-head attention block that is dedicated to channel re-calibration, along with (3) a two-layer MLP that has been enhanced with ReLU, batch normalization, and dropout. This module is used to enhance the non-linearity of the feature space. To further reduce the noise associated with labels and the imbalance in class distribution inherent to the NIH ChestXray14 dataset, we utilize a combined loss that combines BCEWithLogits and Focal Loss as well as extensive data augmentation. On ChestXray14, the average ROC–AUC of CONVFCMAE is 0.852, which is 3.97 percent greater than the state of the art. Ablation experiments demonstrate the individual and collective effectiveness of each component. Grad-CAM visualizations have a superior capacity to localize the pathological regions, and this increases the interpretability of the model. Overall, CONVFCMAE provides a practical, generalizable solution to the problem of extracting features from medical images in a practical manner. Full article
(This article belongs to the Section Medical and Clinical Informatics)
Show Figures

Figure 1

22 pages, 15242 KB  
Article
A Modality Alignment and Fusion-Based Method for Around-the-Clock Remote Sensing Object Detection
by Yongjun Qi, Shaohua Yang, Jiahao Chen, Meng Zhang, Jie Zhu, Xin Liu and Hongxing Zheng
Sensors 2025, 25(16), 4964; https://doi.org/10.3390/s25164964 - 11 Aug 2025
Viewed by 471
Abstract
Cross-modal remote sensing object detection holds significant potential for around-the-clock applications. However, the modality differences between cross-modal data and the degradation of feature quality under adverse weather conditions limit detection performance. To address these challenges, this paper presents a novel cross-modal remote sensing [...] Read more.
Cross-modal remote sensing object detection holds significant potential for around-the-clock applications. However, the modality differences between cross-modal data and the degradation of feature quality under adverse weather conditions limit detection performance. To address these challenges, this paper presents a novel cross-modal remote sensing object detection framework designed to overcome two critical challenges in around-the-clock applications: (1) significant modality disparities between visible light, infrared, and synthetic aperture radar data, and (2) severe feature degradation under adverse weather conditions including fog, and nighttime scenarios. Our primary contributions are as follows: First, we develop a multi-scale feature extraction module that employs a hierarchical convolutional architecture to capture both fine-grained details and contextual information, effectively compensating for missing or blurred features in degraded visible-light images. Second, we introduce an innovative feature interaction module that utilizes cross-attention mechanisms to establish long-range dependencies across modalities while dynamically suppressing noise interference through adaptive feature selection. Third, we propose a feature correction fusion module that performs spatial alignment of object boundaries and channel-wise optimization of global feature consistency, enabling robust fusion of complementary information from different modalities. The proposed framework is validated on visible light, infrared, and SAR modalities. Extensive experiments on three challenging datasets (LLVIP, OGSOD, and Drone Vehicle) demonstrate our framework’s superior performance, achieving state-of-the-art mean average precision scores of 66.3%, 58.6%, and 71.7%, respectively, representing significant improvements over existing methods in scenarios with modality differences or extreme weather conditions. The proposed solution not only advances the technical frontier of cross-modal object detection but also provides practical value for mission-critical applications such as 24/7 surveillance systems, military reconnaissance, and emergency response operations where reliable around-the-clock detection is essential. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

21 pages, 5260 KB  
Article
LapECNet: Laplacian Pyramid Networks for Image Exposure Correction
by Yongchang Li and Jing Jiang
Appl. Sci. 2025, 15(16), 8840; https://doi.org/10.3390/app15168840 - 11 Aug 2025
Viewed by 234
Abstract
Images captured under complex lighting conditions often suffer from local under/ overexposure and detail loss. Existing methods typically process illumination and texture information in a mixed manner, making it difficult to simultaneously achieve precise exposure adjustment and preservation of detail. To address this [...] Read more.
Images captured under complex lighting conditions often suffer from local under/ overexposure and detail loss. Existing methods typically process illumination and texture information in a mixed manner, making it difficult to simultaneously achieve precise exposure adjustment and preservation of detail. To address this challenge, we propose LapECNet, an enhanced Laplacian pyramid network architecture for image exposure correction and detail reconstruction. Specifically, it decomposes the input image into different frequency bands of a Laplacian pyramid, enabling separate handling of illumination adjustment and detail enhancement. The framework first decomposes the image into three feature levels. At each level, we introduce a feature enhancement module that adaptively processes image features across different frequency bands using spatial and channel attention mechanisms. After enhancing the features at each level, we further propose a dynamic aggregation module that learns adaptive weights to hierarchically fuse multi-scale features, achieving context-aware recombination of the enhanced features. Extensive experiments with public benchmarks on the MSEC dataset demonstrated that our method gave improvements of 15.4% in PSNR and 7.2% in SSIM over previous methods. On the LCDP dataset, our method demonstrated improvements of 7.2% in PSNR and 13.9% in SSIM over previous methods. Full article
(This article belongs to the Special Issue Recent Advances in Parallel Computing and Big Data)
Show Figures

Figure 1

12 pages, 806 KB  
Proceeding Paper
Enterococcus faecalis Biofilm: A Clinical and Environmental Hazard
by Bindu Sadanandan and Kavyasree Marabanahalli Yogendraiah
Med. Sci. Forum 2025, 35(1), 5; https://doi.org/10.3390/msf2025035005 - 5 Aug 2025
Viewed by 644
Abstract
This review explores the biofilm architecture and drug resistance of Enterococcus faecalis in clinical and environmental settings. The biofilm in E. faecalis is a heterogeneous, three-dimensional, mushroom-like or multilayered structure, characteristically forming diplococci or short chains interspersed with water channels for nutrient exchange [...] Read more.
This review explores the biofilm architecture and drug resistance of Enterococcus faecalis in clinical and environmental settings. The biofilm in E. faecalis is a heterogeneous, three-dimensional, mushroom-like or multilayered structure, characteristically forming diplococci or short chains interspersed with water channels for nutrient exchange and waste removal. Exopolysaccharides, proteins, lipids, and extracellular DNA create a protective matrix. Persister cells within the biofilm contribute to antibiotic resistance and survival. The heterogeneous architecture of the E. faecalis biofilm contains both dense clusters and loosely packed regions that vary in thickness, ranging from 10 to 100 µm, depending on the environmental conditions. The pathogenicity of the E. faecalis biofilm is mediated through complex interactions between genes and virulence factors such as DNA release, cytolysin, pili, secreted antigen A, and microbial surface components that recognize adhesive matrix molecules, often involving a key protein called enterococcal surface protein (Esp). Clinically, it is implicated in a range of nosocomial infections, including urinary tract infections, endocarditis, and surgical wound infections. The biofilm serves as a nidus for bacterial dissemination and as a reservoir for antimicrobial resistance. The effectiveness of first-line antibiotics (ampicillin, vancomycin, and aminoglycosides) is diminished due to reduced penetration, altered metabolism, increased tolerance, and intrinsic and acquired resistance. Alternative strategies for biofilm disruption, such as combination therapy (ampicillin with aminoglycosides), as well as newer approaches, including antimicrobial peptides, quorum-sensing inhibitors, and biofilm-disrupting agents (DNase or dispersin B), are also being explored to improve treatment outcomes. Environmentally, E. faecalis biofilms contribute to contamination in water systems, food production facilities, and healthcare environments. They persist in harsh conditions, facilitating the spread of multidrug-resistant strains and increasing the risk of transmission to humans and animals. Therefore, understanding the biofilm architecture and drug resistance is essential for developing effective strategies to mitigate their clinical and environmental impact. Full article
(This article belongs to the Proceedings of The 4th International Electronic Conference on Antibiotics)
Show Figures

Figure 1

Back to TopTop