Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (905)

Search Parameters:
Keywords = CBAM

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
24 pages, 5237 KB  
Article
A Precision Weeding System for Cabbage Seedling Stage
by Pei Wang, Weiyue Chen, Qi Niu, Chengsong Li, Yuheng Yang and Hui Li
Agriculture 2026, 16(3), 384; https://doi.org/10.3390/agriculture16030384 - 5 Feb 2026
Abstract
This study developed an integrated vision–actuation system for precision weeding in indoor soil bin environments, with cabbage as a case example. The system integrates lightweight object detection, 3D co-ordinate mapping, path planning, and a three-axis synchronized conveyor-type actuator to enable precise weed identification [...] Read more.
This study developed an integrated vision–actuation system for precision weeding in indoor soil bin environments, with cabbage as a case example. The system integrates lightweight object detection, 3D co-ordinate mapping, path planning, and a three-axis synchronized conveyor-type actuator to enable precise weed identification and automated removal. By integrating ECA and CBAM attention mechanisms into YOLO11, we developed the YOLO11-WeedNet model. This integration significantly enhanced the detection performance for small-scale weeds under complex lighting and cluttered backgrounds. Based on the optimal model performance achieved during experimental evaluation, the model achieved 96.25% precision, 86.49% recall, 91.10% F1-score, and a mean Average Precision (mAP@0.5) of 91.50% calculated across two categories (crop and weed). An RGB-D fusion localization method combined with a protected-area constraint enabled accurate mapping of weed spatial positions. Furthermore, an enhanced Artificial Hummingbird Algorithm (AHA+) was proposed to optimize the execution path and reduce the operating trajectory while maintaining real-time performance. Indoor soil bin tests showed positioning errors of less than 8 mm on the X/Y axes, depth control within ±1 mm on the Z-axis, and an average weeding rate of 88.14%. The system achieved zero contact with cabbage seedlings, with a processing time of 6.88 s per weed. These results demonstrate the feasibility of the proposed system for precise and automated weeding at the cabbage seedling stage. Full article
27 pages, 4033 KB  
Article
DCDW-YOLOv11: An Intelligent Defect-Detection Method for Key Transmission-Line Equipment
by Dezhi Wang, Riqing Song, Minghui Liu, Xingqian Wang, Chengyu Zhang, Ziang Wang and Dongxue Zhao
Sensors 2026, 26(3), 1029; https://doi.org/10.3390/s26031029 - 4 Feb 2026
Viewed by 154
Abstract
The detection of defects in key transmission-line equipment under complex environments often suffers from insufficient accuracy and reliability due to background interference and multi-scale feature variations. To address this issue, this paper proposes an improved defect detection model based on YOLOv11, named DCDW-YOLOv11. [...] Read more.
The detection of defects in key transmission-line equipment under complex environments often suffers from insufficient accuracy and reliability due to background interference and multi-scale feature variations. To address this issue, this paper proposes an improved defect detection model based on YOLOv11, named DCDW-YOLOv11. The model introduces deformable convolution C2f_DCNv3 in the backbone network to enhance adaptability to geometric deformations of targets, and incorporates the convolutional block attention module (CBAM) to highlight defect features while suppressing background interference. In the detection head, a dynamic head structure (DyHead) is adopted to achieve cross-layer multi-scale feature fusion and collaborative perception, along with the WIoU loss function to optimize bounding box regression and sample weight allocation. Experimental results demonstrate that on the transmission-line equipment defect dataset, DCDW-YOLOv11 achieves an accuracy, recall, and mAP of 94.4%, 92.8%, and 96.3%, respectively, representing improvements of 2.8%, 7.0%, and 4.4% over the original YOLOv11, and outperforming other mainstream detection models. The proposed method can provide high-precision and highly reliable defect detection support for intelligent inspection of transmission lines in complex scenarios. Full article
(This article belongs to the Special Issue Image Processing and Analysis for Object Detection: 3rd Edition)
Show Figures

Figure 1

25 pages, 15438 KB  
Article
Day–Night All-Sky Scene Classification with an Attention-Enhanced EfficientNet
by Wuttichai Boonpook, Peerapong Torteeka, Kritanai Torsri, Daroonwan Kamthonkiat, Yumin Tan, Asamaporn Sitthi, Patcharin Kamsing, Chomchanok Arunplod, Utane Sawangwit, Thanachot Ngamcharoensuktavorn and Kijnaphat Suksod
ISPRS Int. J. Geo-Inf. 2026, 15(2), 66; https://doi.org/10.3390/ijgi15020066 - 3 Feb 2026
Viewed by 190
Abstract
All-sky cameras provide continuous hemispherical observations essential for atmospheric monitoring and observatory operations; however, automated classification of sky conditions in tropical environments remains challenging due to strong illumination variability, atmospheric scattering, and overlapping thin-cloud structures. This study proposes EfficientNet-Attention-SPP Multi-scale Network (EASMNet), a [...] Read more.
All-sky cameras provide continuous hemispherical observations essential for atmospheric monitoring and observatory operations; however, automated classification of sky conditions in tropical environments remains challenging due to strong illumination variability, atmospheric scattering, and overlapping thin-cloud structures. This study proposes EfficientNet-Attention-SPP Multi-scale Network (EASMNet), a physics-aware deep learning framework for robust all-sky scene classification using hemispherical imagery acquired at the Thai National Observatory. The proposed architecture integrates Squeeze-and-Excitation (SE) blocks for radiometric channel stabilization, the Convolutional Block Attention Module (CBAM) for spatial–semantic refinement, and Spatial Pyramid Pooling (SPP) for hemispherical multi-scale context aggregation within a fully fine-tuned EfficientNetB7 backbone, forming a domain-aware atmospheric representation framework. A large-scale dataset comprising 122,660 RGB images across 13 day–night sky-scene categories was curated, capturing diverse tropical atmospheric conditions including humidity, haze, illumination transitions, and sensor noise. Extensive experimental evaluations demonstrate that the EASMNet achieves 93% overall accuracy, outperforming representative convolutional (VGG16, ResNet50, DenseNet121) and transformer-based architectures (Swin Transformer, Vision Transformer). Ablation analyses confirm the complementary contributions of hierarchical attention and multi-scale aggregation, while class-wise evaluation yields F1-scores exceeding 0.95 for visually distinctive categories such as Day Humid, Night Clear Sky, and Night Noise. Residual errors are primarily confined to physically transitional and low-contrast atmospheric regimes. These results validate the EASMNet as a reliable, interpretable, and computationally feasible framework for real-time observatory dome automation, astronomical scheduling, and continuous atmospheric monitoring, and provide a scalable foundation for autonomous sky-observation systems deployable across diverse climatic regions. Full article
Show Figures

Figure 1

21 pages, 27867 KB  
Article
An Adaptive Attention DropBlock Framework for Real-Time Cross-Domain Defect Classification
by Shailaja Pasupuleti, Ramalakshmi Krishnamoorthy and Hemalatha Gunasekaran
AI 2026, 7(2), 56; https://doi.org/10.3390/ai7020056 - 3 Feb 2026
Viewed by 145
Abstract
The categorization of real-time defects in heterogeneous domains is a long-standing challenge in the field of industrial visual inspection systems, primarily due to significant visual variations and the lack of labelled information in real-world inspection settings. This work presents the Adaptive Attention DropBlock [...] Read more.
The categorization of real-time defects in heterogeneous domains is a long-standing challenge in the field of industrial visual inspection systems, primarily due to significant visual variations and the lack of labelled information in real-world inspection settings. This work presents the Adaptive Attention DropBlock (AADB) framework, a lightweight deep learning framework that was developed to promote cross-domain defect detection using attention-guided regularization. The proposed architecture integrates the Convolutional Block Attention Module (CBAM) and an organized DropBlock-based regularization scheme, creating a unified and robust framework. Although CBAM-based approaches improve localization of defect-related areas and traditional DropBlock provides a generic spatial regularization, neither of them alone is specifically designed to reduce domain overfitting. To address this limitation, AADB combines attention-directed feature refinement with a progressive, transfer-aware dropout policy that promotes the learning of domain-invariant representations. The proposed model is built on a MobileNetV2 base and trained through a two-phase transfer learning regime, where the first phase consists of pretraining on a source domain and the second phase consists of adaptation to a visually dissimilar target domain with constrained supervision. The overall analysis of a metal surface defect dataset (source domain) and an aircraft surface defect dataset (target domain) shows that AADB outperforms CBAM-only, DropBlock-only, and conventional MobileNetV2 models, with an overall accuracy of 91.06%, a macro-F1 of 0.912, and a Cohen’s k of 0.866. Improved feature separability and localization of error are further described by qualitative analyses using Principal Component Analysis (PCA) and Grad-CAM. Overall, the framework provides a practical, interpretable, and edge-deployable solution to the classification of cross-domain defects in the industrial inspection setting. Full article
Show Figures

Figure 1

26 pages, 6232 KB  
Article
MFE-YOLO: A Multi-Scale Feature Enhanced Network for PCB Defect Detection with Cross-Group Attention and FIoU Loss
by Ruohai Di, Hao Fan, Hanxiao Feng, Zhigang Lv, Lei Shu, Rui Xie and Ruoyu Qian
Entropy 2026, 28(2), 174; https://doi.org/10.3390/e28020174 - 2 Feb 2026
Viewed by 148
Abstract
The detection of defects in Printed Circuit Boards (PCBs) is a critical yet challenging task in industrial quality control, characterized by the prevalence of small targets and complex backgrounds. While deep learning models like YOLOv5 have shown promise, they often lack the ability [...] Read more.
The detection of defects in Printed Circuit Boards (PCBs) is a critical yet challenging task in industrial quality control, characterized by the prevalence of small targets and complex backgrounds. While deep learning models like YOLOv5 have shown promise, they often lack the ability to quantify predictive uncertainty, leading to overconfident errors in challenging scenarios—a major source of false alarms and reduced reliability in automated manufacturing inspection lines. From a Bayesian perspective, this overconfidence signifies a failure in probabilistic calibration, which is crucial for trustworthy automated inspection. To address this, we propose MFE-YOLO, a Bayesian-enhanced detection framework built upon YOLOv5 that systematically integrates uncertainty-aware mechanisms to improve both accuracy and operational reliability in real-world settings. First, we construct a multi-background PCB defect dataset with diverse substrate colors and shapes, enhancing the model’s ability to generalize beyond the single-background bias of existing data. Second, we integrate the Convolutional Block Attention Module (CBAM), reinterpreted through a Bayesian lens as a feature-wise uncertainty weighting mechanism, to suppress background interference and amplify salient defect features. Third, we propose a novel FIoU loss function, redesigned within a probabilistic framework to improve bounding box regression accuracy and implicitly capture localization uncertainty, particularly for small defects. Extensive experiments demonstrate that MFE-YOLO achieves state-of-the-art performance, with mAP@0.5 and mAP@0.5:0.95 values of 93.9% and 59.6%, respectively, outperforming existing detectors, including YOLOv8 and EfficientDet. More importantly, the proposed framework yields better-calibrated confidence scores, significantly reducing false alarms and enabling more reliable human-in-the-loop verification. This work provides a deployable, uncertainty-aware solution for high-throughput PCB inspection, advancing toward trustworthy and efficient quality control in modern manufacturing environments. Full article
(This article belongs to the Special Issue Bayesian Networks and Causal Discovery)
Show Figures

Figure 1

26 pages, 1369 KB  
Article
Progressive Attention-Enhanced EfficientNet–UNet for Robust Water-Body Mapping from Satellite Imagery
by Mohamed Ezz, Alaa S. Alaerjan, Ayman Mohamed Mostafa, Noureldin Laban and Hind H. Zeyada
Sensors 2026, 26(3), 963; https://doi.org/10.3390/s26030963 - 2 Feb 2026
Viewed by 149
Abstract
The sustainable management of water resources and the development of climate-resilient infrastructure depend on the precise identification of water bodies in satellite imagery. This paper presents a novel deep learning architecture that integrates a convolutional block attention module (CBAM) into a modified EfficientNet–UNet [...] Read more.
The sustainable management of water resources and the development of climate-resilient infrastructure depend on the precise identification of water bodies in satellite imagery. This paper presents a novel deep learning architecture that integrates a convolutional block attention module (CBAM) into a modified EfficientNet–UNet backbone. This integration allows the model to prioritize informative features and spatial areas. The model robustness is ensured through a rigorous training regimen featuring five-fold cross-validation, dynamic test-time augmentation, and optimization with the Lovász loss function. The final model achieved the following values on the independent test set: precision = 90.67%, sensitivity = 86.96%, specificity = 96.18%, accuracy = 93.42%, Dice score = 88.78%, and IoU = 79.82%. These results demonstrate improvement over conventional segmentation pipelines, highlighting the effectiveness of attention mechanisms in extracting complex water-body patterns and boundaries. The key contributions of this paper include the following: (i) adaptation of CBAM within a UNet-style architecture tailored for remote sensing water-body extraction; (ii) a rigorous ablation study detailing the incremental impact of decoder complexity, attention integration, and loss function choice; and (iii) validation of a high-fidelity, computationally efficient model ready for deployment in large-scale water-resource and ecosystem-monitoring systems. Our findings show that attention-guided segmentation networks provide a robust pathway toward high-fidelity and sustainable water-body mapping. Full article
19 pages, 2917 KB  
Article
End-to-End Autonomous Decision-Making Method for Intelligent Vehicles Based on ResNet-CBAM-BiLSTM
by Yigao Ning, Xibo Fang, Xuan Zhao, Shu Wang and Jianbo Zheng
Actuators 2026, 15(2), 84; https://doi.org/10.3390/act15020084 - 1 Feb 2026
Viewed by 170
Abstract
To solve the difficulty of autonomous decision-making caused by the complex driving environment and changeable weather conditions, an end-to-end autonomous decision-making method based on residual network (ResNet), convolutional block attention module (CBAM) and bidirectional long short-term memory network (BiLSTM) is proposed for intelligent [...] Read more.
To solve the difficulty of autonomous decision-making caused by the complex driving environment and changeable weather conditions, an end-to-end autonomous decision-making method based on residual network (ResNet), convolutional block attention module (CBAM) and bidirectional long short-term memory network (BiLSTM) is proposed for intelligent vehicles. Firstly, ResNet is used to extract spatial feature information contained in driving scene images. Then, CBAM is adopted to assign weights to each network channel and dynamically focus on important spatial regions in the image. Finally, BiLSTM is constructed to process the contextual features of continuous scenes, and the autonomous decision-making of intelligent vehicles is achieved through the fusion of spatial features and temporal information. On this basis, the proposed network model is trained using a real-world driving dataset and fully tested in various scenarios. Moreover, ablation experiments are conducted to verify the contribution of each module to the overall performance. The results show that the proposed method has better accuracy and stability compared with multiple existing methods, including PilotNet, FCN-LSTM, and DBNet, and its accuracy reaches 90.16% under clear weather conditions, as well as 81.29% under nighttime and snowy weather conditions. Full article
(This article belongs to the Section Actuators for Surface Vehicles)
Show Figures

Figure 1

37 pages, 13544 KB  
Article
Attention-Driven Feature Extraction for XAI in Histopathology Leveraging a Hybrid Xception Architecture for Multi-Cancer Diagnosis
by Shirin Shila, Md. Safayat Hossain, Md Fuyad Al Masud, Mohammad Badrul Alam Miah, Afrig Aminuddin and Zia Muhammad
Mach. Learn. Knowl. Extr. 2026, 8(2), 31; https://doi.org/10.3390/make8020031 - 28 Jan 2026
Viewed by 522
Abstract
The automated and accurate results of classifying histopathology images are necessary in the early detection of cancer, especially the common cancers such as Colorectal Cancer (CRC) and Lung Cancer (LC). Nonetheless, classical deep learning frameworks often face challenges because the intra-class variations are [...] Read more.
The automated and accurate results of classifying histopathology images are necessary in the early detection of cancer, especially the common cancers such as Colorectal Cancer (CRC) and Lung Cancer (LC). Nonetheless, classical deep learning frameworks often face challenges because the intra-class variations are large, the relations across classes are alike, and the quality of images is not stable. In order to eliminate these constraints, a multi-layer diagnostic framework is offered in detail. This process starts with a strong preprocessing pipeline, which involves gamma correction, bilateral filtering, and adaptive CLAHE, resulting in statistically significant changes in image quality quantitative measures. The hybrid attention architecture is presented and includes an Xception backbone, a Convolutional Block Attention Module (CBAM), a Transformer block, and an MLP classifier to successfully combine local features with global context. The proposed model achieved an outstanding performance with a classification of 99.98%, 99.58%, and 99.33% percent on LC25000, CRC-VAL-HE-7K, and NCT-CRC-HE-100K when tested on three publicly available datasets. In order to enhance transparency, very detailed explainability analyses are conducted with the help of layer-wise feature visualization and Grad-CAM. Finally, the real-world example of this framework is presented by its implementation in a web-based platform, which can be a useful and easy-to-use tool in helping to diagnose a pathology. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

21 pages, 3624 KB  
Article
Multi-Scale Feature Fusion and Attention-Enhanced R2U-Net for Dynamic Weight Monitoring of Chicken Carcasses
by Tian Hua, Pengfei Zou, Ao Zhang, Runhao Chen, Hao Bai, Wenming Zhao, Qian Fan and Guobin Chang
Animals 2026, 16(3), 410; https://doi.org/10.3390/ani16030410 - 28 Jan 2026
Viewed by 185
Abstract
In recent years, real-time monitoring of broiler chicken weight has become crucial for assessing growth and health status. Currently, obtaining weight data often relies on manual collection. However, this process is cumbersome, labor-intensive, and inefficient. This paper proposes a broiler carcass weight detection [...] Read more.
In recent years, real-time monitoring of broiler chicken weight has become crucial for assessing growth and health status. Currently, obtaining weight data often relies on manual collection. However, this process is cumbersome, labor-intensive, and inefficient. This paper proposes a broiler carcass weight detection model based on deep learning image segmentation and regression to address these issues. The model first segments broiler carcasses and then uses the pixel area of the segmented region as a key feature for a regression model to predict weight. A custom dataset comprising 2709 images from 301 Taihu yellow chickens was established for this study. A novel segmentation network, AR2U-AtNet, derived from R2U-Net, is proposed. To mitigate the interference of background color and texture on target carcasses in slaughterhouse production lines, the Convolutional Block Attention Module (CBAM) is introduced to enable the network to focus on areas containing carcasses. Furthermore, broilers exhibit significant variations in size, morphology, and posture, which impose high demands on the model’s scale adaptability. Selective Kernel Attention (SKAttention) is therefore integrated to flexibly handle broiler images with diverse body conditions. The model achieved an average Intersection over Union (mIoU) score of 90.45%, and Dice and F1 scores of 95.18%. The regression-based weight prediction achieved an R2 value of 0.9324. The results demonstrate that the proposed method can quickly and accurately determine individual broiler carcass weights, thereby alleviating the burden of traditional weighing methods and ultimately improving the production efficiency of yellow-feather broilers. Full article
(This article belongs to the Section Poultry)
Show Figures

Figure 1

19 pages, 1811 KB  
Article
Defective Wheat Kernel Recognition Using EfficientNet with Attention Mechanism and Multi-Binary Classification
by Duolin Wang, Jizhong Li, Han Gong and Jianyi Chen
Appl. Sci. 2026, 16(3), 1247; https://doi.org/10.3390/app16031247 - 26 Jan 2026
Viewed by 142
Abstract
As a globally significant food crop, the assessment of wheat quality is essential for ensuring food security and enhancing the processing quality of agricultural products. Conventional methods for assessing wheat kernel quality are often inefficient and markedly subjective, which hampers their ability to [...] Read more.
As a globally significant food crop, the assessment of wheat quality is essential for ensuring food security and enhancing the processing quality of agricultural products. Conventional methods for assessing wheat kernel quality are often inefficient and markedly subjective, which hampers their ability to accurately distinguish the complex and diverse phenotypic characteristics of wheat kernels. To tackle the aforementioned issues, this study presents an enhanced recognition method for defective wheat kernels, based on the EfficientNet-B1 architecture. Building upon the original EfficientNet-B1 network structure, this approach incorporates the lightweight attention mechanism known as CBAM (Convolutional Block Attention Module) to augment the model’s capacity to discern features in critical regions. Simultaneously, it modifies the classification head structure to facilitate better alignment with the data, thereby enhancing accuracy. The experiment employs a self-constructed dataset comprising five categories of wheat kernels—perfect wheat kernels, insect-damaged wheat kernels, scab-damaged wheat kernels, moldy wheat kernels, and black germ wheat kernels—which are utilized for training and validation purposes. The results indicate that the enhanced model attains a classification accuracy of 99.80% on the test set, reflecting an increase of 2.6% compared to its performance prior to the enhancement. Furthermore, the Precision, Recall, and F1-score all demonstrated significant improvements. The proposed model achieves near-perfect performance on several categories under controlled experimental conditions, with particularly high precision and recall for scab-damaged and insect-damaged kernels. This study demonstrates the efficacy of the enhanced EfficientNet-B1 model in the recognition of defective wheat kernels and offers novel technical insights and methodological references for intelligent wheat quality assessment. Full article
(This article belongs to the Section Agricultural Science and Technology)
Show Figures

Figure 1

18 pages, 14590 KB  
Article
VTC-Net: A Semantic Segmentation Network for Ore Particles Integrating Transformer and Convolutional Block Attention Module (CBAM)
by Yijing Wu, Weinong Liang, Jiandong Fang, Chunxia Zhou and Xiaolu Sun
Sensors 2026, 26(3), 787; https://doi.org/10.3390/s26030787 - 24 Jan 2026
Viewed by 265
Abstract
In mineral processing, visual-based online particle size analysis systems depend on high-precision image segmentation to accurately quantify ore particle size distribution, thereby optimizing crushing and sorting operations. However, due to multi-scale variations, severe adhesion, and occlusion within ore particle clusters, existing segmentation models [...] Read more.
In mineral processing, visual-based online particle size analysis systems depend on high-precision image segmentation to accurately quantify ore particle size distribution, thereby optimizing crushing and sorting operations. However, due to multi-scale variations, severe adhesion, and occlusion within ore particle clusters, existing segmentation models often exhibit undersegmentation and misclassification, leading to blurred boundaries and limited generalization. To address these challenges, this paper proposes a novel semantic segmentation model named VTC-Net. The model employs VGG16 as the backbone encoder, integrates Transformer modules in deeper layers to capture global contextual dependencies, and incorporates a Convolutional Block Attention Module (CBAM) at the fourth stage to enhance focus on critical regions such as adhesion edges. BatchNorm layers are used to stabilize training. Experiments on ore image datasets show that VTC-Net outperforms mainstream models such as UNet and DeepLabV3 in key metrics, including MIoU (89.90%) and pixel accuracy (96.80%). Ablation studies confirm the effectiveness and complementary role of each module. Visual analysis further demonstrates that the model identifies ore contours and adhesion areas more accurately, significantly improving segmentation robustness and precision under complex operational conditions. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

33 pages, 6275 KB  
Article
TABS-Net: A Temporal Spectral Attentive Block with Space–Time Fusion Network for Robust Cross-Year Crop Mapping
by Xin Zhou, Yuancheng Huang, Qian Shen, Yue Yao, Qingke Wen, Fengjiang Xi and Chendong Ma
Remote Sens. 2026, 18(2), 365; https://doi.org/10.3390/rs18020365 - 21 Jan 2026
Viewed by 230
Abstract
Accurate and stable mapping of crop types is fundamental to agricultural monitoring and food security. However, inter-annual phenological shifts driven by variations in air temperature, precipitation, and sowing dates introduce systematic changes in the spectral distributions associated with the same day of year [...] Read more.
Accurate and stable mapping of crop types is fundamental to agricultural monitoring and food security. However, inter-annual phenological shifts driven by variations in air temperature, precipitation, and sowing dates introduce systematic changes in the spectral distributions associated with the same day of year (DOY). As a result, the “date–spectrum–class” mapping learned during training can become misaligned when applied to a new year, leading to increased misclassification and unstable performance. To tackle this problem, we develop TABS-Net (Temporal–Spectral Attentive Block with Space–Time Fusion Network). The core contributions of this study are summarized as follows: (1) we propose an end-to-end 3D CNN framework to jointly model spatial, temporal, and spectral information; (2) we design and embed CBAM3D modules into the backbone to emphasize informative bands and key time windows; and (3) we introduce DOY positional encoding and temporal jitter during training to explicitly align seasonal timing and simulate phenological shifts, thereby enhancing cross-year robustness. We conduct a comprehensive evaluation on a Cropland Data Layer (CDL) subset. Within a single year, TABS-Net delivers higher and more balanced overall accuracy, Macro-F1, and mIoU than strong baselines, including 2D stacking, 1D temporal convolution/LSTM, and transformer models. In cross-year experiments, we quantify temporal stability using inter-annual robustness (IAR); with both DOY encoding and temporal jitter enabled, the model attains IAR values close to one for major crop classes, effectively compensating for phenological misalignment and inter-annual variability. Ablation studies show that DOY encoding and temporal jitter are the primary contributors to improved inter-annual consistency, while CBAM3D reduces crop–crop and crop–background confusion by focusing on discriminative spectral regions such as the red-edge and near-infrared bands and on key growth stages. Overall, TABS-Net combines higher accuracy with stronger robustness across multiple years, offering a scalable and transferable solution for large-area, multi-year remote sensing crop mapping. Full article
Show Figures

Figure 1

18 pages, 5475 KB  
Article
Small PCB Defect Detection Based on Convolutional Block Attention Mechanism and YOLOv8
by Zhe Sun, Ruihan Ma and Qujiang Lei
Appl. Sci. 2026, 16(2), 1078; https://doi.org/10.3390/app16021078 - 21 Jan 2026
Viewed by 172
Abstract
Automated defect detection in printed circuit boards (PCBs) is a critical process for ensuring the quality and reliability of electronic products. To address the limitations of existing detection methods, such as insufficient sensitivity to minor defects and limited recognition accuracy in complex backgrounds, [...] Read more.
Automated defect detection in printed circuit boards (PCBs) is a critical process for ensuring the quality and reliability of electronic products. To address the limitations of existing detection methods, such as insufficient sensitivity to minor defects and limited recognition accuracy in complex backgrounds, this paper proposes an enhanced YOLOv8 detection framework. The core contribution lies not merely in the integration of the Convolutional Block Attention Module (CBAM), but in a principled and task-specific integration strategy designed to address the multi-scale and low-contrast nature of PCB defects. The complete CBAM is integrated into the multi-scale feature layers (P3, P4, P5) of the YOLOv8 backbone network. By leveraging sequential channel and spatial attention submodules, CBAM guides the model to dynamically optimise feature responses, thereby significantly enhancing feature extraction for tiny, morphologically diverse defects. Experiments on a public PCB defect dataset demonstrate that the proposed model achieves a mean average precision (mAP@50) of 98.8% while maintaining real-time inference speed, surpassing the baseline YOLOv8 model by 9.5%, with the improvements of 7.4% in precision and 12.3% in recall. While the model incurs a higher computational cost (79.4 GFLOPs), it maintains a real-time inference speed of 109.11 FPS, offering a viable trade-off between accuracy and efficiency for high-precision industrial inspection. The proposed model demonstrates superior performance in detecting small-scale defects, making it highly suitable for industrial deployment. Full article
(This article belongs to the Special Issue Digital Technologies Enabling Modern Industries, 2nd Edition)
Show Figures

Figure 1

17 pages, 5869 KB  
Article
Research on Tool Wear Prediction Method Based on CNN-ResNet-CBAM-BiGRU
by Bo Sun, Hao Wang, Jian Zhang, Lixin Zhang and Xiangqin Wu
Sensors 2026, 26(2), 661; https://doi.org/10.3390/s26020661 - 19 Jan 2026
Viewed by 236
Abstract
Aiming to address insufficient feature extraction, vanishing gradients, and low prediction accuracy in tool wear prediction, this paper proposes a hybrid deep neural network based on a Convolutional Neural Network (CNN), Residual Network (ResNet) residual connections, the Convolutional Block Attention Module (CBAM), and [...] Read more.
Aiming to address insufficient feature extraction, vanishing gradients, and low prediction accuracy in tool wear prediction, this paper proposes a hybrid deep neural network based on a Convolutional Neural Network (CNN), Residual Network (ResNet) residual connections, the Convolutional Block Attention Module (CBAM), and a Bidirectional Gated Recurrent Unit (BiGRU). First, a 34-dimensional multi-domain feature set covering the time domain, frequency domain, and time–frequency domain is constructed, and multi-sensor signals are standardized using z-score normalization. A CNN–BiGRU backbone is then established, where ResNet-style residual connections are introduced to alleviate training degradation and mitigate vanishing-gradient issues in deep networks. Meanwhile, CBAM is integrated into the feature extraction module to adaptively reweight informative features in both channel and spatial dimensions. In addition, a BiGRU layer is embedded for temporal modeling to capture bidirectional dependencies throughout the wear evolution process. Finally, a fully connected layer is used as a regressor to map high-dimensional representations to tool wear values. Experiments on the PHM2010 dataset demonstrate that the proposed hybrid architecture is more stable and achieves better predictive performance than several mainstream deep learning baselines. Systematic ablation studies further quantify the contribution of each component: compared with the baseline CNN model, the mean absolute error (MAE) is reduced by 47.5%, the root mean square error (RMSE) is reduced by 68.5%, and the coefficient of determination (R2) increases by 14.5%, enabling accurate tool wear prediction. Full article
(This article belongs to the Section Sensor Networks)
Show Figures

Figure 1

27 pages, 6052 KB  
Article
Wind Turbines Small Object Detection in Remote Sensing Images Based on CGA-YOLO: A Case Study in Shandong Province, China
by Jingjing Ma, Guizhou Wang, Ranyu Yin, Guojin He, Dengji Zhou, Tengfei Long, Elhadi Adam and Zhaoming Zhang
Remote Sens. 2026, 18(2), 324; https://doi.org/10.3390/rs18020324 - 18 Jan 2026
Viewed by 305
Abstract
With the rapid development of high-resolution satellite remote sensing technology, wind turbine detection based on remote sensing imagery has emerged as a crucial research area in renewable energy. However, accurate identification of wind turbines remains challenging due to complex geographical backgrounds and their [...] Read more.
With the rapid development of high-resolution satellite remote sensing technology, wind turbine detection based on remote sensing imagery has emerged as a crucial research area in renewable energy. However, accurate identification of wind turbines remains challenging due to complex geographical backgrounds and their typical appearance as small objects in images, where limited features and background interference hinder detection performance. To address these issues, this paper proposes CGA-YOLO, a specialized network for detecting small targets in high-resolution remote sensing images, and constructs the SDWT dataset, containing Gaofen-2 imagery covering various terrains in Shandong Province, China. The network incorporates three key enhancements: dynamic convolution improves multi-scale feature representation for precise localization; the Convolutional Block Attention Module (CBAM) enhances feature convergence through channel and spatial attention mechanisms; and GhostBottleneck maintains high-resolution details while strengthening feature channels for small targets. Experimental results demonstrate that CGA-YOLO achieves an F1-score of 0.93 and an mAP50 of 0.938 on the SDWT dataset, and obtains an mAP50 of 0.9033 on both RSOD and VEDAI public datasets. CGA-YOLO establishes its superior accuracy over multiple mainstream detection models under identical experimental conditions, confirming its potential as a reliable technical solution for accurate wind turbine identification in complex environments. Full article
Show Figures

Figure 1

Back to TopTop