Next Article in Journal
Eco-Friendly and Sustainable One-Component Polyurethane Syntactic Foams Reinforced with Fly Ash Cenospheres for Acoustic and Thermal Insulation
Previous Article in Journal
Statistical Analysis of Heat Transfer Effects on Flow Patterns Maps in a Flat-Plate Collector/Evaporator with R600a Under Variable Tilt Angles
Previous Article in Special Issue
A Multi-Granularity Random Mutation Genetic Algorithm for Steel Cold Rolling Scheduling Optimization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Online Blast Furnace Condition Recognition Method Based on Spatiotemporal Texture Feature Coupling and Diffusion Networks

1
School of Automation, Central South University, Changsha 410083, China
2
State Key Laboratory of Precision Manufacturing for Extreme Service Performance, Central South University, No. 932, Lushan South Road, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Processes 2025, 13(11), 3416; https://doi.org/10.3390/pr13113416 (registering DOI)
Submission received: 27 September 2025 / Revised: 17 October 2025 / Accepted: 21 October 2025 / Published: 24 October 2025

Abstract

Real-time and accurate identification of the blast furnace (BF) condition is essential for maintaining stability and improving energy efficiency in steelmaking. However, the harsh environment inside the BF makes direct acquisition of the BF condition extremely difficult. To address this challenge, this study proposes an online BF condition recognition method based on spatiotemporal texture feature coupling and diffusion networks (STFC-DN). The method employs a multi-domain Swin-Transformer module (MDSTM) combined with wavelet decomposition and channel attention to extract the gas flow region. A temporal feature pyramid network module (T-FPNM) is then used to capture both the global and local spatiotemporal characteristics of this region. Heuristic clustering and an idempotent generative network (IGN) are introduced to obtain standardized BF condition features, enabling intelligent classification through multi-metric similarity analysis. Experimental results show that the proposed STFC-DN achieves an average accuracy exceeding 98% when identifying four BF conditions: normal, hanging, oblique stockline, and collapsing, with an inference speed of approximately 28 FPS. This approach demonstrates both high accuracy and real-time capability, showing strong potential for advancing the intelligent and sustainable development of the steel industry.

1. Introduction

The blast furnace (BF) is the core facility in the steelmaking process; its operation consumes substantial energy and directly determines steel quality [1,2]. BF conditions serve as key indicators of BF operational stability and smelting efficiency [3,4]. Therefore, precise and real-time monitoring of BF conditions is critically important [5,6]. However, the high-temperature, high-pressure, sealed, and dusty environment inside the BF renders direct acquisition of BF condition information highly challenging [7]. Although the high-temperature industrial endoscope developed by Chen, Yi et al. [8,9,10], and colleagues enables real-time acquisition of BF videos, the absence of auxiliary light sources often causes BF burden surface images to exhibit blurred textures and uneven illumination, severely limiting subsequent BF condition analysis [11]. Therefore, there is an urgent need to develop efficient BF condition recognition methods to achieve intelligent control of the BF ironmaking process [12].
Existing studies on BF condition recognition can be broadly categorized into two approaches: mechanism-based modeling and data-driven, multimodal fusion with intelligent prediction [13].
(1) Mechanism-based modeling and data-driven: BF condition recognition is commonly performed under the constraints of mechanistic models, where operational data are integrated and deep learning is employed for modeling. Xu et al. [14] combined industrial endoscopy with a virtual multi-camera array to achieve real-time 3D measurement of the BF burden surface under extreme environments, demonstrating reliable performance for BF condition monitoring and control. Zhou et al. [15] established quantitative criteria for abnormal condition recognition by detecting coke particle size and temperature distribution in the tuyere region. Duan et al. [16] developed a novel deep neural network, ES-SFRNet, which achieves intelligent recognition of BF tuyere conditions by fusing tuyere images with temporal data. Zhao et al. [17] introduced the GT-MSPC model, which leverages multi-objective optimization for abnormal condition detection and early warning. These methods can achieve recognition accuracy above 95% under specific operating conditions. However, their performance relies heavily on sensor precision and the accurate measurement of internal furnace parameters. When the operating conditions become complex, the prediction error of the models exceeds 10%, which limits their applicability in industrial environments.
(2) Multimodal fusion with intelligent prediction: To overcome the limitations of single data sources, some studies have attempted to integrate radar point clouds, visible/infrared images, and temperature data to construct multi-source information-driven discriminative models. Guo et al. [18] proposed a BF condition classification model based on multi-source fusion and Dynamic Process Adaptive Kernel Fisher Discrimination (DPAKF), which achieved high accuracy. Tian et al. [13] employed RGB-D cameras and digital elevation models to measure burden surface roughness and perform state discrimination, enabling comprehensive BF condition recognition. Lee et al. [19] combined hybrid quantum machine learning with pulverized coal injection control to improve the stability of BF temperature prediction. This type of method achieves an average classification accuracy of 98.2% for BF conditions, representing a 25% improvement in prediction accuracy compared with traditional models. However, issues such as high computational complexity, difficulty in cross-modal data alignment, and limited interpretability still remain.
To address these issues, this paper proposes an online BF condition recognition method based on spatiotemporal texture feature coupling and diffusion networks (STFC-DN). The main contributions are as follows:
(1) To overcome the problem of BF blurred burden surface images that obscure critical gas flow regions, a multi-domain Swin-Transformer module (MDSTM) is proposed to enable precise localization and segmentation of these regions.
(2) To handle the challenge that gas flow regions change dynamically over time and are difficult to comprehensively represent using conventional methods, a temporal feature pyramid network module (T-FPNM) is proposed, which integrates spatial features with temporal dynamic features to achieve multi-scale representation of BF conditions.
(3) To reduce the strong dependence on BF condition labels, alleviate the difficulty of manual annotation, and improve classification accuracy, an intelligent recognition method based on feature generation is introduced. This approach effectively enhances recognition accuracy and robustness under complex BF conditions.
The remainder of this paper is organized as follows: Section 2 introduces BF condition classification and BF burden surface image preprocessing. Section 3 elaborates on the proposed recognition method. Section 4 presents the experimental design and analysis of results. Section 5 concludes the paper and outlines future research directions.

2. Classification of BF Conditions and Color Rendering of BF Burden Surface Images

2.1. Classification and Feature Description of BF Conditions

BF conditions are generally classified into two categories: normal conditions and abnormal conditions. Abnormal conditions primarily include typical types such as hanging, oblique stockline, and collapsing. Table 1 summarizes the characteristic features associated with different BF conditions, and Figure 1 illustrates representative image examples.

2.2. Image Enhancement and Color Rendering of BF Burden Surface

The BF burden surface images captured by a high-temperature industrial endoscope, as illustrated in Figure 2, are of poor quality. To address this limitation, STFC-DN incorporates image clarification and color rendering into the preprocessing stage.
During image clarification, the BF burden surface image is first decomposed into an illumination map and a reflectance map using an encoder–decoder network. Subsequently, gamma correction is applied to the illumination map to improve brightness uniformity, while a bilateral texture filtering-based detail enhancement method [20] is applied to the reflectance map to emphasize the edges and texture features of the gas flow region, as shown in Equation (1).
BS = Illumination Rflectance I out = I in γ R out = B T F R in + i = 1 L 1 Enhance R in B T F R in , k i , k i
where BS denotes the BF burden surface image; Illumination and Rflectance represent the decomposed illumination and reflectance components; γ is the gamma value; I in γ and I out denote the input and enhanced illumination information; BTF R in , k i refers to the bilateral texture filtering operation; k i indicates the filter size of the i-th layer; Enhance d i , k i denotes the enhancement function applied to the texture detail image d i ( d i = R in BTF R in , k i ); and R out represents the enhanced reflectance information.
During color rendering, regions with varying temperature and brightness distributions are mapped into a pseudo-color space to enhance the distinction of gas flow patterns. The experimental results are presented in Figure 3.

3. Methods

STFC-DN first utilizes the MDSTM to precisely extract the gas flow region. Subsequently, the multi-scale texture features and spatial structural features of this region are computed separately and then fused to construct a BF condition representation vector. Finally, an intelligent recognition method based on feature generation is introduced to enable precise identification of the BF condition. The overall framework of the proposed method is illustrated in Figure 4.

3.1. Gas Flow Region Extraction Based on Multi-Domain Swin-Transformer Module (MDSTM)

The MDSTM incorporates multi-scale wavelet decomposition and a channel attention mechanism into the Swin-Transformer [21] to enhance detail capture and global structural perception in gas flow regions. The structure of the MDSTM is illustrated in Figure 5.
First, during the model training stage, a total of 2069 representative blast furnace (BF) burden surface images were selected, with each frame corresponding to one second of the internal furnace state. The dataset included 758 frames of normal BF condition, 547 frames of hanging, 435 frames of oblique stockline, and 329 frames of collapsing. The high-temperature gas flow region in each frame was manually annotated based on three criteria: the regional brightness exceeded the background by more than 25%, the region exhibited plume-like or radially diffused bright textures, and its edge morphology remained stable while changing continuously across adjacent frames. These annotations were used to generate binary supervision labels denoted as M g t . In addition, multi-scale features G f obtained through two-dimensional wavelet transform were introduced as guiding information. Subsequently, the channel attention weights A c were calculated according to the correlation within the gas flow region, and the extracted features were weighted as follows:
G f weighted = A c G f
The feature set G f weighted was then fed into multiple layers of the multi-domain Swin-Transformer module (MDSTM). The window-based multi-head self-attention mechanism (W-MSA) was employed to capture local feature dependencies, while the shifted window strategy (SW-MSA) was used to achieve information interaction across different windows:
G output = SwinBlock ( Shift ( Multi Attention ( G f weight ) ) )
Finally, the output features G output are mapped to the probability space of the gas flow regions through a fully connected layer (FC), yielding the segmentation results, calculated as follows:
P flame = FC ( G output )
The model was trained in a supervised manner, where the input images and their corresponding labels M g t were used for feature learning. The loss function was defined as a weighted combination of binary cross-entropy loss and dice loss, written as follows:
L = λ 1 L B C E ( M p r e d , M g t ) + λ 2 L D i c e ( M p r e d , M g t )
where, λ 1 and λ 2 represent adjustable weighting coefficients.

3.2. Feature Computation of Gas Flow Regions

After the MDSTM precisely extracts the gas flow region from the BF burden surface image, the texture features and spatiotemporal structural characteristics of this region need to be further quantified.

3.2.1. Multi-Scale Texture Features

The multi-scale texture features comprise the perimeter P, area A, brightness B, and sharpness S of the gas flow region, which are combined through weighted fusion, as formulated in Equation (6), written as follows:
F fusion = ω 1 P + ω 2 A + ω 3 B + ω 4 S
where ω 1 ω 4 denotes a learnable weight, which is automatically updated through a Multi-Layer Perceptron (MLP) followed by Softmax normalization:
[ ω 1 , ω 2 , ω 3 , ω 4 ] = Softmax ( MLP ( P , A , B , S ) )
The fused feature preserves geometric shape information while integrating edge characteristics, thereby providing input that combines both global and local information for BF condition recognition.

3.2.2. Spatiotemporal Structural Feature Extraction Based on Temporal FPN Module (T-FPNM)

The structure of the T-FPNM is illustrated in Figure 6. First, a multi-scale convolutional network is employed to extract the spatial features of the gas flow region frame { R t } t = 1 T and to generate multi-level feature maps { F t 1 , F t 2 , , F t L } . Subsequently, feature maps at the same level are stacked into a sequence F stack l and fed into a convolutional long short-term memory (ConvLSTM) network to produce feature maps that incorporate temporal information [22]:
F temporal l = ConvLSTM ( F stack l )
where l denotes the the layer index. Finally, a top-down strategy is employed to iteratively fuse features across layers. The feature map F temporal l + 1 is upsampled to match the resolution of F temporal l and then fused with it via element-wise addition:
F ^ l = Upsample ( F temporal l + 1 ) + F temporal l
By propagating down to the bottom layer, the complete multi-scale spatiotemporal feature set { F ^ 1 , F ^ 2 , , F ^ L } is obtained. This set is then combined with the multi-scale texture features F fusion extracted by the MDSTM to construct the BF condition representation vector { F fusion , F ^ 1 , F ^ 2 , , F ^ L } .

3.3. Intelligent BF Condition Recognition Method Based on Feature Generation

The method first obtains BF condition labels through heuristic fuzzy clustering (HFC), then employs an idempotent generative network (IGN) [23] to generate standardized feature representations of BF conditions, and finally combines multi-metric similarity measurements to achieve BF condition classification. The workflow is shown in Figure 7.
First, the HFC uses the multi-scale texture features of the gas flow region as clustering centers to derive labels Z i for four categories of BF conditions, which are subsequently fed into the IGN network:
G i = IGN ( Z i-embed , η i )
where η i represents noise, G i denote the generated BF condition features. Subsequently, the BF condition representation vector F = { F fusion , F ^ 1 , F ^ 2 , , F ^ L } is compared with the generated features G i :
S cos ( F , G i ) = F · G i | | F | | 2 | | G i | | 2 + ϵ D maha ( F , G i ) = ( F G i ) T 1 ( F G i ) D euc ( F , G i ) = F G i 2 m
where S cos , D maha , and D euc denote the modified cosine similarity, the normalized Mahalanobis distance, and the normalized Euclidean distance, respectively. m represents the feature dimension. The results of the three metrics are integrated using an adaptive weight learning network. Finally, the BF condition category is determined by the maximum discrimination score:
Z pred = arg max i M i
where M i denotes the integrated discrimination score for BF conditions.

4. Experiments

The experimental data were obtained from independent BF condition videos captured by a high-temperature industrial endoscope in a steel plant. The dataset includes four typical BF conditions: normal, hanging, oblique stockline, and collapsing. A total of 162 video clips were selected, each with a duration of approximately 10 s. Sampling at 1 fps resulted in 2956 image frames, including 1083 frames of normal, 782 of hanging, 621 of oblique stockline, and 470 of collapsing. All data were divided into training, validation, and test sets in a ratio of 7:2:1. Representative samples are shown in Figure 8. All experiments were implemented in Python 3.10.8 on the PyCharm Professional 2023.2 platform (JetBrains s.r.o., Prague, Czech Republic), using an NVIDIA GeForce RTX 3090 GPU (NVIDIA Corporation, Santa Clara, CA, USA) with 24 GB of memory.

4.1. Generation of BF Condition Clustering Labels

To reduce reliance on manual annotation, the HFC is employed to automatically generate BF condition labels. The clustering results are illustrated in Figure 9, showing that different BF conditions can be effectively distinguished: normal conditions correspond to three clusters, hanging conditions to four clusters, oblique stockline conditions to five clusters, and collapsing conditions to six clusters. These results validate that the proposed method can automatically uncover potential BF condition structures from image features.

4.2. Accuracy and Efficiency in BF Condition Recognition

To evaluate the overall performance of the STFC-DN, a comparative analysis was conducted against several representative classification networks, including ResNet-50 [24], EfficientNet-B3 [25], HRNet-W32 [26], and the Vision Transformer [27], as shown in Table 2. The evaluation metrics included Precision, Recall, F1-score, and average inference speed, as defined in Equation (13). Precision, Recall, and F1-score were computed from the confusion matrix, which was generated by frame-by-frame comparison between model predictions and manual annotations.
The results indicate that STFC-DN achieved average Precision, Recall, and F1-score values of 98.29%, 97.83%, and 98.16%, respectively, across the four BF conditions, with an average inference speed of 27.79 FPS. These results outperform those of the comparison models and demonstrate that STFC-DN can fully meet the real-time recognition requirements of blast furnace operations.
P r e c i s i o n = T P T P + F P R e c a l l = T P T P + F N F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
where T P , F P , and F N represent the numbers of true positive, false positive, and false negative samples, respectively, and N denotes the number of BF condition categories.

4.3. Feature Analysis and Ablation Experiments of High-Temperature Gas Flow Regions

4.3.1. Multi-Scale Feature Computation of High-Temperature Gas Flow

The multi-scale texture features and spatiotemporal structural features of the gas flow region were computed to quantify the dynamic differences among various BF conditions. Table 3 presents the feature distributions under four representative BF conditions. Overall, in the abnormal conditions, the gas flow region exhibits significant deviations in area, brightness, and spatiotemporal characteristics compared with the normal condition. These deviations indicate stronger local instability and more pronounced dynamic fluctuations within the gas flow region.

4.3.2. Ablation Experiments on the Extraction of High-Temperature Gas Flow Regions

To assess the effectiveness of multi-domain guidance information in the MDSTM, the wavelet transform module was removed, and the model was subsequently retrained. Figure 10 compares the extraction results of the high-temperature gas flow region with and without the wavelet transform module, while Table 4 reports the corresponding BF condition recognition accuracy.
It is evident that the recognition accuracy decreases substantially after removing the multi-domain guidance information, with the decline being particularly pronounced under normal conditions. This finding indicates that frequency-domain details are essential for the precise extraction of gas flow regions and further validates the necessity of the multi-domain feature fusion design.

5. Conclusions

This study addresses the challenges of the harsh internal environment of the BF and the difficulty of obtaining accurate BF condition information by proposing the STFC-DN method. The method employs the MDSTM to achieve precise extraction of the gas flow region, integrates the T-FPNM to characterize its dynamic spatiotemporal evolution, and utilizes feature generation together with a multi-metric discrimination mechanism to accomplish intelligent classification and online recognition of BF conditions. The proposed model can be integrated with the BF process control system to dynamically adjust key operational parameters such as burden distribution eccentricity, tuyere air volume, and pulverized coal injection rate based on recognition results. This enables early warning and prevention of abnormal BF conditions, thereby avoiding interruptions in furnace operation.
However, the performance of STFC-DN remains sensitive to the quality of BF burden surface images and illumination conditions during data acquisition. Its inference speed is limited on devices with low computational capability, and its generalization performance across different furnace types and operating conditions still requires further validation. Future work will focus on model lightweighting, robustness enhancement, and multimodal data fusion to improve the system’s long-term stability and adaptability in complex industrial environments.

Author Contributions

X.J., conceptualization, methodology, and writing—original draft preparation; J.H. (Jie Han), software, data curation, and validation; J.H. (Jianjun He), formal analysis and writing—review and editing; W.G., supervision, project administration, and funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

Supported by the National Natural Science Foundation of China (grant No. 62373377) and the Projects of State Key Laboratory of Precision Manufacturing for Extreme Service Performance under grants ZZYJKT2025-08 and ZZYJKT2025-10.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed at the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Scolari, S.; Dall’Osto, G.; Tuveri, A.; Mombelli, D.; Mapelli, C. Optimization of Red Mud and Blast Furnace Sludge Self-Reducing Briquettes Propaedeutic for Subsequent Magnetic Separation. Metals 2025, 15, 1108. [Google Scholar] [CrossRef]
  2. Xu, D.; Li, Z.; Chen, X.; Wang, Z.; Wu, J. A dielectric-filled waveguide antenna element for 3D imaging radar in high temperature and excessive dust conditions. Sensors 2016, 16, 1339. [Google Scholar] [CrossRef]
  3. Wu, H.; Yu, L.; Chang, S.; Zhang, Y.; Yang, J. Microstructure evolution behavior of blast-furnace coke under different gasification reaction conditions. Coatings 2022, 12, 1116. [Google Scholar] [CrossRef]
  4. Mio, H.; Narita, Y.; Nakano, K.; Nomura, S. Validation of the burden distribution of the 1/3-scale of a blast furnace simulated by the discrete element method. Processes 2019, 8, 6. [Google Scholar] [CrossRef]
  5. La, G.H.; Choi, J.S.; Min, D.J. Investigation on the reaction behaviour of partially reduced iron under blast furnace conditions. Metals 2021, 11, 839. [Google Scholar] [CrossRef]
  6. Li, Y.; Zhang, S.; Zhang, J.; Yin, Y.; Xiao, W.; Zhang, Z. Data-driven multiobjective optimization for burden surface in blast furnace with feedback compensation. IEEE Trans. Ind. Inform. 2019, 16, 2233–2244. [Google Scholar] [CrossRef]
  7. Liu, R.; Gao, Z.Y.; Li, H.Y.; Liu, X.J.; Lv, Q. Research on Blast Furnace Ingredient Optimization Based on Improved Grey Wolf Optimization Algorithm. Metals 2024, 14, 798. [Google Scholar] [CrossRef]
  8. Chen, Z.; Jiang, Z.; Gui, W.; Yang, C. A novel device for optical imaging of blast furnace burden surface: Parallel low-light-loss backlight high-temperature industrial endoscope. IEEE Sens. J. 2016, 16, 6703–6717. [Google Scholar] [CrossRef]
  9. Yi, Z.; Chen, Z.; Jiang, Z.; Gui, W. A novel 3-D high-temperature industrial endoscope with large field depth and wide field. IEEE Trans. Instrum. Meas. 2020, 69, 6530–6543. [Google Scholar] [CrossRef]
  10. Chen, Z.; Wang, X.; Gui, W.; Zhu, J.; Yang, C.; Jiang, Z. A novel sensing imaging equipment under extremely dim light for blast furnace burden surface: Starlight high-temperature industrial endoscope. IEEE/CAA J. Autom. Sin. 2024, 11, 893–906. [Google Scholar] [CrossRef]
  11. Sun, S.; Yu, Z.; Zhang, S.; Xiao, W. Future definition and extraction of the blast furnace 3D burden surface based on intelligent algorithms. Appl. Sci. 2022, 12, 12860. [Google Scholar] [CrossRef]
  12. Shao, S.; Huang, Y.; Shi, H.; Sun, M. Review on simulation and practice of blast furnace burden distribution process. Metall. Res. Technol. 2025, 122, 610. [Google Scholar] [CrossRef]
  13. Tian, J.; Tanaka, A.; Gao, D.; Liu, Z.; Hou, Q.; Chen, X. Characterization of the Blast Furnace Burden Surface: Experimental Measurement and Roughness Statistics. ISIJ Int. 2023, 63, 1217–1225. [Google Scholar] [CrossRef]
  14. Xu, T.; Chen, Z.; Jiang, Z.; Huang, J.; Gui, W. A real-time 3D measurement system for the blast furnace burden surface using high-temperature industrial endoscope. Sensors 2020, 20, 869. [Google Scholar] [CrossRef]
  15. Zhou, D.; Xu, K.; Bai, J.; He, D. On-line detecting the tuyere coke size and temperature distribution of raceway zone in a working blast furnace. Fuel 2022, 316, 123349. [Google Scholar] [CrossRef]
  16. Duan, Y.; Liu, X.; Liu, R.; Li, X.; Li, H.; Li, H.; Sun, Y.; Zhang, Y.; Lv, Q. A novel anomaly detection and classification algorithm for application in tuyere images of blast furnace. Eng. Appl. Artif. Intell. 2025, 139, 109558. [Google Scholar] [CrossRef]
  17. Zhao, L.T.; Yang, T.; Yan, R.; Zhao, H.B. Anomaly detection of the blast furnace smelting process using an improved multivariate statistical process control model. Process Saf. Environ. Prot. Trans. Inst. Chem. Eng. Part B 2022, 166, 617–627. [Google Scholar] [CrossRef]
  18. Guo, K.; Zhang, Y.X.; Zhang, S.; Xiao, W.D. Classification model for blast furnace status based on multi-source information. Eng. Appl. Artif. Intell. 2025, 141, 109728. [Google Scholar] [CrossRef]
  19. Lee, N.; Shin, M.; Sagingalieva, A.; Tripathi, A.J.; Pinto, K.; Melnikov, A. Predictive control of blast furnace temperature in steelmaking with hybrid depth-infused quantum neural networks. arXiv 2025, arXiv:2504.12389. [Google Scholar] [CrossRef]
  20. Zhi-cheng, H.; Chuan, W.; Hang, Y.; Ming, Z. Image detail enhancement method based on multi-scale bilateral texture filter. Chin. Opt. 2016, 9, 423–431. [Google Scholar] [CrossRef]
  21. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  22. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  23. Shocher, A.; Dravid, A.; Gandelsman, Y.; Mosseri, I.; Rubinstein, M.; Efros, A.A. Idempotent generative network. arXiv 2023, arXiv:2311.01462. [Google Scholar] [CrossRef]
  24. Wen, L.; Li, X.; Gao, L. A transfer convolutional neural network for fault diagnosis based on ResNet-50. Neural Comput. Appl. 2020, 32, 6111–6124. [Google Scholar] [CrossRef]
  25. Alhichri, H.; Alswayed, A.S.; Bazi, Y.; Ammour, N.; Alajlan, N.A. Classification of remote sensing images using EfficientNet-B3 CNN model with attention. IEEE Access 2021, 9, 14078–14094. [Google Scholar] [CrossRef]
  26. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–19 June 2019; pp. 5693–5703. [Google Scholar]
  27. Han, K.; Wang, Y.; Chen, H.; Chen, X.; Guo, J.; Liu, Z.; Tang, Y.; Xiao, A.; Xu, C.; Xu, Y.; et al. A survey on vision transformer. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 87–110. [Google Scholar] [CrossRef]
Figure 1. Illustrations of BF conditions: (a) normal condition, (b) hanging, (c) oblique stockline, and (d) collapsing.
Figure 1. Illustrations of BF conditions: (a) normal condition, (b) hanging, (c) oblique stockline, and (d) collapsing.
Processes 13 03416 g001
Figure 2. BF burden surface image acquired by high-temperature industrial endoscope: (ac) BF burden surface images obtained from different batches.
Figure 2. BF burden surface image acquired by high-temperature industrial endoscope: (ac) BF burden surface images obtained from different batches.
Processes 13 03416 g002
Figure 3. BF burden surface image after enhancement and pseudo-color rendering: (a,c) clarified BF burden surface images and (b,d) color-rendered BF burden surface images.
Figure 3. BF burden surface image after enhancement and pseudo-color rendering: (a,c) clarified BF burden surface images and (b,d) color-rendered BF burden surface images.
Processes 13 03416 g003
Figure 4. Overall framework of the proposed STFC-DN method.
Figure 4. Overall framework of the proposed STFC-DN method.
Processes 13 03416 g004
Figure 5. Structure of the proposed MDSTM.
Figure 5. Structure of the proposed MDSTM.
Processes 13 03416 g005
Figure 6. Structure of the proposed T-FPNM.
Figure 6. Structure of the proposed T-FPNM.
Processes 13 03416 g006
Figure 7. Workflow of the proposed method.
Figure 7. Workflow of the proposed method.
Processes 13 03416 g007
Figure 8. BF Burden surface images captured by a high-temperature industrial endoscope: (ad) images from different batches.
Figure 8. BF Burden surface images captured by a high-temperature industrial endoscope: (ad) images from different batches.
Processes 13 03416 g008
Figure 9. Clustering results of BF condition labels obtained by HFC. (a) normal condition, (b) hanging, (c) oblique stockline, and (d) collapsing.
Figure 9. Clustering results of BF condition labels obtained by HFC. (a) normal condition, (b) hanging, (c) oblique stockline, and (d) collapsing.
Processes 13 03416 g009
Figure 10. Extraction results of the high-temperature gas flow region: (a) original image; (b) extraction with multi-domain guidance information; (c) extraction without multi-domain guidance information.
Figure 10. Extraction results of the high-temperature gas flow region: (a) original image; (b) extraction with multi-domain guidance information; (c) extraction without multi-domain guidance information.
Processes 13 03416 g010
Table 1. Classification and Characteristics of BF Conditions.
Table 1. Classification and Characteristics of BF Conditions.
BF ConditionCharacteristics
Normal conditionStable combustion inside the BF, uniform gas flow distribution, consistent brightness, clear gas flow morphology, and sharp image display.
HangingThe burden surface cannot descend steadily, permeability decreases, gas flow is obstructed, and central gas flow brightness intensifies.
Oblique stocklineUneven burden descent, unreasonable gas flow distribution, asymmetric gas flow pattern, and low brightness.
CollapsingSudden massive collapse of burden, chaotic and irregular image appearance, and low brightness.
Table 2. Recognition accuracy and inference speed of different models under various BF conditions.
Table 2. Recognition accuracy and inference speed of different models under various BF conditions.
ModelBF ConditionPrecision (%)Recall (%)F1-Score (%)Inference Speed (FPS)
ResNet-50Normal condition96.2295.8896.0020.48
Hanging95.8195.2595.5219.57
Oblique stockline95.1894.5494.8519.67
Collapsing98.9599.0098.9725.49
EfficientNet-B3Normal condition96.9896.4196.6713.78
Hanging96.3695.9396.1112.64
Oblique stockline95.7495.0095.4312.49
Collapsing99.1298.9099.0114.95
HRNet-W32Normal condition95.4894.7695.0024.69
Hanging95.2794.3794.8824.13
Oblique stockline94.8693.8594.2923.49
Collapsing98.7498.4298.5525.41
Vision TransformerNormal condition96.5195.9896.2626.42
Hanging96.1495.4595.8425.73
Oblique stockline95.9594.9895.4825.98
Collapsing99.0098.8598.9726.91
MDSTMNormal condition98.2497.7298.4228.16
Hanging97.8697.3397.5627.91
Oblique stockline97.5896.9297.2826.73
Collapsing99.4799.3499.3628.37
Note: The bold values indicate the best results among all models under each BF condition.
Table 3. Comparison of Gas Flow Region Features under Different BF Conditions.
Table 3. Comparison of Gas Flow Region Features under Different BF Conditions.
BF ConditionArea (/px2)Perimeter (/px)BrightnessSharpnessSpatiotemporal Feature
Normal condition 3.28 × 10 4 2.27 × 10 4 0.9880.0530.045
Hanging 9.80 × 10 3 1.24 × 10 4 1.0000.0430.065
Oblique stockline 1.71 × 10 4 2.58 × 10 4 0.9990.0410.061
Collapsing 3.51 × 10 4 1.75 × 10 3 0.9960.0010.094
Table 4. Comparison of BF condition recognition with and without multi-domain guidance.
Table 4. Comparison of BF condition recognition with and without multi-domain guidance.
ModelCondition
NormalHangingOblique StocklineCollapsing
With multi-domain guidance98.24%97.86%97.58%99.47%
Without multi-domain guidance93.22%96.81%97.36%98.64%
Note: The bold values indicate the best recognition accuracy under each BF condition.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ji, X.; Han, J.; He, J.; Gui, W. An Online Blast Furnace Condition Recognition Method Based on Spatiotemporal Texture Feature Coupling and Diffusion Networks. Processes 2025, 13, 3416. https://doi.org/10.3390/pr13113416

AMA Style

Ji X, Han J, He J, Gui W. An Online Blast Furnace Condition Recognition Method Based on Spatiotemporal Texture Feature Coupling and Diffusion Networks. Processes. 2025; 13(11):3416. https://doi.org/10.3390/pr13113416

Chicago/Turabian Style

Ji, Xiao, Jie Han, Jianjun He, and Weihua Gui. 2025. "An Online Blast Furnace Condition Recognition Method Based on Spatiotemporal Texture Feature Coupling and Diffusion Networks" Processes 13, no. 11: 3416. https://doi.org/10.3390/pr13113416

APA Style

Ji, X., Han, J., He, J., & Gui, W. (2025). An Online Blast Furnace Condition Recognition Method Based on Spatiotemporal Texture Feature Coupling and Diffusion Networks. Processes, 13(11), 3416. https://doi.org/10.3390/pr13113416

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop