Next Article in Journal
Taste Sensor Assessment of Bitterness in Medicines: Overview and Recent Topics
Previous Article in Journal
Visual Navigation Algorithms for Aircraft Fusing Neural Networks in Denial Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Industry Image Classification Based on Stochastic Configuration Networks and Multi-Scale Feature Analysis

1
Artificial Intelligence Research Institute, China University of Mining and Technology, Xuzhou 221116, China
2
Sunyueqi Honors College, China University of Mining and Technology, Xuzhou 221116, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(15), 4798; https://doi.org/10.3390/s24154798
Submission received: 14 June 2024 / Revised: 18 July 2024 / Accepted: 23 July 2024 / Published: 24 July 2024
(This article belongs to the Section Physical Sensors)

Abstract

:
For industry image data, this paper proposes an image classification method based on stochastic configuration networks and multi-scale feature extraction. The multi-scale features are extracted from images of different scales using deep 2DSCN, and the hidden features of multiple layers are also connected together to obtain more informational features. The integrated features are fed into SCNs to learn a classifier which improves the recognition rate for different categories. In the experiments, a handwritten digit database and an industry hot-rolled steel strip database are used, and the comparison results demonstrate the proposed method can effectively improve the classification accuracy.

1. Introduction

With the development of computer technology, deep learning methods have been widely used for image recognition and classification [1]. Since demonstrating their strong feature representation ability, CNNs with various structures have been developed [2,3,4,5,6] and widely used in image processing and analysis. For the industry image classification task, Masci et al. [7] presented a classification approach based on Max-Pooling CNNs with the features directly extracted from the pixel representation of the steel defect images. Lee et al. [8] proposed a classification method based on CNNs and class activation maps to implement a fast decision-making process. Chen et al. [9] proposed to combine three deep CNN models trained individually, and the average strategy was used to obtain defect classification. Konovalenko et al. [10] proposed a classifier based on two deep residual neural networks in which the hyper-parameters of the optimal model were selected through various investigations. Li et al. [11] proposed a CNN-T model by merging the CNN and Transformer encoder, and obtained an improvement in classification accuracy compared with the pure CNN. Feng et al. [12] introduced a ResNet50 classifier with two additional FcaNet and convolutional block attention modules to deal with steel surface defect data. To deal with the image classification problem, CNNs with different structures that consist of various modules have been developed and obtained better classification accuracy. However, in order to obtain the optimal model, the selection of hyper-parameters in the back-propagation algorithm requires a lot of prior work.
Random neural networks have also been developed for data analysis modeling; the weights and bias of the hidden layers are randomly assigned based on the randomized algorithm [13]. The feed-forward neural networks with random weights (NNRWs) was proposed in [14]. By adding the direct link between the input layer and output layer, Pao et al. [15,16,17] proposed the random vector functional link (RVFL) neural network. A stack RVFL was introduced in [18] using the negative correlation learning strategy. Lu et al. [19] proposed an image data recognition method using two dimensional neural networks with random weights (2D-NNRWs). The stochastic configuration networks (SCNs) in [20] were proposed to improve the learning process by introducing a supervised mechanism. Ensemble methods using base SCN models were introduced in [21,22] to deal with large-scale modeling. The base SCN model was expanded by using multiple hidden layers, and a deep network framework was built in [23,24]. For matrix data, a two dimensional stochastic configuration network (2DSCN) [25] has been developed, which has advantages in image data analytics. On this basis, Li et al. [26] proposed an improved SCN with vision patch fusion, which improves the network’s feature representation ability by extracting randomly fused image features from three-channel images. Li et al. [27] proposed to generate a different convolution kernel with physical meaning using a supervised learning mechanism, and a deep convolutional neural network was constructed for working condition recognition.
In this paper, stochastic configuration networks are used in the proposed image classification method. For feature extraction, compared with a single hidden layer, SCNs with deep structures can extract image features with low computing time. Moreover, because of randomly assigned weights and bias in the stochastic configuration algorithm, images of different scales are used to make up a loss of image details when constructing the deep networks. These multi-scale features are fused through a fully connected layer, and the integrated features are used to build a classifier based on SCNs to complete the identification of different categories. The main contributions are summarized as follows:
(1)
Designed an image classification framework based on SCNs for extracting features from multi-scale images.
(2)
Investigated the influence of different scales’ training data and different network structures on feature extraction results.
(3)
Demonstrated the advantage of the proposed randomized learning method in steel surface defect data.
The remainder of this paper is organized as follows. Section 2 overviews the stochastic configuration networks. Section 3 details the proposed method with deep feature extraction SCNs and a learned classifier. Section 4 elaborates on the experiments and discusses the experimental results. Finally, the conclusion is presented in Section 5.

2. Stochastic Configuration Networks

As improved randomized networks, SCNs were proposed in [20] under a supervised mechanism. Based on the stochastic configuration algorithm, the weights and biases of the hidden layer are gradually configured until an SCN is built. The modeling process of the SCN model with a single hidden layer can be described as follows:
(1)
Given a training dataset { X , T } with N samples, where X = { x 1 , x 2 , , x N } , x i R d , T = { t 1 , t 2 , , t N } , t i R m .
(2)
Suppose the SCN has been configured with L 1 hidden nodes, the output can be calculated by
f L 1 ( X ) = l = 1 L 1 β l g l ( w l T X + b l ) , L = 1 , 2 , . ( f 0 = 0 )
where w l R d , b l R are the weights and bias, β l = [ β l , 1 , β l , 2 , , β l , m ] is the weight of the output layer and g is an activation function. The current residual error is
e L 1 = T f L 1 ( X ) = [ e L 1 , 1 ( X ) , e L 1 , 2 ( X ) , , e L 1 , m ( X ) ] .
(3)
Suppose the error e L 1 does not reach the stop condition, then a new node should be added and the weight and bias w L , b L are configured to calculate the hidden output
h L ( X ) = g L ( w L T X + b L ) .
Denote a set of variables ξ L , q , q = 1 , 2 , , m as follows:
ξ L , q = ( e L 1 , q ( X ) T · h L ( X ) ) 2 h L ( X ) T · h L ( X ) ( 1 r μ L ) e L 1 , q ( X ) T e L 1 , q ( X )
where 0 < r < 1 , μ L is a sequence of real numbers and 0 < μ L 1 r , lim L μ L = 0 .
(4)
The configured weights should satisfy the following constraint condition
ξ L , q 0 , q = 1 , 2 , , m .
Based on the SC-I algorithm in [20], a set of candidate parameters w L n , b L n are randomly assigned under the condition Equation (5). The corresponding parameters are selected as the final values when q = 1 m ξ L , q obtains the maximum value.
(5)
Calculate the output weight β . For the configured and fixed hidden weights and bias, the hidden feature matrix is H = [ h 1 , h 2 , , h L ] with h l = g ( w l X + b l ) , l = 1 , 2 , , L . Then, the output weight β = [ β 1 , β 2 , , β L ] can be calculated
β = arg min β | | f l = 1 L β l h l | | 2 2 .
In the stochastic configuration algorithm, the hidden layer weights w l , l = 1 , 2 , , L are firstly taken from a certain distribution (e.g., Uniform distribution or Gaussian distribution), and the bias b l can be calculated by
b l = w l T X , l = 1 , 2 , , L ,
where X is randomly taken from the training dataset [28]. Then, the supervision mechanism (5) is employed to select proper weight values and add new hidden nodes. The deepSCN is an expansion of the SCN framework and more details about the theory and algorithm can be found in [23].
Similarly, for the image data, i.e., x i R d 1 × d 2 , a 2DSCN model can be built under the supervised constraint Equation (5), and the output of the hidden node is
h l ( X ) = g l ( u l T X v l + b l ) , l = 1 , 2 , . , L ,
where u l R d 1 , v l R d 2 are the hidden layer weight parameters and they are randomly assigned, and the bias is calculated via b l = u l T X v l with randomly sampled data X . The weight of the output layer β l is calculated by Equation (6), and the model output is
f L ( X ) = l = 1 L β l h l .
Moreover, by combining the learning processes of a 2DSCN and deepSCN, the construction of a deep 2DSCN can be described as follows. When the structure of deep networks is in series, the output of the current layer is regarded as the input data of the next layer. Due to the stochastic configuration algorithm, the weights and bias of the hidden layers are randomly generated from a certain distribution and selected with the inequality condition (5) satisfied. The weight parameters of the first hidden layer and the following hidden layers are u l , v l , b l and w k , b k ; the corresponding hidden features are extracted to calculate the parameter β and establish the deep 2DSCN model.

3. Proposed Multi-Scale Image Classification Based on SCNs

In order to extract multi-scale information and improve the feature representation ability, an SCN-based classification method is proposed in this paper. The overview of the proposed method is shown in Figure 1, the multi-scale images are obtained by downsampling from the source image data, and the images of each scale are used to construct a deep SCN model for feature extraction, respectively. These extracted features are combined via a fully connected layer, and an SCN-based classifier is built to obtain the classification result.

3.1. Multi-Scale Feature Extraction Based on deepSCNs

During the pre-processing process, images of different scales are obtained from the original data by performing kernel convolution operators and downsampling. For an image I, the new image of a different scale can be obtained by
I n e w = D p [ C o n v ( I , σ ) ] ,
where σ is the Gaussian kernel and D p is the downsampling operator.
Given the training dataset { X , T } , where X = { x 1 , x 2 , , x N } , x i R d 1 × d 2 and target label T = { t 1 , t 2 , , t N } , t i R m . The Gaussian kernels σ n are used to obtain K sub-dataset { X n , T n } , n = 1 , 2 , , K by
X n = D p [ C o n v ( X , σ n ) ] , n = 1 , 2 , , K .
where σ n , n = 1 , 2 , , K are different kernels and are used to obtain images of different scales. Since the label is unchanged, T n = T , n = 1 , 2 , , K .
For each sub-dataset { X n , T n } , the deep SCN model is built for feature extraction. The detailed modeling process is described in Algorithm 1. Firstly, the images X n are used to extract the feature matrix H 1 of the first hidden layer, i.e.,
H 1 ( X n ) = [ h 1 ( X n ) , h 2 ( X n ) , , h L ( X n ) ] ,
where h l ( X n ) = g ( u l T X n v l + b l ) . If the number of hidden nodes is equal to L m a x or the early stop condition is satisfied, the corresponding weights and bias { u l , v l , b l } are fixed, and the parameter learning of the next layer proceeds.
For the second hidden layer, since the network is connected in series, the input data are the feature vectors in the matrix H 1 ( X n ) which is extracted from the layer above. For each feature vector, the weights and bias w l , b l are randomly assigned through the inequality constraint and the feature matrix H 2 ( X n ) is calculated. The network structure is built step by step with the adding of hidden nodes until the stop condition is met.
Suppose the deep SCN model has M hidden layers, for each sub-dataset, the final feature matrix can be obtained by combining the hidden layer features,
H ( X n ) = [ H 1 ( X n ) , H 2 ( X n ) , , H M ( X n ) ] .
Algorithm 1: Deep SCNs for images.
Sensors 24 04798 i001

3.2. SCN-Based Classifier

For the multi-scale feature extraction, the deep SCN models are established by executing Algorithm 1 on each sub-dataset, and the corresponding feature matrix H ( X n ) is calculated as the output.
A fully connected layer is used to fuse the features extracted from images of different scales; the fully connected feature vector can be calculated by
H = [ H ( X 1 ) , H ( X 2 ) , , H ( X K ) ] .
Then, the feature data H are used for building the SCN classifier, and the classification results are obtained.
The proposed classification method consists of two stages: the multi-scale feature analysis and classifier learning. For multi-scale feature analysis, the sub-images of different scales are obtained by various convolution kernels operators. Image features are extracted from sub-images by building a deep SCN model, and in order to reduce the information loss among hidden layers, the hidden features of each layer are combined together. Then, the connected features are used to build the classification model, and an SCN-based classifier is learned to identify different categories.

4. Performance Evaluation

4.1. Experimental Setup

In this section, the Handwritten Digit Recognition CVL database [29] and the Northeastern University (NEU) surface defect database [30,31] are used for experimental comparison. The CVL database serves as a standard database for digit recognition and classification and is usually used for performance comparison among various algorithms. The NEU dataset has six kinds of typical surface defects of the hot-rolled steel strip, including rolled-in scale (RS), patches (Pa), crazing (Cr), pitted surface (PS), inclusion (In) and scratches (Sc), as shown in Figure 2. The dataset includes 1800 images with the size 299 × 299 . Due to the small sample size, the deep neural network model is prone to overfitting. The window sampling method is applied to expand the number of samples to 10,000, and the new sample images are resized to 64 × 64 . In the experiment, the image data are randomly sampled, with 80 % as the training data and the remaining 20 % as testing data.
In the proposed method, for the parameters used to build SCNs, we take the maximum number of candidate layers M m a x = 5 , the training tolerance error ϵ = 0.0001 , the range of the hidden weights λ = { 0.5 , 1 , 5 , 30 , 50 , 100 } , r = { 0.9 , 0.99 , 0.999 , 0.9999 , 0.99999 } , the maximum number of candidate nodes T m a x = 100 and g is the sigmoid activation function. All the simulation experiments in this paper are performed by using the MATLAB R2023a on a computer with a 3.00 GHz CPU and 16 GB memory.

4.2. Results and Discussion

We first evaluate the classification results of deep SCNs with different numbers of hidden layers. For the CVL database, the number of hidden layers M m a x is set to { 1 , 2 , 3 , 4 , 5 } with L m a x = 500 . Since the SCNs are built by incremental learning in Algorithm 1, the classification accuracy of different hidden nodes L is calculated. As shown in Figure 3, with the increase in hidden layer nodes, the classification accuracy of the training and testing process increases. For M = 1 , M = 2 , M = 3 , we find that the results of the test data are very close when L = 500 , while the computing time shown in Table 1 is completely different. For the NEU dataset, the number of hidden layers M m a x is set with L m a x = 1000 , as shown in Figure 4. For the hidden layer M = 2 , the values of the classification result are close to that of M = 1 before L = 500 ; then, the values get higher with the increment in hidden nodes L. However, the curve values of M = 4 and M = 5 decrease and the classification effect becomes worse.
The comparison results of different M show that large M can extract more useful and reliable feature information for image representation, the built deep SCN model can obtain a good recognition rate, while the computing time is shorter. In addition, due to the limited number of training samples, the deep model may become overfitted. Different from the back-propagation algorithm, the deep network of the proposed method is built by the stochastic configuration algorithm; the modeling process is an incremental learning process, meaning it can adaptively adjust the network structure.
In order to analyze the influence of image size on modeling, the original image data are downsampled to obtain samples of different sizes. The original sizes of the images in the CVL database and the NEU dataset are 28 × 28 and 64 × 64 , respectively. Since the CVL database consists of images with handwritten digits, the image size has a significant impact on molding. As shown in Figure 5, image samples of three different scales are used to build the SCN and deepSCN models. The comparison results show that the recognition rate gets higher as the image size becomes smaller, no matter whether using the SCN model or deepSCN model. The comparison results on the NUE dataset are shown in Figure 6. Different from the CVL dataset, the training recognition rate gets lower as the image size becomes smaller. With the addition of hidden nodes L, the recognition rate keeps increasing, and the model built with image data from scale 2 achieves high testing accuracy. In the deep network, the input of the current hidden layer is the extracted feature map from the above layer. The features extracted from scale 2 images can still be used to characterize the images, whereas the images from scale 3 have a loss of detail, resulting in low recognition rates. Moreover, the computing time is calculated and is shown in Table 2. In general, the larger the image, the longer the computing time. However, compared with the SCN model, the deepSCN model achieves a comparable recognition rate and is less time-consuming.
For the image classification task, the proposed method consists of different parts, including downsampling to obtain images of different scales, building the deep SCN model for feature extraction and building an SCN classifier. For the three parts, the ablation experiments are performed. As shown in Figure 7, the SCN and deepSCN methods are built with only one-scale images (i.e., the original image dataset). The ensemble-SCN and ensemble-deepSCN methods are built with images from multi-scales and their outputs are integrated into the final classification results. The proposed-SCN and the proposed-deepSCN methods are the proposed method in which the connected feature vectors are extracted by the SCN and deepSCN models.
As shown in Figure 7, because the training set only contains images from one scale, the recognition rate of the SCN and deepSCN methods is the lowest; see the yellow curves. With training images collected from multiple scales, the recognition rate of the ensemble-SCN and ensemble-deepSCN methods is greatly improved, while when the number of hidden nodes L exceeds 500, the recognition rate begins to decrease due to overfitting. The proposed method can solve the overfitting problem by adding an additional classifier to optimize the fully connected feature vector. Through the experimental comparison, the proposed-deepSCN method can obtain a good recognition accuracy, and due to the stochastic configuration algorithm, the building of networks with a deep structure takes less cost time than the single hidden layer structure.
In the experiment, the proposed method is compared with some randomized learning methods, such as the 2DNNRW- [19], 2DSCN- [25], deepSCN- [23], DNNE- [18] and SCNE [22]-based methods. Table 3 shows the recognition results of the comparison methods. It can be seen that the 2DNNRW-, 2DSCN- and deepSCN-based methods obtain low classification accuracy as the training data are constructed from single-scale images. The DNNE- and SCNE-based methods built with multi-scale images obtain higher accuracy along with the proposed method. In addition, when an additional classifier is added, the proposed method achieves a higher recognition rate. The comparison results show that the multi-scale feature representation and classifier of the proposed method can effectively improve the classification accuracy.

5. Conclusions

In this paper, an image classification method using SCNs is proposed. Due to its incremental learning algorithm, the network structure of deep SCNs can be adaptively constructed with the weight parameters of hidden layers randomly assigned. The experimental results demonstrate the advantage of the proposed method for solving classification problems, using the multi-scale extracted features, and an additional classifier can effectively improve classification accuracy. For future work, the weight configuration of the convolutional kernel can be discussed to optimize SCNs’ modeling on image data.

Author Contributions

Conceptualization, Q.W.; methodology, Q.W.; software, Q.W.; validation, Q.W.; formal analysis, Q.W.; investigation, Q.W., D.L., H.T., Y.Q. and D.Z.; resources, Q.W.; data curation, Q.W., D.L., H.T. and Y.Q.; writing—original draft preparation, Q.W.; writing—review and editing, Q.W. and D.Z.; visualization, Q.W. and D.Z.; supervision, Q.W.; project administration, Q.W.; funding acquisition, Q.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Fundamental Research Funds for the Central Universities (2022QN1045).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Rawat, W.; Wang, Z. Deep Convolutional Neural Networks for Image Classification: A Comprehensive Review. Neural Comput. 2017, 29, 2352–2449. [Google Scholar] [CrossRef]
  2. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; Volume 25, pp. 1097–1105. [Google Scholar]
  3. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  4. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  5. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 1–14. [Google Scholar]
  6. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual Attention Network for Image Classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6450–6458. [Google Scholar]
  7. Masci, J.; Meier, U.; Ciresan, D.; Schmidhuber, J.; Fricout, G. Steel defect classification with Max-Pooling Convolutional Neural Networks. In Proceedings of the 2012 International Joint Conference on Neural Networks (IJCNN), Brisbane, QLD, Australia, 10–15 June 2012; pp. 1–6. [Google Scholar]
  8. Lee, S.Y.; Tama, B.A.; Moon, S.J.; Lee, S. Steel Surface Defect Diagnostics Using Deep Convolutional Neural Network and Class Activation Map. Appl. Sci. 2019, 9, 5449. [Google Scholar] [CrossRef]
  9. Chen, W.; Gao, Y.; Gao, L.; Li, X. A New Ensemble Approach based on Deep Convolutional Neural Networks for Steel Surface Defect classification. Procedia CIRP 2018, 72, 1069–1072. [Google Scholar] [CrossRef]
  10. Konovalenko, I.; Maruschak, P.; Brevus, V. Steel Surface Defect Detection Using an Ensemble of Deep Residual Neural Networks. J. Comput. Inf. Sci. Eng. 2021, 22, 014501. [Google Scholar] [CrossRef]
  11. Li, S.; Wu, C.; Xiong, N. Hybrid Architecture Based on CNN and Transformer for Strip Steel Surface Defect Classification. Electronics 2022, 11, 1200. [Google Scholar] [CrossRef]
  12. Feng, X.; Gao, X.; Luo, L. A ResNet50-Based Method for Classifying Surface Defects in Hot-Rolled Strip Steel. Mathematics 2021, 9, 2359. [Google Scholar] [CrossRef]
  13. Scardapane, S.; Wang, D. Randomness in neural networks: An overview. WIREs Data Min. Knowl. Discov. 2017, 7, e1200. [Google Scholar] [CrossRef]
  14. Schmidt, W.F.; Kraaijveld, M.A.; Duin, R.P.W. Feedforward neural networks with random weights. In Proceedings of the 11th IAPR International Conference on Pattern Recognition, The Hague, The Netherlands, 30 August–3 September 1992; pp. 1–4. [Google Scholar]
  15. Pao, Y.H.; Takefuji, Y. Functional-link net computing: Theory, system architecture, and functionalities. Computer 1992, 25, 76–79. [Google Scholar] [CrossRef]
  16. Pao, Y.H.; Phillips, S.M.; Sobajic, D.J. Neural-net computing and the intelligent control of systems. Int. J. Control 1992, 56, 263–289. [Google Scholar] [CrossRef]
  17. Igelnik, B.; Pao, Y.H. Stochastic choice of basis functions in adaptive function approximation and the functional-link net. IEEE Trans. Neural Netw. 1995, 6, 1320–1329. [Google Scholar] [CrossRef] [PubMed]
  18. Alhamdoosh, M.; Wang, D. Fast decorrelated neural network ensembles with random weights. Inf. Sci. 2014, 264, 104–117. [Google Scholar] [CrossRef]
  19. Lu, J.; Zhao, J.; Cao, F. Extended feed forward neural networks with random weights for face recognition. Neurocomputing 2014, 136, 96–102. [Google Scholar] [CrossRef]
  20. Wang, D.; Li, M. Stochastic Configuration Networks: Fundamentals and Algorithms. IEEE Trans. Cybern. 2017, 47, 3466–3479. [Google Scholar] [CrossRef]
  21. Huang, C.; Li, M.; Wang, D. Stochastic configuration network ensembles with selective base models. Neural Netw. 2021, 137, 106–118. [Google Scholar] [CrossRef]
  22. Wang, D.; Cui, C. Stochastic configuration networks ensemble with heterogeneous features for large-scale data analytics. Inf. Sci. 2017, 417, 55–71. [Google Scholar] [CrossRef]
  23. Wang, D.; Li, M. Deep Stochastic Configuration Networks with Universal Approximation Property. In Proceedings of the 2018 International Joint Conference on Neural Networks, Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–8. [Google Scholar]
  24. Wang, D.; Felicetti, M.J. Stochastic Configuration Machines for Industrial Artificial Intelligence. arXiv 2023, arXiv:2308.13570. [Google Scholar]
  25. Li, M.; Wang, D. 2-D Stochastic Configuration Networks for Image Data Analytics. IEEE Trans. Cybern. 2019, 51, 359–372. [Google Scholar] [CrossRef]
  26. Li, R.; Jiao, W.; Zhu, Y. Improved stochastic configuration networks with vision patch fusion method for industrial image classification. Inf. Sci. 2024, 670, 120570. [Google Scholar] [CrossRef]
  27. Li, W.-T.; Tong, Q.-Q.; Wang, D.-H.; Wu, G.-C. Research on fused magnesium furnace working condition recognition method based on deep convolutional stochastic configuration networks. Acta Autom. Sin. 2024, 50, 527–545. [Google Scholar]
  28. Dudek, G. Generating random weights and biases in feedforward neural networks with random hidden nodes. Inf. Sci. 2019, 481, 33–56. [Google Scholar] [CrossRef]
  29. Kleber, F.; Fiel, S.; Diem, M.; Sablatnig, R. CVL-DataBase: An Off-Line Database for Writer Retrieval, Writer Identification and Word Spotting. In Proceedings of the 2013 12th International Conference on Document Analysis and Recognition, Washington, DC, USA, 25–28 August 2013; pp. 560–564. [Google Scholar]
  30. Song, K.; Yan, Y. A noise robust method based on completed local binary patterns for hot-rolled steel strip surface defects. Appl. Surf. Sci. 2013, 285, 858–864. [Google Scholar] [CrossRef]
  31. He, Y.; Song, K.; Meng, Q.; Yan, Y. An End-to-End Steel Surface Defect Detection Approach via Fusing Multiple Hierarchical Features. IEEE Trans. Instrum. Meas. 2020, 69, 1493–1504. [Google Scholar] [CrossRef]
Figure 1. The overview of the proposed method.
Figure 1. The overview of the proposed method.
Sensors 24 04798 g001
Figure 2. The NEU surface defect data.
Figure 2. The NEU surface defect data.
Sensors 24 04798 g002
Figure 3. Recognition rate of different M on CVL database. (a) Training performance; (b) testing performance.
Figure 3. Recognition rate of different M on CVL database. (a) Training performance; (b) testing performance.
Sensors 24 04798 g003
Figure 4. Recognition rate of different M on NEU database. (a) Training performance; (b) testing performance.
Figure 4. Recognition rate of different M on NEU database. (a) Training performance; (b) testing performance.
Sensors 24 04798 g004
Figure 5. Performance comparison on images with different sizes on CVL database. (a,b) Training and testing results of the SCN model; (c,d) training and testing results of the deepSCN model.
Figure 5. Performance comparison on images with different sizes on CVL database. (a,b) Training and testing results of the SCN model; (c,d) training and testing results of the deepSCN model.
Sensors 24 04798 g005
Figure 6. Performance comparison on images with different sizes on NEU database. (a,b) Training and testing results of the SCN model; (c,d) training and testing results of the deepSCN model.
Figure 6. Performance comparison on images with different sizes on NEU database. (a,b) Training and testing results of the SCN model; (c,d) training and testing results of the deepSCN model.
Sensors 24 04798 g006
Figure 7. Performance comparison with different classifiers on NEU database. (a,b) Training and testing performance with SCN-based feature extraction; (c,d) training and testing performance with deepSCN-based feature extraction.
Figure 7. Performance comparison with different classifiers on NEU database. (a,b) Training and testing performance with SCN-based feature extraction; (c,d) training and testing performance with deepSCN-based feature extraction.
Sensors 24 04798 g007
Table 1. Computing time of deep SCN model with different layers.
Table 1. Computing time of deep SCN model with different layers.
M = 1M = 2M = 3M = 4M = 5
CVL3004.3078826.2544417.0548275.0342198.1505
NEU4084.9769679.8829386.1502317.3123161.1162
Table 2. Computing time of different image sizes.
Table 2. Computing time of different image sizes.
ScaleImage SizedeepSCNSCN
CVLscale 128 × 28438.75384349.2825
scale 214 × 14383.05003823.0818
scale 37 × 7403.40733910.1134
SEUscale 164 × 64302.72454089.6932
scale 232 × 32240.27812924.6824
scale 316 × 16221.68832583.3933
Table 3. Comparison results on NEU dataset.
Table 3. Comparison results on NEU dataset.
ModelAccuracyRecallPrecisionF1 Score
2DNNRW0.76910.61670.88640.7275
2DSCN0.81680.69490.91900.7914
deepSCN0.81250.68730.91700.7857
DNNE0.79170.65560.90070.7589
SCNE0.85380.75740.93840.8382
Proposed0.85590.76060.93980.8407
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Q.; Liu, D.; Tian, H.; Qin, Y.; Zhao, D. Industry Image Classification Based on Stochastic Configuration Networks and Multi-Scale Feature Analysis. Sensors 2024, 24, 4798. https://doi.org/10.3390/s24154798

AMA Style

Wang Q, Liu D, Tian H, Qin Y, Zhao D. Industry Image Classification Based on Stochastic Configuration Networks and Multi-Scale Feature Analysis. Sensors. 2024; 24(15):4798. https://doi.org/10.3390/s24154798

Chicago/Turabian Style

Wang, Qinxia, Dandan Liu, Hao Tian, Yongpeng Qin, and Difei Zhao. 2024. "Industry Image Classification Based on Stochastic Configuration Networks and Multi-Scale Feature Analysis" Sensors 24, no. 15: 4798. https://doi.org/10.3390/s24154798

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop