Next Article in Journal
Globally Guided Deep V-Network-Based Motion Planning Algorithm for Fixed-Wing Unmanned Aerial Vehicles
Previous Article in Journal
Probability-Based LIDAR–Camera Calibration Considering Target Positions and Parameter Evaluation Using a Data Fusion Map
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exploration of MPSO-Two-Stage Classification Optimization Model for Scene Images with Low Quality and Complex Semantics

Department of Information Engineering, Engineering University of PAP, Xi’an 710086, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(12), 3983; https://doi.org/10.3390/s24123983
Submission received: 30 April 2024 / Revised: 30 May 2024 / Accepted: 17 June 2024 / Published: 19 June 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Currently, complex scene classification strategies are limited to high-definition image scene sets, and low-quality scene sets are overlooked. Although a few studies have focused on artificially noisy images or specific image sets, none have involved actual low-resolution scene images. Therefore, designing classification models around practicality is of paramount importance. To solve the above problems, this paper proposes a two-stage classification optimization algorithm model based on MPSO, thus achieving high-precision classification of low-quality scene images. Firstly, to verify the rationality of the proposed model, three groups of internationally recognized scene datasets were used to conduct comparative experiments with the proposed model and 21 existing methods. It was found that the proposed model performs better, especially in the 15-scene dataset, with 1.54% higher accuracy than the best existing method ResNet-ELM. Secondly, to prove the necessity of the pre-reconstruction stage of the proposed model, the same classification architecture was used to conduct comparative experiments between the proposed reconstruction method and six existing preprocessing methods on the seven self-built low-quality news scene frames. The results show that the proposed model has a higher improvement rate for outdoor scenes. Finally, to test the application potential of the proposed model in outdoor environments, an adaptive test experiment was conducted on the two self-built scene sets affected by lighting and weather. The results indicate that the proposed model is suitable for weather-affected scene classification, with an average accuracy improvement of 1.42%.

1. Introduction

Multimedia resources such as images and videos have become the main media through which people acquire information nowadays [1]. How to quickly find the desired semantics from a large number of redundant scene images has become a popular task in computer vision in recent years. Image scene semantic classification has made contributions to improving work efficiency in various industries, such as rescue agencies which need to determine the type of disaster scene to achieve timely search and rescue in affected areas; public security organs which need to assess the topography of riot scenes to achieve effective control; and transportation agencies which need to judge the distribution of facilities in a short period of time after the planning of paths and so on. However, it is often affected by harsh environmental conditions and the large number of scene elements, along with strong interference, which results in low-quality first-line image information, which is unfavorable for subsequent classification. For example, misclassifying lanes, bridges, embankments, and rural roads can create obstacles for commanders’ decision-making.
Scholars have paid more attention to the setting of the network architecture in order to improve the accuracy of complex semantic scene image classification, constantly innovating methods and achieving certain results [2,3,4]. However, the test objects of the above methods are mostly international standard clear image scene sets, seldom considering the negative impact of low-quality scene images on classification [5]. Even though scholars have gradually discovered that the underlying cause of the decline in classification performance is low-quality images [6,7], the conclusions are mostly applicable to remote sensing images [8] or clear image sets with artificial noise [9], without considering low-quality real-measured images that would be affected by various external factors. For example, network propagation delays can degrade the high-definition quality of images [10], severe weather and dense smoke environments can cause semantic blurring of images [11], poor lighting conditions can lead to misreading of image information [12], etc. These all indicate that it is necessary to explore the causes of low-quality real-measured images and find matching classification methods.
However, scholars have found that even with high-precision CNN architectures, classification performance can still drop significantly when faced with low-quality images [9]. This is because the internal pixel structure of the image neighborhood in the actual environment degrades [13], causing a sharp drop in image resolution and leading to poor image quality [14]. Therefore, it is imperative to add image preprocessing operations [7] to improve the classification effect of low-quality scene image classification models.
To develop a practical model for low-quality scene image classification, this paper makes two innovative contributions: 1. We propose a new method. By refining 21 existing scene classification methods based on two focal points of structural complexity and semantic clarity, a two-stage classification optimization model based on the Mix-optimization PSO Algorithm (MPSO) is proposed. To this end, we introduce an improved hybrid PSO algorithm to optimize the parameters of the Level-1 classification module DCNN-ELM network and the Level-2 reconstruction module SRCNN network in a nested iterative manner, thereby improving the semantic classification effect of low-quality complex scene images. 2. We build a real-measured low-quality dataset. By constructing multi-complex semantic news scene datasets and low-quality datasets under the influence of visible light and environment, the research gap in low-quality image datasets is filled, providing data support for the three different types of experiments with existing scene datasets.
The structure of the paper is arranged as follows: Section 2 summarizes the existing classification methods based on the classification objects. Section 3 introduces the model architecture and principles proposed based on Section 2. Section 4 conducts feasibility and applicability tests. Section 5 focuses on discussing the experimental results and the contributions made by the model, reflecting on the shortcomings and providing prospects. Section 6 concludes the entire paper.

2. Related Research Progress

After collecting a large amount of literature, this article summarizes the current scene classification methods into two categories based on structural complexity and semantic clarity.

2.1. Development Status of Scene Image Classification Methods Based on Structural Complexity

Due to the rich semantic content of scene images, the structural complexity has a significant impact on classification performance. To illustrate this issue, we leverage the results of the previously published research [3] and use the DCNN-ELM algorithm to perform classification tests on our self-built military weapon image datasets and the low-quality scene datasets built in this paper. The classification results for representative scene images are recorded in Figure 1. It is evident that for images with clear semantic content such as “warplane” and “blast”, the main subject is clearly distinguishable from the background, resulting in a high classification accuracy rate of up to 99%. However, for images with high natural environmental influence such as “forest” and “fire”, which are more susceptible to natural environmental influence, the classification performance drops to 84% due to the difficulty in distinguishing the subject from the background. Below, we will detail scene classification methods for different semantic complexities.
For efficient recognition of rough shapes and geometric structures in single-target scenarios [2], a large number of methods have emerged that extract features first and then combine them with classifiers to achieve recognition. Currently, scholars are constantly innovating in feature extraction methods, such as extracting Gist descriptor underlying features [15], fusing local features [16], or fusing features at different levels [17]. In order to further improve accuracy, the Bag-of-Visual-Words (BoVW) and Gaussian Mixture Model (GMM) have been quickly discovered and applied in scene image classification for their high robustness. In 2024, scholars used the Boruta feature selection algorithm for geomorphic features to distinguish archaeological objects [18]. However, the aforementioned techniques require manual feature extraction and are not widely applicable. They also have semantic ambiguity issues for scenes with cluttered backgrounds. For example, items involved in indoor blind training scenarios are often complex and confusing [19], and outdoor traffic scenes often contain multiple elements such as pedestrians, vehicles, and buildings [20,21]. The similarity and diversity between these image pixels [22,23] make image interference information more abundant, and traditional classification methods lose feature details, resulting in low recognition rates.
For the accurate recognition of complex semantic scenes mentioned above, scholars have frequently innovated around network optimization and parameter adjustment methods, mainly reflected in the methods outlined below.
One approach is to deepen the number of network layers. In order to overcome the problem of insufficient training data, scholars have increased the number of network layers on the basis of convolutional neural network (CNN) architecture to extract more semantic features [22], and have successively developed GoogLeNet [23], the ResNet model [24], and the ResNet model combined with Inception architecture [25] to optimize classification performance. However, multi-layer convolution inevitably loses key information, resulting in limited accuracy.
A second approach is to optimize the network architecture. In 2020, scholars streamlined the classification structure to ensure the classification accuracy and execution speed of indoor scenes [17]. For example, in 2021, scholars combined low-level manual features with deep CNN multi-stage features (HF-MSFs) based on the spatial pyramid structure [19], and improved the architecture to achieve multi-feature adaptive fusion [26] in 2023. In the same year, FcA et al. first applied the upsampling structure to remote sensing image scene classification [8] and achieved good results.
The third approach is to innovate classifiers. In recent years, CNN-ELM based on CNN architecture [11] has been widely applied in image classification with the advantages of high efficiency and high accuracy [27], which also provides ideas for scene image classification in this article.
The fourth approach is to improve the initialization method of network parameters. Scholars have designed a network parameter assignment method [4,6,28] to enhance the accuracy of the classification network. Additionally, in 2023, scholars encoded image segments into sequences and combined them with convolutional architectures to form the Segment Anything Model (SAM) [29] for automatic semantic segmentation of scene images. In 2024, this architecture achieved good results in medical image detection, but its precision for classifying unclear cancer cell images was not high [30].
However, most of the above methods are based on noise-free and clear images, and the high-accuracy classification of low-quality images is still in the initial stage of research.

2.2. Progress in Low-Quality Image Classification Methods Based on Semantic Clarity

Low image quality often affects subsequent classification [20,21]. To illustrate this issue, we applied the DCNN-ELM algorithm to classify and compare the same scene using the IC Image Sets and WD Image Sets established in this paper. For example, as shown in Figure 2, the PSNR value of the left image is 0.89 dB higher than that of the right image, clearly indicating that the right image’s edge is deformed due to the inherent resolution of the equipment. In the end, the correct classification rate of the left image is 1.2% higher than that of the right image.
Of course, image quality is not only affected by the inherent resolution of photography equipment, but also by factors such as weather conditions, geographical environment, and time period, which can lead to a decrease in semantic clarity and thus affect classification results. For example, underwater mapping tasks that are subject to the Refraction Effect [30] usually have to take into account multiple factors affecting the image quality such as the position of the capturing camera (above or below water), water refraction, water depth, and light propagation, which undoubtedly have a detrimental effect on the subsequent scene categorization. At present, there are two main optimization schemes for improving the accuracy of low-quality image classification: classification image preprocessing and modifying the classification network structure.
The first type focuses on adding image preprocessing operations before the classification network [31] to enhance image features. Taking semantic fuzzy image classification affected by noise as an example, early scholars made full use of underlying features such as color and texture or HOG features to design filter structures [32] to enhance detail features. However, it is unavoidable to find that an “enhanced image” is not a “clear image”, as there is a gap in feature distribution between the two which cannot benefit from low-quality image classification under the influence of complex scenes [7], nor can it guarantee that similar structural areas are uniformly repaired [13]. Subsequently, the domain adaptation method [33] became mainstream, and the use of neural network architecture became the focus of research.
Scholars have found that image preprocessing methods based on super-resolution reconstruction technology [34] are particularly effective for classification improvement. Taking infrared image classification, which has gradually become a research hotspot since 2020, as an example, feature-based reconstruction methods emerged in 2021 [35], and models with covariant function regularization were proposed in 2022 [36]. However, the above processing operations mostly ignore the optimization of subsequent classification networks, and the design of quantitative indicators is not comprehensive enough.
The second type focuses on optimizing the structure of low-quality image classification networks to avoid the loss of critical information. Taking the limited dataset and difficulty in obtaining high-quality remote sensing image classification as an example, in order to achieve high accuracy classification based on complex spectral features, network deployment was innovated based on CNN architecture in 2018 [37]. Extreme structure DenseNet was launched in 2019 [38], while scholars first used upsampling structure for image scene classification in 2021, and designed DRSNet architecture [8]. Finally, lightweight newCNN was proposed in 2022 [39].
Taking inspiration from the above, on the one hand, image reconstruction pre-operation is introduced in order to better complete the task of self-collected low-quality image classification, and on the other hand, the existing classification network architecture is improved to comprehensively enhance classification accuracy.

3. Two-Stage Classification Optimization Algorithm Model Based on MPSO Method

To explore the best approach for classifying low-quality scenes that satisfy complex structures and semantic ambiguity, this paper proposes to construct a two-stage hyperparameter particle swarm optimization strategy based on a convolutional neural network architecture, as shown in Figure 3.
An improved fusion-based particle swarm optimization (MPSO) algorithm is introduced to iteratively optimize the parameters of the Level-1 classification module DCNN-ELM network and the Level-2 reconstruction module SRCNN network. Each stage’s selected particle combination represents one configuration of the optimized integrated network.
Two types of parameters, the size and number of connection weights between the input, output, and implicit layers of the first stage reconstruction network constitute the initial population size. The population in each type of parameter is identified using [p_1, p_2, p_3 …p_m]. The five elements of the second-stage classification network, namely the number of convolutional layers (consistent with the number of pooling layers), the step size, the convolution sum size, the number of convolutional kernels, and the learning rate, which together constitute the initial population size, are regarded as hyperparameter populations and are identified using [p_11, p_12, p_13 …p_1k] [p_21, p_22, p_23 …p_2k] …[p_m1, p_m2, p_m3 …p_mk] for identification. The first stage is based on SRCNN architecture, using Mix-optimization PSO Algorithm (MPSO) group exploration thinking to calculate parameter particle fitness and find the best set of hyperparameters suitable for the current reconstruction network. Then, entering the second stage, based on DCNN-ELM architecture, randomly initialize the network parameters and iterate. When the accuracy is close to 70%, calculate the occurrence probability of each category after docking with ELM. If the current optimal particle fitness value is met, call the ELM layer, calculate the weight of the hidden layer, and cache the β matrix.
In summary, the first stage of the group involves searching space to find the best set of reconstruction hyperparameters, and the second stage of the group uses the sub-search space to find the best set of classification architecture parameters for these reconstructed layers. In this way, the combination of hyperparameters selected after completing each layer cycle constitutes a new classification model. Each section is described in detail below.

3.1. SRCNN Structure for Level-1 Reconstruction Stage

Considering the essential influence of image resolution, in order to realize the preprocessing operation of low-quality images, the SRCNN model in Figure 4 is selected for the level-1 reconstruction of this model, and the calculation method is shown in Table 1.
Considering the running time, the initial value is chosen to be 64, the second is chosen to be 32, and the initial values of the three-layer filters f 1 × f 1 , f 2 × f 2 , f 3 × f 3 are 9 × 9 , 1 × 1 , 5 × 5 , respectively. The learning rate of the first two layers of the network is η 1 = η 2 = 10 4 , and the learning rate of the last layer is η 3 = 10 5 . The picture channel is chosen to be 1.

3.2. DCNN-ELM Structure for Level-2 Classification Stage

After level-1 completes a small cycle of reconstruction network parameter selection for each execution, the DCNN-ELM network architecture is selected and enters level-2 for iterative parameter selection.
After a small cycle of selecting and reconstructing network parameters is completed in level-1, the parameter iteration based on DCNN-ELM network architecture is carried out in level-2. With the increase in the number of convolution layers, the image pixel value represents more details of the image after being acted upon by the convolution kernel, which is helpful for the judgment of the final result. At the same time, weight sharing in CNN makes the number of parameters only related to the size of the convolution kernel and the number of feature maps. This is reflected in the smaller convolution kernel, which helps to reduce parameter complexity. Therefore, information such as the number of feature maps, the size of the convolution kernel, and the step size constitute the initial population scale in the second stage, and the initialization model setting is set as shown in Figure 5. The calculation method of feature extraction is listed in Table 2.
This model takes the output (n × 1 dimensions) after three times pooling as the input of ELM, and the hidden layer H contains n u neurons, i.e., any node j 1 , n u , which corresponds to m neurons in the output layer. Then, for N training samples i 1 , N , the mapping relationship between the input feature X i n p u t = x i 1 , x i 2 , , x in T and output label Y o u t p u t = y i 1 , y i 2 , , y im T can be calculated by Equation (1):
Y o u t p u t = j = 1 n u β j g ( W j X i n p u t + b j )
W j = w j 1 , w j 2 , , w j n denotes the connection weight of the n-th neuron of the input to the nu-th neuron of the hidden layer. β j = β 1 j , β 2 j , , β m j denotes the connection weight of the nu-th neuron of the hidden layer to the m-th neuron of the output.
Normally, the actual output is T. If the number of neurons in the hidden layer is equal to the number of samples in the training set, the actual output is simplified according to Equation (1) as:
H β = T
T is the matrix transpose of T and is the implicit layer output matrix, in the form of Equation (3):
H = g ( w 1 x 1 + b 1 ) g ( w 2 x 1 + b 2 ) g ( w n u x 1 + b n u ) g ( w 1 x 2 + b 1 ) g ( w 2 x 2 + b 2 ) g ( w n u x 2 + b n u ) g ( w 1 x N + b 1 ) g ( w 2 x N + b 2 ) g ( w n u x N + b n u ) N × n u
w and b are usually kept constant during training, so usually β is the only parameter to be learned in ELM. This requires the introduction of the Moore–Penrose generalized inverse H to solve for β and obtain β = H T , which gives the output after training is complete:
Y o u t p u t = g ( X i n p u t ) β = h ( X i n p u t ) H T
The ELM layer is invoked based on the classification accuracy, and the MPSO algorithm is applied in combination with the cache weights to iteratively search for optimization of the population in this loop, in order to complete the optimal parameter combinations of the classification network under the current optimal reconstruction network architecture.

3.3. Mix-Optimization PSO Algorithm

As seen in Section 3.1 and Section 3.2, the basic CNN architecture has narrow hyperparameters such as weights, filter sizes, numbers, step sizes, and activation functions, as well as generalized hyperparameters such as the learning rate and the order of the number of layers, which collectively determine the convergence performance of the model and greatly affect the classification accuracy. Given that CNN hyperparameter fine-tuning is a multimodal function optimization problem, scholars currently prefer to combine the Gradient Descent method and optimize the network hyperparameters with the PSO method [40] to improve the classification performance, which is reflected in Equation (5).
v m ( t + 1 ) = ω v m ( t ) + c 1 a 1 p b e s t m x m ( t ) + c 2 a 2 g b e s t m x m ( t ) x m ( t + 1 ) = x m ( t ) + v m ( t + 1 )
where t is the number of iterations, ω is the inertia weight, r 1 and r 2 are constants between [0, 1], and a1 and a2 are the learning factors. p b e s t m stands for the individual optimum, g b e s t m stands for the group optimum, v m t + 1 is the post-update velocity, and x m t + 1 is the post-update position.
However, it cannot be ignored that the multi-layered CNN network parameters require more particles than the classic particle swarm optimization algorithm, and the complex iteration number leads to a decline in optimization ability, thereby adversely affecting the classification performance. To avoid this problem as much as possible, many kinds of literature have proposed methods to improve the inertia weight (ω), social coefficient (a1), and cognitive coefficient (a2), and have successfully been applied to DCNN [27] and CNN-ELM [5] network classification. Inspired by the existing research [40], we proposed an MPSO optimization mode based on the iteration combination, which on the one hand weakens the group cognition in the early stage of the iteration cycle and strengthens the overall search ability later, accelerates the adjustment of the learning factor in real-time and correcting the particle optimization trajectory in a timely manner. On the other hand, we optimized the inertia weight in a stage-wise manner to assist the learning rate. The strategy is as follows:
① Changing learning factors
a 1 = a 1 max + a 1 max a 1 min * t t max a 2 = a 2 max + a 2 max a 2 min * t t max
where a 1 max , a 1 min are often the maximum and minimum values of a 1 , which are taken as 2.5 and 0.5, respectively. a 2 max , a 2 min are the maximum and minimum values of a 2 , which are taken as 2.5 and 0.5, respectively. t is the current number of iterations, and t max is the maximum number of iterations.
② Optimizing inertia weights
ω t = 0.9 t < 0.2 t max 1 1 + e 10 t 2 t max / t max o t h e r w i s e
The initial population size at the time of the experiment was 40, and the spatial dimensions were consistent with the weight threshold.

4. Experiment Results

As shown in Figure 6, the experiment mainly consists of two modules. The first module mainly tests the feasibility of the proposed architecture. For the existing scene set and the scene set formed by intercepting low-quality news keyframes, the rationality of the classification structure and the necessity of the pre-reconstruction link are verified, respectively, so as to comprehensively evaluate the classification adaptability and high accuracy of the model in this paper.
The second module mainly evaluates the applicability of the model. For the low-quality image set of news scenes and the self-built low-quality image set affected by the environment, the application potential of the model is explored by realizing multi-scene semantic classification with different clarity.

4.1. Exploring the Rationality of Classification Model Architecture

In this part, three types of datasets are selected as shown in Table 3, and the validation set and test set are set according to 2:8.
Moreover, we have completed the fitness evaluation of our model on the 15-scene dataset and selected an appropriate batch size value as in Figure 7 to demonstrate the advantages of introducing MPSO for parameter adjustment.
Generation refers to the number of genetic cycles, and accuracy rate represents the average accuracy rate after classification. We found that a high accuracy rate can be maintained under the condition of fewer parameters, and the accuracy rate tends to stabilize after seven generations of cycles. However, the accuracy rate starts to decrease with the growth in batch size, and the accuracy rate tends to be optimal in the genetic cycle when the batch size is chosen to be 25.

4.1.1. Intuitive Comparison of Multi-Method Classification Effects

In order to more comprehensively illustrate the applicability and classification performance of the classification structure proposed in this paper, we roughly classify the 20 scene classification methods mentioned in the previous section into three major categories: feature extraction improvement, network architecture updating, and classifier design. We carry out comparative experiments with the methods in this paper, and record the classification results in Table 4.
For the 15-scene dataset, the algorithm proposed in this paper is 1.54% higher than the best existing method, ResNet-ELM. For the MIT-Inoor dataset, the algorithm proposed in this paper is 0.18% higher than the best available method Places365-VGG. For Sun datasets, although 0.1% lower than the best available method CNNang-SVM, the comprehensive advantage remains ahead. After an in-depth analysis of the reasons, we found the following results:
Accuracy improves with a decrease in the number of classification labels. Although a large dataset helps to improve the classification, the accuracy decreases when the variety of classification labels increases. Therefore, the 15-scene dataset with fewer labels has more advantageous classification accuracy compared to the other two datasets.
Along with the increase in structure complexity, the accuracy decreases. The overall effect is that the recognition accuracy of indoor scenes is lower than that of outdoor scenes. We attribute this to the fact that indoor scenes contain too many trivial structures, which are semantically rich and complex and adversely affect feature judgments. Optimizing the classifier is more conducive to improving the classification effect of the outdoor scene dataset; improving the classification network architecture is more advantageous in classifying the indoor scene dataset. Our method still maintains excellent performance for complex scenes due to the advantages of combining the above two approaches: initializing the network weights using swarm intelligence algorithms, and streamlining the network structure against ELM to extract more detailed features.

4.1.2. Comparison of Classification Effect of Multi-Semantic Labels

We take 15 scenes with excellent classification performance as an example. The number of each tag is recorded in Figure 8, and the classification results are recorded in the form of a confusion matrix in Figure 9. The comprehensive analysis reveals that:
Consistent with the previous conclusions, the proposed architecture achieves good classification results in outdoor scenes such as “coast”, “forest”, and “suburb”, while it is inferior in indoor scenes with high confusion levels such as “kitchen”, “living room”, and “bedroom”.
Meanwhile, we were surprised to find that the “suburb” scene dataset, although smaller in number than the other scene sets, generally has a high image resolution, which contributes to a classification accuracy of 99.99%. This also means that datasets with higher classification accuracy tend to have higher PSNR values. Inspired by this, we propose to think about whether image resolution affects classification performance, i.e., does the addition of a pre-reconstruction session contribute to higher classification accuracy?

4.2. Experiments on the Necessity of Adding a Pre-Reconstruction Architecture

4.2.1. The Effect of Low-Quality Images (Low Resolution) on the Degree of Classification Performance

Due to the disadvantage of the limited number of existing low-quality datasets, we reduce the resolution of the 15-scene dataset images, i.e., select different upscale factors to realize the Bicubic interpolation operation. Figure 10 shows the visual effects and PSNR values of outdoor scenes “coast”, “highway”, “open country”, “forest”, and indoor scenes “living room” and “bedroom” under different upscales.
The PSNR values of outdoor scenes with high classification accuracy are generally higher than those of indoor scenes. Based on semantic complexity, the 15 scenes are roughly divided into three categories: indoor, outdoor, and mixed indoor-outdoor scenes. The proposed method is used to classify them, and the results are recorded in Table 5.
As the value of upscale increases, the classification performance continuously declines. When upscale is set to 3, the accuracy decreases by 0.37% compared to when upscale is set to 2, and when upscale is set to 4, the accuracy decreases by 0.53% compared to when upscale is set to 3. This shows that as the image resolution decreases, the loss of key information gradually increases the impact on classification accuracy.
Upon further investigation, it was found that the degradation of outdoor image classification performance after degradation was higher than that of indoor images. For example, when the upscale was changed from 2 to 3, the outdoor image classification performance decreased by 0.43% while the indoor image classification performance decreased by 0.25%. To investigate the impact of low-resolution scene images on classification performance, we recorded the classification results at different resolutions in Figure 11.
From the figure, it can be concluded that the impact of degradation on semantic complexity and confusing scenes is greater than that on scenes with a single structure. In particular, the impact on outdoor scenes that are easily affected by the environment (such as “coast” and “open country”) is greater than that on indoor scenes (such as “office” and “living room”). We hypothesize that this is due to the omission of typical features and loss of edge details in the extraction process of low-resolution images. For example, “coast” and “open country” are easily confused, which shows the importance of adding pre-reconstruction operation before classification.

4.2.2. The Degree of Contribution of the Pre-Reconstruction Structure to the Classification Performance Improvement in This Paper

This part of the research object comes from 1068 short videos on the CCTV website, from which 4861 keyframes are selected and categorized into “Stage”, “Conference”, “Woodland”, “Fire”, “Building”, “Flood” and “Highway”. The specific composition and quality ratio are given in Table 6. The test set and training set are divided according to the ratio of 2:8, and the input size is uniformly 256 × 256.
After the performing degradation process (upscale taken as 2, 3, and 4) for the above image set, based on the classification architecture proposed in this paper, six different preprocessing methods were compared, including Bilateral Filter, dehazing method, SC, ANR, SRCNN, and MPSO SRCNN. The classification results are recorded in Figure 12.
From Figure 12, it is intuitive to find that the introduction of image preprocessing operations improves the classification accuracy of low-quality images. The contribution rate of the reconstruction-based preprocessing method to the improvement of classification network accuracy is generally higher than that of the image enhancement method, especially the method proposed in this paper, which has the best effect. At the same time, it was found that when upscale is set to 3, the classification accuracy of the proposed method can be significantly different from that of other methods, so upscale should be set to 3 in subsequent research.
In addition, the classification improvement rate of the model in this article for outdoor scenes is higher than that for indoor scenes. We speculate that this is because outdoor scenes are often affected by weather conditions and lighting conditions, resulting in poor image quality. The pre-reconstruction step of the model in this article helps to improve classification performance. Inspired by this, we ask whether the model can achieve improvement in classification performance for low-quality images taken in real-life scenarios.

4.3. Application Test of the Model for Low-Quality Complex Scene Classification

4.3.1. Exploration of the Types of Low-Quality Images Suitable for the Proposed Model

We conducted experiments for seven news scenes divided into two categories of better and worse according to their semantic clarity, using or not using pre-reconstruction as a comparison, and recorded the results in Table 7.
From Table 7, we can see that the semantic classification accuracy of the better scene was improved by 1.53%, and the semantic classification accuracy of the poor scene was improved by 2.11%. It is obvious that the improvement in low-quality images is greater, especially for the “fire” and “building” semantic categories, which were improved by 3.82% and 4.53%, respectively. The advantage of semantic classification for the “stage” semantic category in indoor images is higher than that for the “Conference” semantic category with good lighting. This shows that introducing preprocessing operations is more effective for image classification in scenarios affected by lighting and environment. Therefore, in the next part of the experiment, we will focus on studying the value of the proposed model in classifying low-quality images collected under weather and lighting conditions.

4.3.2. Exploration of the Application Value of the Proposed Model in Improving Classification Performance

This part of the experiment is conducted for the special environment self-built low-quality image dataset (the composition is shown in Table 8). IC Image Sets represent the infrared image set built by the professional infrared camera after being affected by illumination change, with a total of 1087 pictures. WD Image Sets represent the image sets created by cell phones after being affected by weather disturbances (rain, snow, haze), totaling 1174 images.
Firstly, the visual effect of using different algorithms in the pre-reconstruction stage of the model is recorded in Figure 13 and Figure 14, and it is clearly found that the reconstructed network pattern in this paper can better emphasize the detailed features. In particular, the semantic reconstruction of “Building” has obvious effects. In the local zoomed-in image, it is found that the window outline and the building structure have been repaired, which is more helpful for the subsequent categorization. The “Woodland” semantic shows better performance in IC Image Sets, and the edges of leaves are sharpened to a higher degree. The “Traffic” semantics show superior performance in WD Image Sets, with wheels and street lights being reconstructed to be more readable.
Secondly, in Figure 15, the experimental results using the two-stage classification model proposed in this paper versus the classification-only model are documented. The specific improvement is recorded in Table 7.
From Figure 15, it is found that the model introducing the reconstruction phase improves the correct rate of semantic classification for both datasets. The improvement of semantic classification accuracy for poor-quality scenarios is stronger than that for better-quality scenarios, especially for the discrimination of “Building”, which is consistent with the findings in Figure 13 and Figure 14, and shows that the reconstruction stage contributes greatly to the improvement of classification performance.
From Table 9, it is found that for better quality images, IC-Image Sets have better classification results with an average improvement of 0.67%. For poor-quality images, WD-Image Sets have better classification performance with an average improvement of 1.42%. In general, the semantic improvement of poor-quality scenes is higher: for the “Building” scene, the classification accuracy of IC-Image Sets and WD-Image Sets is improved by 2.08% and 3.52%, respectively. For the “Traffic” scenario, WD-Image Sets are more suitable, with an improvement of 0.67%. For the “Woodland” scenario, IC-Image Sets are more effective, with an improvement of 0.92%.

5. Discussion

The comprehensive experiments conducted in this study demonstrate that the proposed two-stage classification model effectively enhances the classification performance for the measured low-quality scene images.
Firstly, we focused on the feasibility of the classification stage model. Consistent with the focus of Section 2.1, we selected three general scene datasets, including indoor and outdoor scenes, based on the complexity of scene structures. We not only demonstrated the architectural stability through parameter selection but also proved its feasibility through comparisons with existing methods. The experimental results reveal that the proposed architecture excels in datasets with fewer classification labels and a higher proportion of outdoor scenes. Additionally, it exhibits higher accuracy in classifying images with higher resolutions. This observation prompts us to consider the contribution of improving the resolution of low-quality images to enhance the overall classification accuracy of the model.
Secondly, we investigated the rationality of the pre-reconstruction stage model. Aligning with the focus of Section 2.2, we established seven news scene image sets based on the causes of low image quality. By integrating diverse image quality enhancement techniques into the unified classification architecture outlined in this paper, we aim to underscore the superiority of the proposed reconstruction model. The results indicate that the proposed architecture achieves better classification accuracy for low-quality scene images affected by environmental and lighting factors. Moreover, the improvement is more significant for images with lower PSNR values. This finding suggests that we should explore the application potential of the proposed model by testing it on practical low-resolution images.
Finally, we tested the practicality of the two-stage classification model. Using equipment to capture and establish IC and WD Image Sets, we conducted comparative experiments between the two-stage and single-stage methods for low-resolution images, employing fewer labels based on the aforementioned results. Both visually and quantitatively, the experimental data clearly demonstrates the practical value of the proposed model.
Overall, the key advantage of the method presented in this paper lies in its innovative integration with practical low-resolution scene images to conceive an optimal classification model. A series of validation experiments fully demonstrate its application capabilities, complementing the test results of practical low-quality image research datasets. This undoubtedly provides a foundation for the analysis of recorded videos containing a large number of image frames. For instance, in emergency rescue missions, UAVs often capture live videos in low-light or harsh environments, where a classification model is urgently needed to assist commanders in quickly identifying critical lifeforms from complex images.
However, it is also recognized that the proposed method has limitations, primarily due to its heavy reliance on semantic elements in input images. If the effective or target information in an image is too small while the interfering information is excessive, it can lead to misclassification and semantic confusion. To address this issue, future research should focus on introducing practical dynamic video target detection, considering the target information in consecutive keyframes for comprehensive judgment. Additionally, optimizing the two-stage classification effect by combining the advantages of the SAM architecture for scenes with distinct edge delineation and achieving automated segmentation are also key areas for further work.

6. Conclusions

This paper focuses on the problem of semantic image classification of low-quality complex scenes, and proposes the MPSO two-stage classification optimization model, which is found to have the following advantages through three sets of experiments: 1. The classification effect of the scenes with a few kinds of complex semantics is better than the existing methods. 2. The optimized network design of the pre-reconstruction module greatly improves the classification effect of low-quality scene images. 3. The optimized network design of the pre-reconstruction module still maintains good application potential even when classifying outdoor scenes with large environmental influences. However, there still exists a scene with a lot of useless interfering information and low utilization of useful information. The next step is to focus on how to remove a large amount of interfering information in the pre-reconstruction module to help realize efficient recognition.

Author Contributions

Conceptualization, K.L.; Methodology, K.L.; Software, K.L. and Q.Z.; Validation, X.S. and X.D.; Formal analysis, R.W. and Q.Z.; Investigation, X.D.; Resources, R.W.; Data curation, R.W., X.D. and Q.Z.; Writing—original draft, K.L.; Writing—review & editing, X.S.; Project administration, K.L.; Funding acquisition, K.L. and X.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [National Natural Science Foundation of China] grant number [62006246] and [Xi’an Association for Science and Technology Youth Support Program] grant number [959202413100].

Data Availability Statement

All relevant data are within the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Deng, S.; Qiu, J. A Study on Image Posting Behaviors on Social Media Platforms. Book Inf. Sci. Knowl. 2024, 40, 10–20. [Google Scholar]
  2. Khan, A.; Chefranov, A.; Demirel, H. Image scene geometry recognition using low-level features fusion at multi-layer deep CNN. Neurocomputing 2021, 440, 111–126. [Google Scholar] [CrossRef]
  3. Kavitha, D.; Radha, V. TEXNET: A deep convolutional neural network model to recognize text in natural scene images. J. Eng. Sci. Technol. 2021, 16, 1782–1799. [Google Scholar]
  4. Kexin, L.; Siyuan, L. Discussion on the degree of influence on the performance of scene semantic classification after applying SRCNN reconstruction model. In Proceedings of the International Conference on Network Communication and Information Security (ICNIS 2021), Sanya, China, 3–5 December 2021; SPIE: Bellingham, WA, USA, 2022; Volume 12175, pp. 101–109. [Google Scholar]
  5. Dodge, S.; Karam, L. Understanding how image quality affects deep neural networks. In Proceedings of the 2016 Eighth International Conference on Quality of Multimedia Experience (QoMEX), Lisbon, Portugal, 6–8 June 2016; pp. 1–6. [Google Scholar]
  6. Ye, A.; Zhou, X.; Miao, F. Innovative Hyperspectral Image Classification Approach Using Optimized CNN and ELM. Electronics 2022, 11, 775. [Google Scholar] [CrossRef]
  7. da Costa, G.B.P.; Contato, W.A.; Nazare, T.S.; Neto, J.E.; Ponti, M. An empirical study on the effects of different types of noise in image classification tasks. arXiv 2016, arXiv:1609.02781. [Google Scholar] [CrossRef]
  8. Fc, A.; Jin, Y. DRSNet: Novel architecture for small patch and low-resolution remote sensing image scene classification. Int. J. Appl. Earth Obs. Geoinf. 2021, 104, 102577. [Google Scholar]
  9. Hendrycks, D.; Dietterich, T. Benchmarking Neural Network Robustness to Common Corruptions and Perturbations. arXiv 2019, arXiv:1903.12261. [Google Scholar] [CrossRef]
  10. Jian, Z. A Target Classification Method for Complex Scenarios. Ph.D. Thesis, Hangzhou University of Electronic Science and Technology, Hangzhou, China, 2020. [Google Scholar]
  11. Duan, M.; Li, K.; Yang, C.; Li, K. A hybrid deep learning CNN–ELM for age and gender classification. Neurocomputing 2017, 275, 448–461. [Google Scholar] [CrossRef]
  12. Riza, N.A. Low Image Contrast Detection in a Bright Light Interference HDR Scene Using Smart CAOS Camera. IEEE Photon-Technol. Lett. 2023, 35, 321–324. [Google Scholar] [CrossRef]
  13. Wang, Y.; Cao, Y.; Zha, Z.J.; Zhang, J.; Xiong, Z. Deep Degradation Prior for Low-Quality Image Classification. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  14. Chen, J.; Li, S.; Lin, L.; Wan, J.; Li, Z. No-reference blurred image quality assessment method based on structure of structure features. Signal Process. Image Commun. 2023, 118, 117008. [Google Scholar] [CrossRef]
  15. Oliva, A.; Torralba, A. Modeling the shape of the scene: A holistic representation of the spatial envelope. Int. J. Comput. Vis. 2001, 42, 145–175. [Google Scholar] [CrossRef]
  16. Zou, J.; Li, W.; Chen, C.; Du, Q. Scene classification using local and global features with collaborative representation fusion. Inf. Sci. 2016, 348, 209–226. [Google Scholar] [CrossRef]
  17. Khan, A.; Chefranov, A.; Demirel, H. Image-level structure recognition using image features, templates, and ensemble of classifiers. Symmetry 2020, 12, 1072. [Google Scholar] [CrossRef]
  18. Janowski, Ł.; Pydyn, A.; Popek, M.; Gajewski, J.; Gmińska-Nowak, B. Towards better differentiation of archaeological objects based on geomorphometric features of a digital elevation model, the case of the Old Oder Canal. Archaeol. Prospect. 2024, 1–12. [Google Scholar] [CrossRef]
  19. Zhu, Z.; Li, D.; Hu, Y.; Li, J.; Liu, D.; Li, J. Indoor scene segmentation algorithm based on full convolutional neural network. Neural Comput. Appl. 2021, 33, 8261–8273. [Google Scholar] [CrossRef]
  20. Li, Y.; Zhu, C.; Liu, Y.; Hong, Y.; Wang, J. Geometric and Semantic Analysis of Road Image Sequences for Traffic Scene Construction. Neurocomputing 2021, 465, 336–349. [Google Scholar] [CrossRef]
  21. Ou, X.; Wang, H.; Li, W.; Zhang, G.; Chen, S. A scene segmentation algorithm combining the body and the edge of the object. Inf. Process. Manag. 2022, 59, 102840. [Google Scholar] [CrossRef]
  22. Liu, S.; Tian, G.; Xu, Y. A novel scene classification model combining ResNet based transfer learning and data augmentation with a filter. Neurocomputing 2019, 338, 191–206. [Google Scholar] [CrossRef]
  23. Tang, P.; Wang, H.; Kwong, S. G-MS2F: GoogLeNet based multi-stage feature fusion of deep CNN for scene recognition. Neurocomputing 2017, 225, 188–197. [Google Scholar]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  25. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar] [CrossRef]
  26. Lu, K.; Cheng, J.; Li, H.; Ouyang, T. MFAFNet: A Lightweight and Efficient Network with Multi-Level Feature Adaptive Fusion for Real-Time Semantic Segmentation. Sensors 2023, 23, 6382. [Google Scholar] [CrossRef]
  27. Shetty, V.; Singh, M.; Salunkhe, S.; Rathod, N. Comparative Analysis of Diferent Classifcation Techniques. SN Comput. Sci. 2022, 3, 50. [Google Scholar] [CrossRef]
  28. Singh, P.; Chaudhury, S.; Panigrahi, B.K. Hybrid MPSO-CNN: Multi-level Particle Swarm optimized Hyperparameters of Convolutional Neural Network. Swarm Evol. Comput. 2021, 63, 100863. [Google Scholar] [CrossRef]
  29. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.-Y.; et al. Segment Anything. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 3992–4003. [Google Scholar]
  30. Zhang, Y.; Shen, Z.; Jiao, R. Segment anything model for medical image segmentation: Current applications and future directions. Comput. Biol. Med. 2024, 171, 108238. [Google Scholar] [CrossRef]
  31. Agrafiotis, P.; Skarlatos, D.; Georgopoulos, A.; Karantzalos, K. DepthLearn: Learning to Correct the Refraction on Point Clouds Derived from Aerial Imagery for Accurate Dense Shallow Water Bathymetry Based on SVMs-Fusion with LiDAR Point Clouds. Remote Sens. 2019, 11, 2225. [Google Scholar] [CrossRef]
  32. Liu, D.; Wen, B.; Liu, X.; Wang, Z.; Huang, T. When image denoising meets high-level vision tasks: A deep learning approach. In Proceedings of the International Joint Conference on Artificial Intelligence. International Joint Conferences on Artificial Intelligence Organization. arXiv 2018, arXiv:1706.04284. [Google Scholar] [CrossRef]
  33. Bousmalis, K.; Silberman, N.; Dohan, D.; Erhan, D.; Krishnan, D. Unsupervised Pixel-Level Domain Adaptation with Generative Adversarial Networks. In Proceedings of the Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar] [CrossRef]
  34. Tan, W.; Yan, B.; Bare, B. Feature super-resolution: Make machine see more clearly. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  35. Zhang, Z.; Zheng, W.; Ma, Z.; Yin, L.; Xie, M.; Wu, Y. Infrared star image denoising using regions with deep reinforcement learning. Infrared Phys. Technol. 2021, 117, 103819. [Google Scholar] [CrossRef]
  36. Zhao, J.; Cheng, R. Blind infrared images reconstruction using covariogram regularization from a regular pentagon. Infrared Phys. Technol. 2022, 120, 103963. [Google Scholar] [CrossRef]
  37. Cheng, Y.; Wang, D.; Zhou, P.; Zhang, T. Model Compression and Acceleration for Deep Neural Networks: The Principles, Progress, and Challenges. IEEE Signal Process. Mag. 2018, 35, 126–136. [Google Scholar] [CrossRef]
  38. An, Z.; Zhang, J.; Sheng, Z.; Er, X.; Lv, J. RBDN: Residual Bottleneck Dense Network for Image Super-Resolution. IEEE Access 2021, 9, 103440–103451. [Google Scholar] [CrossRef]
  39. Wang, W.; Ma, J.; Xu, C.; Zhang, Y.; Ding, Y.; Yu, S.; Zhang, Y.; Liu, Y. Low-rank sparse feature selection for image classification. Expert Syst. Appl. 2022, 189, 115685. [Google Scholar] [CrossRef]
  40. Wang, Y.; Zhang, H.; Zhang, G. cPSO-CNN: An efficient PSO-based algorithm for fine-tuning hyper-parameters of convolutional neural networks. Swarm Evol. Comput. 2019, 49, 114–123. [Google Scholar] [CrossRef]
Figure 1. Semantic classification performance for images with different scene complexity.
Figure 1. Semantic classification performance for images with different scene complexity.
Sensors 24 03983 g001
Figure 2. Visual comparison of the same scene taken by different shooting devices.
Figure 2. Visual comparison of the same scene taken by different shooting devices.
Sensors 24 03983 g002
Figure 3. Flowchart of the algorithm.
Figure 3. Flowchart of the algorithm.
Sensors 24 03983 g003
Figure 4. SRCNN architecture.
Figure 4. SRCNN architecture.
Sensors 24 03983 g004
Figure 5. DCNN-ELM initialization model.
Figure 5. DCNN-ELM initialization model.
Sensors 24 03983 g005
Figure 6. Experimental module design.
Figure 6. Experimental module design.
Sensors 24 03983 g006
Figure 7. Effect of batch-size selection on network optimal fitness (classification accuracy).
Figure 7. Effect of batch-size selection on network optimal fitness (classification accuracy).
Sensors 24 03983 g007
Figure 8. Scene Category and Quantity.
Figure 8. Scene Category and Quantity.
Sensors 24 03983 g008
Figure 9. Classification Obfuscation Matrix of 15-scene Dataset.
Figure 9. Classification Obfuscation Matrix of 15-scene Dataset.
Sensors 24 03983 g009
Figure 10. Comparison of image quality affected by inherent limitations of equipment.
Figure 10. Comparison of image quality affected by inherent limitations of equipment.
Sensors 24 03983 g010
Figure 11. The influence of image semantic quality on classification accuracy.
Figure 11. The influence of image semantic quality on classification accuracy.
Sensors 24 03983 g011
Figure 12. Classification of low-quality news images using different preprocessing methods based on the classification network proposed in this article.
Figure 12. Classification of low-quality news images using different preprocessing methods based on the classification network proposed in this article.
Sensors 24 03983 g012
Figure 13. Pre-reconstruction stage visualization of IC Image Sets.
Figure 13. Pre-reconstruction stage visualization of IC Image Sets.
Sensors 24 03983 g013
Figure 14. Pre-reconstruction stage visualization of WD Image Sets.
Figure 14. Pre-reconstruction stage visualization of WD Image Sets.
Sensors 24 03983 g014
Figure 15. Comparison of classification results with and without the introduction of a two-stage classification model.
Figure 15. Comparison of classification results with and without the introduction of a two-stage classification model.
Sensors 24 03983 g015
Table 1. SRCNN calculation flow.
Table 1. SRCNN calculation flow.
Calculation StepCalculation Method
1. Image block extraction and feature representation: a set of trained bases (e.g., Haar, DCT, etc.) are used to represent the input extracted image blocks, and a set of optimized filters associated with the bases are used to convolve the image. F 1 ( Y i ) = h ( W 1 Y i + b 1 ) = max ( 0 , W 1 Y i + b 1 )
2. Nonlinear mapping: the first layer is represented as an n1-dimensional vector and is transformed into an n2-dimensional vector by nonlinear mapping. F 2 ( Y i ) = max ( 0 , W 2 F 1 ( Y i ) + b 2 )
3. Image reconstruction: the high-resolution feature blocks extracted from the convolutional layer are aggregated, and then the filtering of the feature map is realized in the pattern of averaging. F ( Y i ) = W 3 F 2 ( Y i ) + b 3
Table 2. Calculation steps for feature extraction.
Table 2. Calculation steps for feature extraction.
Calculation StepCalculation MethodParameter Meaning
1. Convolutional layer calculation x i j k + 1 = R K l p = 1 P i q = 1 Q i x K l ( s i + p , s j + q ) k ω K l ( p , q ) k + 1 + b i j K l denotes the number of feature maps connected to the kth convolutional layer, i.e., the number of channels. P i and Q i are the currently set filter width and height, and S is the sliding step. ω K l ( p , q ) k + 1 denotes the value of the convolutional kernel connected to the k + 1st layer of feature maps located at (p, q). R is chosen to be the nonlinear excitation function ReLUs. b i j is the bias term.
2. Calculation of pooling layer X i j k + 1 = max X ( i , j ) k , X ( i + 1 , j + 1 ) k , , X ( i + P i , j + Q i ) k
Table 3. Listing of datasets.
Table 3. Listing of datasets.
DatasetsTotalCategoryNumber of Each CategoryData Content
15-Scene448515200~4004 indoor; 9 outdoor;
1 mixed indoor-outdoor
SUN397108,754397>100175 indoor; 220 outdoor;
2 mixed indoor-outdoor
MIT Inoor15,62067>10067 indoor; five categories: “Store”, “Home”, “Public spaces”, “Leisure”, and “Working place”.
Table 4. Comparison of scene classification effects of 21 methods.
Table 4. Comparison of scene classification effects of 21 methods.
Dataset15-SceneSUNMIT-Inoor
Optimization method based on feature extraction improvementSIFT-SPM81.4--
SIFT-GAIJIN---
PlaceNet-50.3668.24
MOP-CNN-51.98
CENTRIST + SVM--36.9
BOP--46.1
Optimization methods based on network architecture improvementImageNet-AlexNet84.0542.6156.79
ImageNet-GoogleNet84.9543.8859.48
ImageNet-VGG86.2848.2964.87
VGG-16-56.275.67
Places365-AlexNet89.2556.1270.72
Places365-GoogleNet91.2558.3773.30
Places365-VGG91.9763.2476.53
Optimization methods based on classifier replacementGoogleNet-SVM92.09--
GoogleNet-ELM92.18--
ResNet-SVM93.4--
ResNet-ELM92.64--
CNNang-SVM-6958.4
CNN-SVM-55.14-
DCNN-ELM--69.27
Associative convolutional architectureSAM--70.26
The methodology proposed in this paper94.1868.976.71
Table 5. Comparison of classification performance for multi-class scene images of different quality (%).
Table 5. Comparison of classification performance for multi-class scene images of different quality (%).
Original ImageUpscale = 2Upscale = 3Data Content
Indoor (4 types)89.7889.6389.3888.97
outdoor (9 types)96.1095.7595.3294.74
mixed indoor-outdoor (2 types)91.7291.4291.1690.69
All 15 types94.1893.8893.5192.98
Table 6. Composition of Low-Quality News Image Scene Set.
Table 6. Composition of Low-Quality News Image Scene Set.
SemanticsStageConferenceWoodlandFireBuildingFlood
Number725698640720703659716
Better350334320324346315355
Poor375364320396357344361
Number725698640720703659716
Table 7. Classification of News Scene Images with Different Semantic Clarity (%).
Table 7. Classification of News Scene Images with Different Semantic Clarity (%).
SemanticsStageConferenceWoodlandFireBuildingFloodHighwayAvg
betterWithout pretreatment-Net90.7086.1490.1393.2290.4688.6793.5090.44
MPSO SRCNN-Net91.1286.8990.8495.7994.7389.2594.8291.97
poorWithout pretreatment-Net90.2486.0387.4988.2685.5986.9992.1988.15
MPSO SRCNN-Net90.9286.8090.9592.0890.1288.0392.7490.26
Table 8. Self-constructed low-quality datasets with different semantic clarity.
Table 8. Self-constructed low-quality datasets with different semantic clarity.
SemanticsIC Image SetsWD Image Sets
SumBetterPoorSumBetterPoor
Woodland271135136263130133
Traffic 366175191351162189
Building450200250560250310
Table 9. Specific enhancements (%).
Table 9. Specific enhancements (%).
“Building”“Traffic”“Woodland”Avg
IC-Image SetsBetter quality1.15%0.04%0.76%0.67%
Poor quality2.08%0.49%0.92%1.28%
WD-Image SetsBetter quality0.02%0.14%0.65%0.35%
Poor quality3.52%0.67%0.41%1.42%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, K.; Wang, R.; Song, X.; Deng, X.; Zhu, Q. Exploration of MPSO-Two-Stage Classification Optimization Model for Scene Images with Low Quality and Complex Semantics. Sensors 2024, 24, 3983. https://doi.org/10.3390/s24123983

AMA Style

Liu K, Wang R, Song X, Deng X, Zhu Q. Exploration of MPSO-Two-Stage Classification Optimization Model for Scene Images with Low Quality and Complex Semantics. Sensors. 2024; 24(12):3983. https://doi.org/10.3390/s24123983

Chicago/Turabian Style

Liu, Kexin, Rong Wang, Xiaoou Song, Xiaobing Deng, and Qingchao Zhu. 2024. "Exploration of MPSO-Two-Stage Classification Optimization Model for Scene Images with Low Quality and Complex Semantics" Sensors 24, no. 12: 3983. https://doi.org/10.3390/s24123983

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop