Next Article in Journal
Predicting the Global Potential Suitable Distribution of Fall Armyworm and Its Host Plants Based on Machine Learning Models
Next Article in Special Issue
Deep Learning-Based Detection of Oil Spills in Pakistan’s Exclusive Economic Zone from January 2017 to December 2023
Previous Article in Journal
Improving Hyperspectral Image Classification with Compact Multi-Branch Deep Learning
Previous Article in Special Issue
Radiation Feature Fusion Dual-Attention Cloud Segmentation Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Methodology for Severe Convective Cloud Identification Using Lightweight Neural Network Model Ensembling

College of Meteorology and Oceanography, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(12), 2070; https://doi.org/10.3390/rs16122070
Submission received: 21 April 2024 / Revised: 18 May 2024 / Accepted: 5 June 2024 / Published: 7 June 2024
(This article belongs to the Special Issue Deep Learning for Satellite Image Segmentation)

Abstract

:
This study introduces an advanced ensemble methodology employing lightweight neural network models for identifying severe convective clouds from FY-4B geostationary meteorological satellite imagery. We have constructed a FY-4B based severe convective cloud dataset by a combination of algorithms and expert judgment. Through the ablation study of a model ensembling combination of multiple specialized lightweight architectures—ENet, ESPNet, Fast-SCNN, ICNet, and MobileNetV2—the optimal EFNet (ENet- and Fast-SCNN-based network) not only achieves real-time processing capabilities but also ensures high accuracy in severe weather detection. EFNet consistently outperformed traditional, heavier models across several key performance indicators: achieving an accuracy of 0.9941, precision of 0.9391, recall of 0.9201, F1 score of 0.9295, and computing time of 18.65 s over the test dataset of 300 images (~0.06 s per 512 × 512 pic). ENet shows high precision but misses subtle clouds, while Fast-SCNN has high sensitivity but lower precision, leading to misclassifications. EFNet’s ensemble approach balances these traits, enhancing overall predictive accuracy. The ensemble method of lightweight models effectively aggregates the diverse strengths of the individual models, optimizing both speed and predictive performance.

1. Introduction

Severe convective clouds, typically associated with thunderstorms, high winds, hail, and other extreme weather phenomena, are characterized by robust convective clusters formed under specific humidity conditions in an unstable atmospheric layer [1]. These clouds, featuring a relatively low cloud base and a significantly higher cloud top, contain intense internal vertical airflows [2,3,4,5] and are crucial for meteorological forecasting and disaster warning systems, potentially mitigating the impacts of natural disasters [6,7,8]. With advances in remote sensing technology [2,9,10,11,12,13], the automatic detection and classification of these clouds using satellite images has become feasible, enhancing the timeliness and accuracy of meteorological services. Simultaneously, the use of deep learning techniques, particularly convolutional neural networks [14,15,16] (CNNs), has greatly improved the detection and classification of severe convective clouds in satellite imagery. Cloud segmentation based on deep learning is a long-standing and ongoing area of research in the field of remote sensing. This includes general cloud segmentation without differentiating cloud types [5,11,14,17,18,19,20,21,22], as well as segmentation of severe convective clouds [2,4,6,8,15,23,24,25,26,27], which continue to be actively studied, underscoring the enduring significance of this problem. In practical applications, particularly where real-time processing of vast datasets is required, lightweight neural networks have garnered attention due to their lower computational demands and rapid processing capabilities. Lightweight models such as ENet [28], ESPNet [29], Fast-SCNN [30], ICNet [31], and MobileNetV2 [32], though smaller, demonstrate performance comparable to larger networks like U-Net [33] and DeepLabV3 [34] in semantic segmentation tasks. These networks maintain high accuracy while significantly enhancing computational speed, making them suitable for use in resource-constrained environments [5,16,35].
Photogrammetry [36,37,38] and GNSS radio occultation [39,40] are exemplary in providing precise object details and valuable profile data, respectively, but satellite remote sensing images offer distinct advantages for large-scale environmental monitoring. Satellite imagery excels in providing extensive geographical coverage and frequent updates [41], capabilities that are indispensable for tracking dynamically changing severe convective clouds. This comprehensive and continuous coverage allows for the real-time observation of weather patterns on a global scale, which is not feasible with photogrammetry’s relatively limited area coverage or the atmospheric profiles provided by GNSS radio occultation. Additionally, the integration of multispectral imaging in satellites enhances the detection and analysis of various cloud properties and atmospheric conditions, leading to more accurate weather forecasting and climate studies. Therefore, while other methods provide valuable insights at different scales and dimensions, satellite remote sensing remains crucial for holistic and continuous monitoring.
To further improve recognition accuracy and the generalizability of models, ensemble techniques have been incorporated into deep learning. By aggregating predictions from multiple models, ensemble methods can effectively reduce the error rates that might occur in individual models, thereby achieving higher accuracy [3,42,43]. Additionally, ensemble methods enhance the predictive capability of models on unseen samples, which is particularly vital under the fluctuating conditions of meteorological phenomena [24,25,44]. As lightweight neural network technologies [45] continue to advance, their integration into ensemble frameworks for recognizing severe convective clouds offers significant research value and practical prospects. Integrating multiple lightweight models not only balances efficiency and accuracy but also enhances the robustness of the system [35].
For light, medium, or large neural networks, researchers have made many meaningful and significant improvements to make them more adaptable to cloud recognition tasks [20,24,42,46,47,48,49,50] (including the identification of severe convective clouds). However, beyond these improvements, we can also consider how to use model ensemble techniques to better integrate and utilize existing lightweight neural networks, enabling them to exhibit capabilities and computational efficiency far beyond a single large neural network.
The model ensemble techniques have been widely used in many fields [3,7,11,19,51,52,53], but they have not yet been applied to the task of remote sensing image recognition of severe convective clouds. In the field of meteorology, model integration techniques have been extensively utilized for analyzing synoptic scale weather phenomena [54] and enhancing meteorological forecasts [55,56]. This approach has also been widely adopted in the domain of cloud segmentation within satellite remote sensing imagery [3,11,19]. However, these efforts predominantly employ integrated models for RGB channel segmentation in visible light satellite images, without differentiation of cloud types, and do not engage the infrared and water vapor channels necessary for segmenting severe convective clouds. Currently, the singular application of model integration techniques to severe convective weather focuses on extreme weather events, including the use of radar reflectivity data for thunderstorm warnings [57] and high-resolution rapid refresh (HRRR) 1–24 h forecasts to predict catastrophic weather conditions over the United States [58]. Additionally, existing cloud segmentation generally involves the integration of lightweight and non-lightweight networks to achieve higher accuracy, albeit at a marginally slower speed than non-lightweight networks [3], or solely through non-lightweight network integrations [11,19], yet these methods typically underemphasize computational speed. Therefore, we innovatively propose a multi-lightweight model integration technique for severe convective cloud segmentation, which not only surpasses the accuracy of single non-lightweight networks but also improves speed. The novelty of this research lies in the successful first-time application of this multi-lightweight model integration approach specifically for severe convective cloud segmentation.
Given the ample excellent research in the field [2,4,6,7,8,15,23,24,25,26,27,59,60], we do not propose a specific neural network. While developing new networks has its merits, optimizing the use of existing networks is equally valuable [61]. Our work achieves efficient, high-precision cloud detection by enhancing the utilization of existing lightweight networks. We demonstrate the efficacy of integrating multiple lightweight neural network models, a more impactful approach likely to see broader adoption among researchers than introducing a single network. This study employs five lightweight neural networks—ENet, ESPNet, Fast-SCNN, ICNet, MobileNetV2—as submodules in an ensemble to achieve efficient and accurate detection of severe convective clouds. The FY-4B satellite [62,63] is the first operational satellite of China’s new generation of geostationary meteorological satellites, the Fengyun-4 series, with significant improvements compared to FY-4A [64,65]. Therefore, this article uses data from FY-4B and takes the severe convective weather in the southeast of the Eurasian continent and nearby sea areas as the target of research. Comparative experiments demonstrate that this ensemble method not only maintains high recognition accuracy but also significantly improves processing speed, showcasing the potential and advantages of lightweight deep learning models in practical applications.

2. Study Area and Data

2.1. Study Area

The study area is located between 0–45°N and 100–160°E (Figure 1), primarily encompassing the southeastern part of the Eurasian continent and adjacent maritime regions. This region exhibits significant seasonal and regional characteristics in severe convective weather activities, influenced by both topography and regional atmospheric circulation patterns. The area includes a wide range of terrestrial and marine zones, such as China, Japan, the Korean Peninsula, the South China Sea, and the East China Sea. Each area’s unique weather patterns are driven by a combination of geographical, oceanographic, and atmospheric factors [66,67], ranging from intense thunderstorms in China’s southeast, influenced by moist and cool air masses, to Japan’s typhoon-induced convective systems during summer. The Korean Peninsula shows sharp seasonal contrasts in convective activity, while the South and East China Seas are hotspots for cyclonic developments and associated severe weather, respectively, fueled by warm sea temperatures and monsoonal interactions.

2.2. FY-4B Data

Fengyun-4B [62,63] (FY-4B) is a new-generation geostationary meteorological satellite in the Fengyun series from China, an important member following FY-4A. It is specifically designed for meteorological observation, climate environment monitoring, and natural disaster warning. Compared to FY-4A, FY-4B has improved performance [62,68,69], providing more accurate and comprehensive meteorological data to meet the growing demand for meteorological services. As of 1 April 2024, the satellite’s subsatellite point is located at 105°E.
The Advanced Geosynchronous Radiation Imager (AGRI) is one of the core payloads of the FY-4B satellite. It is a high-performance multi-channel imager primarily used to obtain high-quality atmospheric, cloud, and surface images. The AGRI sensor inherits the technical advantages of similar sensors on FY-4A, while optimizing and improving some performance indicators. The successful deployment of FY-4B and its AGRI sensor has further enhanced the capabilities in the field of geostationary meteorological observation, which is of great significance for improving the accuracy of weather forecasts, deepening the understanding of the Earth’s climate system, and effectively responding to natural disasters. Through coordination with other meteorological satellite systems, FY-4B can provide more robust and comprehensive support for global meteorological services. The main features and capabilities of the FY-4B AGRI sensor include:
  • Multi-channel imaging. AGRI is equipped with multiple observation channels, including visible light, near-infrared, mid-infrared, and far-infrared bands, supporting wide-band atmospheric, cloud, and surface observations. This multi-channel observation capability allows AGRI to capture detailed features under different meteorological and environmental conditions.
  • High spatial resolution. AGRI provides a spatial resolution of up to 500 m in the visible light and near-infrared channels, and even higher resolution in other channels, enabling it to capture more detailed meteorological and surface information.
  • Fast scanning capability. AGRI provides rapid observation updates, including global scans every 15 min, regional scans every 5 min, and rapid scans of key areas every minute, greatly improving the monitoring and response speed to extreme weather events.
The observational parameters for each channel are presented in Table 1.

2.3. Data Preprocessing

The radiometric calibration and geographical repositioning of FY-4B data are crucial steps before FY-4B data labeling and severe convective cloud dataset construction. Radiometric calibration ensures that the satellite’s sensor outputs accurately reflect the electromagnetic energy it captures. This process corrects any system biases or errors, translating raw data into meaningful measurements that can be compared over time and with other sensors. Geographical repositioning, on the other hand, aligns the satellite data with geographic coordinates on Earth’s surface. Here, the geographic repositioning includes converting nominal projection to a 0.04° lat-lon grid projection, and converting the 0.04° lat-lon grid projection back to nominal projection. This is essential for accurate mapping and analysis, as it corrects any positional discrepancies due to the satellite’s orbit, sensor geometry, or Earth’s rotation. Together, these processes ensure that the data from FY-4B is reliable and precise for practical applications.
Radiometric calibration [70] mainly includes two steps: Firstly, Level 1 data for the solar reflective bands requires the conversion of DN values to reflectance, radiance, or apparent reflectance. Secondly, for the infrared bands, Level 1 data requires the conversion of DN values to radiance ( W / ( m 2 · s r · u m ) ) or brightness temperature (K) according to the lookup table.
In geographical repositioning [8,71], we need to convert nominal projection to a 0.04° lat-lon grid projection. The FY-4 satellite employs the geostationary orbit nominal projection defined by the CGMS LRIT/HRIT global standard, with geographic coordinates calculated based on the WGS84 reference ellipsoid. Projection transformation involves mapping the pre-projection data points according to the projection transformation formula into the coordinates of the post-projection image. We first determine the geographic extent of the projection transformation and calculate the number of rows and columns in the post-projection image. Then, we map each pixel from the original cloud map through the lat-lon projection transformation formula to the post-projection image. When converting the FY-4B data from 0.04° lat-lon grid projection to nominal projection, the opposite operation can be performed.

2.4. FY-4B Data Labeling: Severe Convective Cloud Dataset Construction

The geographical regions from which the dataset samples were collected are shown in Figure 1, spanning the period from 1 June 2022, to 30 June 2023. All samples were selected from areas with distinct convective cloud clusters, ensuring that each sample contained convective clouds. Regarding the number of samples used for the cloud segmentation neural network, Tian et al. [3] employed 22,432 samples of 256 × 256 pixels, while Ma et al. used 7200 samples of the same size [19]. Zhang et al. [72] utilized 784 samples of 125 × 125 pixels and 2543 samples of 256 × 256 pixels. Li et al. [73] used 3000 samples of 512 × 512 pixels for single cloud category classification in lightweight network model. As this study focuses on ensemble training based on lightweight models, the demand of lightweight models for sample size is relatively small. Therefore, a compromise of 3000 samples of 512 × 512 pixels was adopted.
We have constructed a severe convective cloud dataset using 3000 multi-channel satellite data samples from the AGRI onboard the FY-4B satellite, each with dimensions of 512 × 512 pixels. This dataset was developed through a combination of algorithms and expert judgment. The labeling process focuses primarily on images from the FY-4B satellite, based on actual needs. In the labeling process, severe convective cloud formations such as convective cells, thunderstorm clusters, and squall lines are identified. The process eliminates interference from cirrus clouds in the longwave infrared data.
Before labeling, data from reflective and infrared bands are preprocessed—reflective band data are calibrated as reflectance, and infrared band data as brightness temperature. Data from channels at 0.47, 0.65, and 0.825 μm are combined to create true-color images. For example, the water vapor channel data (WV at 6.25 μm) and longwave infrared channel data (LWIR at 10.8 μm) are separated to create a water vapor-infrared brightness temperature difference ( B T D ), calculated as B T D = T B B W V T B B L W I R .
After calculating the B T D , each remote sensing image is subjected to dynamic thresholding to coarsely extract mesoscale convective cloud clusters, identifying their approximate regions and applying morphological operations to fill gaps within the convective clusters. The dynamic B T D threshold of convective clouds in each period is specifically set based on the expert visual interpretation of the images during that time frame. Then, combined with the synthesized true-color cloud map, the expert manual visual interpretation corrects misidentifications of convective areas (typically caused by high-altitude clouds like cirrus), further enhancing the labeling accuracy.
The severe convective cloud dataset example is illustrated in Figure 2, with convective cloud clusters identified and edges detected using the Canny operator [74]. The Canny operator, a multi-stage edge detection algorithm, is widely employed in computer vision tasks for its optimal performance in identifying object boundaries within digital images [74,75]. This operator applies a sequence of Gaussian filters to smooth the image, reducing noise while preserving the integrity of edge structures [76]. The algorithm then computes the intensity gradients, followed by non-maximum suppression to eliminate spurious edge pixels [77]. Finally, hysteresis thresholding is applied to track and connect the remaining edge pixels, resulting in a binary edge map. The Canny operator’s adaptability to various image conditions and its ability to produce clear, continuous edges make it a preferred choice for edge detection in numerous applications like delineation of convective cloud clusters in meteorological image analysis [78].

3. Method

The framework for our method of severe convective cloud identification using lightweight neural network model ensembling is illustrated in Figure 3. Initially, from the aforementioned dataset of severe convective clouds, we extract three-channel information from 3000 samples of 512 × 512 resolution, specifically at 6.25 μm, B T D   ( 6.25 10.8   μ m ) , and B T D   ( 10.8 12   μ m ) . This forms a four-dimensional array (3000 × 3 × 512 × 512). Concurrently, the corresponding binary labels are extracted, resulting in another four-dimensional array (3000 × 1 × 512 × 512). Subsequently, these 3000 data samples are randomly divided into training, validation, and test sets with a ratio of 8:1:1. Following this partition, we employ model ensembling and ablation study techniques to test the performance of 31 combinations ( k = 1 5 C 5 , k ) of five different lightweight neural networks (ENet, ESPNet, Fast-SCNN, ICNet, and MobineNetV2) in identifying severe convective clouds to determine the optimal combination. Upon identifying the most effective ensemble, we build a framework for severe convective cloud identification based on this optimal lightweight neural network ensemble, and compare its efficacy with traditional, non-lightweight neural networks such as U-Net and DeepLabV3, thereby validating the effectiveness of our approach.
In the context of satellite remote sensing, infrared and water vapor channels are indispensable for the segmentation of severe convective clouds due to their capability to detect essential atmospheric conditions [4,8,25,26,27]. These channels excel in capturing thermal variations and monitoring moisture dynamics—key factors in identifying and predicting severe weather events. Infrared imagery, effective in delineating cold cloud tops indicative of severe conditions, and water vapor channels, crucial for assessing moisture content essential for storm intensity and development, operate continuously, thus facilitating round-the-clock monitoring and significantly enhancing forecasting accuracy. Conversely, visible light and shortwave infrared (SWIR) channels present limitations: they are bound to daylight operations and lack comprehensive monitoring of atmospheric water vapor, reducing their effectiveness in continuous severe weather prediction. Due to these constraints and the unavailability of nighttime data from FY4 series satellites in visible and SWIR channels [8,64,70], we have opted to utilize the water vapor and longwave infrared band data from the FY4B satellite. The selected 6.25 μm, B T D   ( 6.25 10.8   μ m ) , and B T D   ( 10.8 12   μ m ) have previously been employed in the FY4A series for the task of severe convective cloud segmentation, as demonstrated by Chen et al. [8]. Additionally, the selection of these three channels is justified as they sufficiently enable high-precision segmentation of severe convective clouds. The innovation of this paper is reflected in surpassing traditional networks in both speed and accuracy. Given that the three channels are adequate, introducing additional channels would result in marginal improvements and could potentially reduce the computational efficiency of the model.

3.1. Parameter Configuration

This experiment was conducted using the PyTorch2.2.0 + cu121 (Python 3.10.14) deep learning framework [79] on an NVIDIA RTX A5000 GPU equipped with 24 GB of memory. The adaptive moment estimation (Adam) optimizer [80] was employed, configured with the exponential decay rate for the first moment estimates β 1 = 0.9 , the exponential decay rate for the second-moment estimates β 2 = 0.999 , and a very small constant ϵ = 10 8 to prevent any division by zero in the implementation.

3.2. Loss Function

In the task of binary classification for severe convective cloud detection, the binary cross-entropy loss [81] (BCE) is employed to effectively measure model accuracy. Cross-entropy serves as a crucial metric in evaluating classification models by quantifying the divergence between predicted and true probability distributions. A cross-entropy score of zero represents an ideal where predictions perfectly match actual labels, whereas scores approaching infinity indicate severe mispredictions, with log loss amplifying errors where predicted probabilities of the correct class near zero. This measure not only prioritizes accuracy but also considers the probabilistic confidence of predictions, enhancing both the robustness and reliability of model evaluations. It is advantageous for gradient-friendly properties that facilitate effective model training via backpropagation. BCE quantifies the discrepancy between predicted probabilities and actual binary labels, defined mathematically as:
BCE = 1 N i = 1 N y i log y i ^ + 1 y i log 1 y i ^
where N is the sample count, y i represents the true label, and y i ^ denotes the predicted probability of severe convective cloud presence.

3.3. Model Evaluation Index

To objectively evaluate the performance of our models in the binary classification of severe convective clouds, we employ five standard metrics. Each metric provides insights into different aspects of model accuracy and robustness, thus allowing a comprehensive assessment across various dimensions of prediction performance.
  • Accuracy: Defined as the ratio of correctly predicted results to the total results, accuracy offers a straightforward measure of overall model effectiveness. Specifically, the accuracy measures the proportion of total predictions that are correct. It ranges from 0 to 1, where 1 indicates perfect accuracy, and 0 indicates complete inaccuracy.
Accuracy = Number   of   Correct   Predictions Total   Number   of   Predictions
2.
Precision: Precision assesses the accuracy of positive predictions and evaluates the incidence of false positives. It is particularly useful in situations where the cost of a false positive is high. Like accuracy, precision ranges from 0 to 1, with 1 being perfect (no false positives) and 0 indicating all positive predictions are incorrect.
Precision = True   Positives True   Positives + False   Positives
3.
Recall: Also known as sensitivity, recall measures the model’s ability to detect all actual positives. Its range is also between 0 and 1, where 1 means all true positives are correctly identified, and 0 signifies no true positives are detected.
Recall = True   Positives True   Positives + False   Negatives
4.
F1 Score: The F1 score is the harmonic mean of precision and recall, providing a balance between the two when an equal importance is assumed. It is particularly useful when the cost of false positives and false negatives is high. The F1 score also ranges from 0 to 1, where 1 is the best possible score, indicating perfect precision and recall.
F 1   Score = 2 × Precision × Recall Precision + Recall
5.
Intersection over Union (IoU): Also known as the Jaccard index, IoU is the ratio of the intersection to the union of the predicted and true labels. An IoU of 1 indicates a perfect prediction where the predicted labels or boundaries completely coincide with the ground truth, while an IoU of 0 signifies no overlap at all between the predicted and actual labels.
IoU = Area   of   Overlap Area   of   Union
6.
Overall Performance (OP): To synthesize the insights provided by individual metrics into a single performance indicator, we introduce the OP index. The higher the OP score, the better, as it suggests a model that performs well across all key aspects of binary classification. OP is computed as the sum of the standardized scores of the five aforementioned metrics:
OP = Accuracy + Precision + Recall + F 1   Score + IoU
This composite metric provides a holistic view of the model performance, encompassing accuracy, precision, sensitivity, and the balance between precision and recall, as well as the degree of overlap between predicted and true classes. OP is particularly useful when selecting the optimal combination from a large set of models. Here, “optimal” refers to the best overall performance, rather than necessarily being the top performer in every individual metric.

3.4. Training

The models were trained for a total of 30 epochs, the loss records and detailed parameters setting of which are shown in Figure 4 and Table 2, respectively.
For most networks, the training loss tends to stabilize after 20 epochs, and it generally remains stable after 30 epochs, with the exception of MobileNetV2, which exhibits slight fluctuations in the later stages (Figure 4). The fluctuations in validation loss are more pronounced, with ESPNet experiencing the largest variations in validation loss. Notably, MobileNetV2 shows significant fluctuations during the 5–10 epoch range, but eventually stabilizes. Importantly, the lightweight network ENet demonstrates the fastest decline in validation loss at the onset of training and achieves the lowest stable loss subsequently. The performance of non-lightweight networks, such as U-Net and DeepLabV3, in both training and validation is commendable as well.
However, when focusing on training duration and model size (as shown in Table 2), the advantages of lightweight models become apparent. Even when aggregating the training times of the five lightweight networks, the required training time, the number of model parameters, and the model size are all less than those of U-Net and DeepLabV3. Thus, the methodology proposed in this paper for severe convective cloud identification using an ensemble of lightweight neural network models aims to leverage this compactness of lightweight network architectures.
Our findings also indicate that networks with excessively small model parameters do not yield desirable results. As evidenced in Table 2, ICNet’s parameters are at least two orders of magnitude smaller than those of other networks. When a model’s parameter count is particularly low, its training loss tends to be significantly greater than that of more complex models with larger parameter counts. This discrepancy stems from the constrained expressive, learning, and optimization capabilities of models with minimal parameters, which tend to underfit, as demonstrated by the high loss observed in Figure 4.

3.5. Building the EF Network via Model Ensembling and Ablation Studies

In this section, we have achieved computational efficiency and accuracy that surpass those of single non-lightweight networks by ensembling various lightweight networks with the optimal combination. To ascertain the most effective model ensembling combination for severe convective cloud identification, we utilize ablation study techniques to evaluate the efficacy of 31 different combinations ( k = 1 5 C 5 , k ) involving five distinct lightweight neural networks (ENet, ESPNet, Fast-SCNN, ICNet, and MobileNetV2).
Figure 5 and Figure 6 illustrate the results of these 31 model combinations. During the evaluation of the 31 model combinations, we conducted separate computations and recorded the resulting parameters for each combination, executing a total of 31 runs. In the model ensembling process, we fuse the outputs of the last linear layer of each network within the combinations, which are the raw score logits not yet normalized by probabilities. Utilizing logits directly in the loss function enhances numerical stability as transforming logits to probabilities via softmax before computing the cross-entropy loss involves logarithmic and exponential operations, which can lead to numerical instability. This approach also improves computational efficiency by avoiding dual transformations (first softmax, then log). For each specific combination, all modules within the ensemble were simultaneously executed, and their logits were summed to obtain the final output probability. Based on this output, the accuracy and runtime of each combination were obtained. The process of incrementally adding or removing lightweight neural network modules to identify the critical components and optimal combination is referred to as an ablation study in this paper. In this context, the ablation study serves not only to assess performance but also as a method for determining the most favorable ensemble configuration.
Initially, we hypothesized that the full model combination of ENet + ESPNet + Fast-SCNN + ICNet + MobileNetV2 (E + ES + F + I + M) would yield the highest-quality results, albeit with slightly lower computational efficiency. However, upon comparing the performance parameters of these 31 model combinations (as shown in Table 3), we discovered that the dual-model combination of ENet + Fast-SCNN (E + F) actually provided the highest overall quality, ranking first in five metrics (test loss, accuracy, F1, IoU, and OP). Additionally, due to its simpler dual model structure, the E + F combination also offers advantages in computation speed. Consequently, we have chosen to construct EFNet (ENet and Fast-SCNN based network) based on the dual-model combination of E + F.

3.6. Parallel Computing

Given our approach uses an ensemble of multiple lightweight submodels, the availability of multiple GPU resources would enable the utilization of multi-GPU hardware acceleration to parallelize the inference process of these submodels. Consequently, we propose a model parallelization architecture that allows multiple submodels to be executed concurrently on different GPUs.
Specifically, we have defined a new model class named CombinedModel, which accepts an arbitrary number of submodels as input parameters. During the forward propagation process, each submodel independently processes input data on its respective GPU, generating intermediate results. Subsequently, through efficient cross-GPU communication primitives, the outputs from all submodels are aggregated and summed to produce the final prediction result. This fine-grained model parallelization method allows us to handle models without being constrained by the memory limits of a single GPU.
During the implementation phase, we leveraged high-level APIs to facilitate more efficient and streamlined training and evaluation of model combinations. The program organizes multiple submodels into a list, iterates over them, and automatically distributes input data across multiple GPUs for parallel computation while aggregating gradients from each GPU, thereby accelerating the training process. Simultaneously, during the inference stage, we employ a context manager to disable gradient calculations, as the inference process does not require backpropagation. This approach reduces unnecessary memory consumption and computational overhead, thereby enhancing inference efficiency. In summary, by utilizing the high-level APIs provided by PyTorch, we achieve more efficient and readable implementation of model combination training and evaluation while optimizing inference performance, laying the foundation for realizing high accuracy and low latency objectives.

4. Results

4.1. Qualitative Analysis

By comparing the results of severe convective cloud identification using EFNet, U-Net, DeepLabV3, ENet, ESPNet, Fast-SCNN, ICNet, and MobileNetV2 (as shown in Figure 7), it is evident that the identification quality of single lightweight neural networks is generally inferior to that of non-lightweight networks (U-Net, DeepLabV3).
We found significant misclassification issues with ICNet (Figure 7k), MobileNetV2 (Figure 7l), ENet (Figure 7h), and Fast-SCNN (Figure 7j). The common issue of ICNet, MobileNetV2, and ENet is the extensive underreporting of clouds (an excess of green areas), with ICNet and Fast-SCNN additionally suffering from a substantial number of false positives (an excess of red areas). However, the combined EFNet (Figure 7e) does not exhibit both issues and achieves significantly better performance compared to U-Net (Figure 7f), and shows mixed results when compared to DeepLabV3 (Figure 7g). When the quality of results is comparable and slightly superior, EFNet is preferred due to its significantly higher computational efficiency.
It is noteworthy that in this study, ENet (Figure 7h) exhibits a considerable number of false negatives (evidenced by the extensive green areas), yet it has fewer false positives. Conversely, Fast-SCNN (Figure 7j) shows a higher number of false positives (indicated by the extensive red areas), but fewer false negatives. The integration of these two models into EFNet (Figure 7e) results in a reduction in both false positives and false negatives. This outcome serves as the most intuitive demonstration of the significance of model integration in improving predictive accuracy and balance.

4.2. Performance Metrics Analysis

Although qualitative analysis suggests that DeepLabV3 appears to perform better than U-Net, in reality, the performance metrics of DeepLabV3 do not show any advantage over U-Net (as indicated in Table 4). Furthermore, compared to the non-lightweight networks U-Net and DeepLabV3, EFNet maintains an advantage in five key metrics (Test Loss, Accuracy, F1, IoU, OP) and also demonstrates greater computational speed.

5. Discussion

Severe convective cloud systems are a critical element in weather forecasting, involving complex physical and dynamical processes that make it challenging for a single model to accurately capture all relevant features. Therefore, the use of ensembles of multiple models allows for a more comprehensive analysis and identification of these complex weather systems.
Our analysis suggests that severe convective cloud identification is well suited to ensembles of multiple lightweight neural networks, though such studies are currently scarce. Firstly, the problem of severe convective cloud identification covers a broader feature space, with characteristics that may vary across different temporal and spatial ranges. Lightweight networks can be optimized for specific features or sections of an image, and integrating various networks can cover a wider array of characteristics comprehensively. For instance, ENet (Figure 7h) exhibits numerous false negatives, evidenced by extensive green areas, indicating a conservative predictive approach with high specificity but reduced sensitivity. This results in fewer false positives, demonstrating ENet’s tendency to miss certain convective cloud formations that are subtle or complex. In contrast, Fast-SCNN (Figure 7j) shows a liberal predictive behavior with a higher number of false positives (extensive red areas), indicating high sensitivity but lower specificity. This model identifies a broader range of features as convective clouds, increasing the likelihood of misclassification. The ensembling of these models into EFNet (Figure 7e) effectively balances these characteristics, reducing both false positives and false negatives. This demonstrates the power of model ensembling to enhance predictive accuracy and achieve a more balanced performance by leveraging the complementary strengths of Enet’s precision and Fast-SCNN’s recall. This approach illustrates the broader utility of model ensembles in machine learning to provide a comprehensive view of complex phenomena.
Secondly, ensembling multiple lightweight neural networks can enhance computational efficiency while maintaining model accuracy. Lightweight neural networks typically require fewer computational resources and storage space, making deployment feasible in resource-constrained environments such as direct satellite processing or real-time monitoring systems. Additionally, the parallel processing capabilities of multiple lightweight networks further enhance efficiency. The ensembling also reduces the risk of overfitting, as lightweight models, with fewer parameters compared to large deep networks, are less prone to overfitting. Ensembling multiple lightweight models can preserve the generalization capabilities of the model while improving prediction stability. Lastly, it enhances the robustness of the model; in the face of varying meteorological conditions and environmental changes, a single model may struggle to adapt to all scenarios. Model ensembling, by combining the predictions of multiple models, can provide more robust predictions when faced with unknown or anomalous data.
This paper introduces the EFNet model based on model ensembling methods and an ablation study. This model achieves the aforementioned advantages, ensuring that the quality of severe convective cloud identification surpasses that of U-Net and DeepLabV3 while attaining greater computational efficiency with a smaller model and fewer parameters.

6. Conclusions

The implementation of an ensemble of lightweight neural networks for the detection of severe convective clouds has demonstrated significant advancements in meteorological imaging applications, with the computational time, test loss, accuracy, recall, F1, IoU, and OP surpassing U-Net and DeepLabV3. Based on FY-4B datasets, we constructed a severe convective cloud dataset that supports the deep learning of severe convective cloud segmentation. Our methodology effectively combines the rapid processing abilities of lightweight models with the robustness required for accurate weather forecasting. The ensemble model consistently outperformed traditional, heavier models across several key performance indicators, achieving an accuracy of 0.9941, precision of 0.9391, recall of 0.9201, F1 score of 0.9295, IoU of 0.8684, and computing time of 18.65 s over the test dataset of 300 images. These metrics underscore the efficacy of our approach in maintaining high accuracy while reducing computational demands, making it potential for real-time monitoring and analysis.
This research moves us closer to more dynamic and precise meteorological service systems capable of better predicting and mitigating the impacts of severe weather conditions. The dynamic mechanisms are difficult for a single neural network to fully grasp; only a multi-model ensemble can more easily extract them. However, the multi-model ensemble of non-lightweight networks always faces the problem of low computational efficiency. Here, we demonstrated that the ensembles of several lightweight models can achieve slightly better results using less computing time and power, which means the ensemble of lightweight networks could resolve this issue to some extent. Future work will focus on further refining the ensemble techniques and exploring their applicability to other types of satellite data and environmental monitoring tasks, potentially enhancing the predictive capabilities and operational efficiency of weather services globally.

Author Contributions

Conceptualization, J.Z. and M.H.; methodology, J.Z.; software, J.Z.; validation, J.Z. and M.H.; formal analysis, J.Z.; investigation, J.Z.; resources, J.Z.; data curation, M.H.; writing—original draft preparation, J.Z.; writing—review and editing, J.Z.; visualization, J.Z.; supervision, M.H.; project administration, M.H.; funding acquisition, M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The FY-4B data used are available and can be freely downloaded at https://satellite.nsmc.org.cn/PortalSite/Data/Satellite.aspx (accessed on 10 April 2024). The FY-4B lookup table can be freely downloaded at http://www.nsmc.org.cn/nsmc/cn/satellite/FY4B.html (accessed on 10 April 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ma, R.; Sun, J.; Yang, X. An eight-year climatology of the warm-season severe thunderstorm environments over North China. Atmos. Res. 2021, 254, 105519. [Google Scholar] [CrossRef]
  2. Hang, R.; Wang, J.; Ge, L.; Shi, C.; Wei, J. Convective Cloud Detection From Himawari-8 Advanced Himawari Imager Data Using a Dual-Branch Deformable Convolutional Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 7490–7500. [Google Scholar] [CrossRef]
  3. Tian, Y.; Pang, S.; Qu, Y. Fusion Cloud Detection of Multiple Network Models Based on Hard Voting Strategy. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 6646–6649. [Google Scholar]
  4. Li, T.; Wu, D.; Wang, L.; Yu, X. Recognition algorithm for deep convective clouds based on FY4A. Neural Comput. Appl. 2022, 34, 21067–21088. [Google Scholar] [CrossRef]
  5. Li, L.; Li, X.; Jiang, L.; Su, X.; Chen, F. A review on deep learning techniques for cloud detection methodologies and challenges. Signal Image Video Process. 2021, 15, 1527–1535. [Google Scholar] [CrossRef]
  6. Han, L.; Sun, J.; Zhang, W. Convolutional neural network for convective storm nowcasting using 3-D Doppler weather radar data. IEEE Trans. Geosci. Remote Sens. 2019, 58, 1487–1495. [Google Scholar] [CrossRef]
  7. Han, D.; Lee, J.; Im, J.; Sim, S.; Lee, S.; Han, H.J. A novel framework of detecting convective initiation combining automated sampling, machine learning, and repeated model tuning from geostationary satellite data. Remote Sens. 2019, 11, 1454. [Google Scholar] [CrossRef]
  8. Chen, Q.; Yin, X.; Li, Y.; Zheng, P.; Chen, M.; Xu, Q. Recognition of Severe Convective Cloud Based on the Cloud Image Prediction Sequence from FY-4A. Remote Sens. 2023, 15, 4612. [Google Scholar] [CrossRef]
  9. Bai, C.; Zhang, M.; Zhang, J.; Zheng, J.; Chen, S. LSCIDMR: Large-scale satellite cloud image database for meteorological research. IEEE Trans. Cybern. 2021, 52, 12538–12550. [Google Scholar] [CrossRef] [PubMed]
  10. Fu, Y.; Mi, X.; Han, Z.; Zhang, W.; Liu, Q.; Gu, X.; Yu, T. A Machine-Learning-Based Study on All-Day Cloud Classification Using Himawari-8 Infrared Data. Remote Sens. 2023, 15, 5630. [Google Scholar] [CrossRef]
  11. Kai, Z.; Jiansheng, L.; Jianfeng, Y.; Wen, O.; Gaojie, W.; Xun, Z. A cloud and snow detection method of TH-1 image based on combined ResNet and DeeplabV3+. Acta Geod. Cartogr. Sin. 2020, 49, 1343. [Google Scholar]
  12. Lee, Y.; Kummerow, C.D.; Ebert-Uphoff, I. Applying machine learning methods to detect convection using GOES-16 ABI data. Atmos. Meas. Techn. Discuss. 2020, 2020, 1–28. [Google Scholar]
  13. Li, W.; Zhang, F.; Lin, H.; Chen, X.; Li, J.; Han, W. Cloud detection and classification algorithms for Himawari-8 imager measurements based on deep learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
  14. Ge, W.; Yang, X.; Jiang, R.; Shao, W.; Zhang, L.; Sensing, R. CD-CTFM: A Lightweight CNN-Transformer Network for Remote Sensing Cloud Detection Fusing Multiscale Features. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 4538–4551. [Google Scholar] [CrossRef]
  15. Liu, Q.; Li, Y.; Yu, M.; Chiu, L.S.; Hao, X.; Duffy, D.Q.; Yang, C.J. Daytime rainy cloud detection and convective precipitation delineation based on a deep neural Network method using GOES-16 ABI images. Remote Sens. 2019, 11, 2555. [Google Scholar] [CrossRef]
  16. Ukkonen, P.; Mäkelä, A. Evaluation of machine learning classifiers for predicting deep convection. J. Adv. Model. Earth Syst. 2019, 11, 1784–1802. [Google Scholar] [CrossRef]
  17. Gong, C.; Long, T.; Yin, R.; Jiao, W.; Wang, G. A Hybrid Algorithm with Swin Transformer and Convolution for Cloud Detection. Remote Sens. 2023, 15, 5264. [Google Scholar] [CrossRef]
  18. Guo, B.; Zhang, F.; Li, W.; Zhao, Z.; Sensing, R. Cloud Classification by machine learning for Geostationary Radiation Imager. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4102814. [Google Scholar] [CrossRef]
  19. Ma, N.; Sun, L.; Wang, Q.; Yu, Z.; Liu, S. Improved cloud detection for Landsat 8 images using a combined neural network model. Remote Sens. Lett. 2020, 11, 274–282. [Google Scholar] [CrossRef]
  20. Xu, C.; Geng, S.; Wang, D.; Zhou, M. Cloud detection of space-borne video remote sensing using improved Unet method. In Proceedings of the International Conference on Algorithms, High Performance Computing, and Artificial Intelligence (AHPCAI 2021), Sanya, China, 19–21 November 2021; pp. 297–303. [Google Scholar]
  21. Zhang, Z.; Iwasaki, A.; Xu, G.; Song, J. Cloud detection on small satellites based on lightweight U-net and image compression. J. Appl. Remote Sens. 2019, 13, 026502. [Google Scholar] [CrossRef]
  22. Zhou, K.; Zheng, Y.; Dong, W.; Wang, T. A deep learning network for cloud-to-ground lightning nowcasting with multisource data. J. Atmos. Ocean. Technol. 2020, 37, 927–942. [Google Scholar] [CrossRef]
  23. Molina, M.J.; Gagne, D.J.; Prein, A.F. A benchmark to test generalization capabilities of deep learning methods to classify severe convective storms in a changing climate. Earth Space Sci. 2021, 8, e2020EA001490. [Google Scholar] [CrossRef]
  24. Rumapea, H.; Zarlis, M.; Efendy, S.; Sihombing, P. Improving Convective Cloud Classification with Deep Learning: The CC-Unet Model. Int. J. Adv. Sci. Eng. Inf. Technol. 2024, 14, 28. [Google Scholar] [CrossRef]
  25. Yang, K.; Wang, Z.; Deng, M.; Dettmann, B. Improved tropical deep convective cloud detection using MODIS observations with an active sensor trained machine learning algorithm. Remote Sens. Environ. 2023, 297, 113762. [Google Scholar] [CrossRef]
  26. Yang, Y.; Zhao, C.; Sun, Y.; Chi, Y.; Fan, H. Convective Cloud Detection and Tracking Using the New-Generation Geostationary Satellite Over South China. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4103912. [Google Scholar] [CrossRef]
  27. Zhang, X.; Wang, T.; Chen, G.; Tan, X.; Zhu, K. Convective clouds extraction from Himawari–8 satellite images based on double-stream fully convolutional networks. IEEE Geosci. Remote Sens. Lett. 2019, 17, 553–557. [Google Scholar] [CrossRef]
  28. Paszke, A.; Chaurasia, A.; Kim, S.; Culurciello, E. Enet: A deep neural network architecture for real-time semantic segmentation. arXiv 2016, arXiv:1606.02147. [Google Scholar]
  29. Mehta, S.; Rastegari, M.; Caspi, A.; Shapiro, L.; Hajishirzi, H. Espnet: Efficient spatial pyramid of dilated convolutions for semantic segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 552–568. [Google Scholar]
  30. Poudel, R.P.; Liwicki, S.; Cipolla, R. Fast-scnn: Fast semantic segmentation network. arXiv 2019, arXiv:1902.04502. [Google Scholar]
  31. Zhao, H.; Qi, X.; Shen, X.; Shi, J.; Jia, J. Icnet for real-time semantic segmentation on high-resolution images. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 405–420. [Google Scholar]
  32. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  33. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. pp. 234–241. [Google Scholar]
  34. Yurtkulu, S.C.; Şahin, Y.H.; Unal, G. Semantic segmentation with extended DeepLabv3 architecture. In Proceedings of the 2019 27th Signal Processing and Communications Applications Conference (SIU), Sivas, Turkey, 24–26 April 2019; pp. 1–4. [Google Scholar]
  35. Wang, C.-H.; Huang, K.-Y.; Yao, Y.; Chen, J.-C.; Shuai, H.-H.; Cheng, W.-H. Lightweight deep learning: An overview. IEEE Consum. Electron. Mag. 2022, 99, 1–12. [Google Scholar] [CrossRef]
  36. Colomina, I.; Molina, P. Unmanned aerial systems for photogrammetry and remote sensing: A review. ISPRS J. Photogramm. Remote Sens. 2014, 92, 79–97. [Google Scholar] [CrossRef]
  37. Schenk, T. Introduction to Photogrammetry; The Ohio State University: Columbus, OH, USA, 2005; Volume 106. [Google Scholar]
  38. Seiz, G.; Shields, J.; Feister, U.; Baltsavias, E.P.; Gruen, A. Cloud mapping with ground-based photogrammetric cameras. Int. J. Remote Sens. 2007, 28, 2001–2032. [Google Scholar] [CrossRef]
  39. Hammouti, M.; Gencarelli, C.N.; Sterlacchini, S.; Biondi, R. Volcanic clouds detection applying machine learning techniques to GNSS radio occultations. GPS Solut. 2024, 28, 116. [Google Scholar] [CrossRef]
  40. Kaplan, E.D.; Hegarty, C. Understanding GPS/GNSS: Principles and Applications; Artech House: London, UK, 2017. [Google Scholar]
  41. Farooq, B.; Manocha, A. Satellite-based change detection in multi-objective scenarios: A comprehensive review. Remote Sens. Appl. Soc. Environ. 2024, 34, 101168. [Google Scholar] [CrossRef]
  42. Abdollahi, A.; Pradhan, B.; Alamri, A.M. An ensemble architecture of deep convolutional Segnet and Unet networks for building semantic segmentation from high-resolution aerial images. Geocarto Int. 2022, 37, 3355–3370. [Google Scholar] [CrossRef]
  43. O’Donncha, F.; Zhang, Y.; Chen, B.; James, S.C. Ensemble model aggregation using a computationally lightweight machine-learning model to forecast ocean waves. J. Mar. Syst. 2019, 199, 103206. [Google Scholar] [CrossRef]
  44. Kühnlein, M.; Appelhans, T.; Thies, B.; Nauss, T. Improving the accuracy of rainfall rates from optical satellite sensors with machine learning—A random forests-based approach applied to MSG SEVIRI. Remote Sens. Environ. 2014, 141, 129–143. [Google Scholar] [CrossRef]
  45. Qian, Z.; Wang, D.; Shi, X.; Yao, J.; Hu, L.; Yang, H.; Ni, Y.J. Lightning Identification Method Based on Deep Learning. Atmosphere 2022, 13, 2112. [Google Scholar] [CrossRef]
  46. Hu, K.; Zhang, D.; Xia, M.J. CDUNet: Cloud detection UNet for remote sensing imagery. Remote Sens. 2021, 13, 4533. [Google Scholar] [CrossRef]
  47. Hu, K.; Zhang, D.; Xia, M.; Qian, M.; Chen, B.J. LCDNet: Light-weighted cloud detection network for high-resolution remote sensing images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 4809–4823. [Google Scholar] [CrossRef]
  48. Luo, C.; Feng, S.; Yang, X.; Ye, Y.; Li, X.; Zhang, B.; Chen, Z.; Quan, Y. LWCDnet: A lightweight network for efficient cloud detection in remote sensing images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  49. Qian, J.; Ci, J.; Tan, H.; Xu, W.; Jiao, Y.; Chen, P. Cloud Detection Method Based on Improved DeeplabV3+ Remote Sensing Image. IEEE Access 2024, 12, 9229–9242. [Google Scholar] [CrossRef]
  50. Yao, X.; Guo, Q.; Li, A.J. Light-weight cloud detection network for optical remote sensing images with attention-based deeplabv3+ architecture. Remote Sens. 2021, 13, 3617. [Google Scholar] [CrossRef]
  51. Chang, S.; Li, Y.; Shi, C.; Guo, D. Combined Effects of the ENSO and the QBO on the Ozone Valley over the Tibetan Plateau. Remote Sens. 2022, 14, 4935. [Google Scholar] [CrossRef]
  52. England, S.L.; Liu, G.; Withers, P.; Yiğit, E.; Lo, D.; Jain, S.; Schneider, N.M.; Deighan, J.; McClintock, W.E.; Mahaffy, P.R.J.J.o.G.R.P. Simultaneous observations of atmospheric tides from combined in situ and remote observations at Mars from the MAVEN spacecraft. J. Geophys. Res. Planets 2016, 121, 594–607. [Google Scholar] [CrossRef]
  53. Gao, J.W.; Rong, Z.J.; Klinger, L.; Li, X.Z.; Liu, D.; Wei, Y. A Spherical Harmonic Martian Crustal Magnetic Field Model Combining Data Sets of MAVEN and MGS. Earth Space Sci. 2021, 8, e2021EA001860. [Google Scholar] [CrossRef]
  54. Chen, B.-F.; Kuo, Y.-T.; Huang, T.-S. A deep learning ensemble approach for predicting tropical cyclone rapid intensification. Atmos. Sci. Lett. 2023, 24, e1151. [Google Scholar] [CrossRef]
  55. Singla, P.; Duhan, M.; Saroha, S. An ensemble method to forecast 24-h ahead solar irradiance using wavelet decomposition and BiLSTM deep learning network. Earth Sci. Inform. 2022, 15, 291–306. [Google Scholar] [CrossRef] [PubMed]
  56. Lin, C.-Y.; Chang, Y.-S.; Abimannan, S. Ensemble multifeatured deep learning models for air quality forecasting. Atmos. Pollut. Res. 2021, 12, 101045. [Google Scholar] [CrossRef]
  57. Guastavino, S.; Piana, M.; Tizzi, M.; Cassola, F.; Iengo, A.; Sacchetti, D.; Solazzo, E.; Benvenuto, F. Prediction of severe thunderstorm events with ensemble deep learning and radar data. Sci. Rep. 2022, 12, 20049. [Google Scholar] [CrossRef] [PubMed]
  58. Sha, Y.; Sobash, R.A.; Gagne, D.J. Generative ensemble deep learning severe weather prediction from a deterministic convection-allowing model. Artif. Intell. Earth Syst. 2024, 3, e230094. [Google Scholar] [CrossRef]
  59. Lei, B.; Yang, L.; Xu, Z. Using convolutional neural network to classify convective cloud on radar echoes. In Proceedings of the 2019 International Conference on Meteorology Observations (ICMO), Chengdu, China, 28–31 December 2019; pp. 1–3. [Google Scholar]
  60. Zhou, K.; Zheng, Y.; Li, B.; Dong, W.; Zhang, X. Forecasting different types of convective weather: A deep learning approach. J. Meteorol. Res. 2019, 33, 797–809. [Google Scholar] [CrossRef]
  61. Ganaie, M.A.; Hu, M.; Malik, A.K.; Tanveer, M.; Suganthan, P.N. Ensemble deep learning: A review. Eng. Appl. Artif. Intell. 2022, 115, 105151. [Google Scholar] [CrossRef]
  62. Zhang, P.; Xu, Z.; Guan, M.; Xie, L.; Xian, D.; Liu, C. Progress of Fengyun Meteorological Satellites Since 2020. Chin. J. Space Sci. 2022, 42, 724–732. [Google Scholar] [CrossRef]
  63. Zhu, Z.; Shi, C.; Gu, J. Characterization of bias in Fengyun-4B/AGRI infrared observations using RTTOV. Remote Sens. 2023, 15, 1224. [Google Scholar] [CrossRef]
  64. Ma, Z.; Zhu, S.; Yang, J. FY4QPE-MSA: An all-day near-real-time quantitative precipitation estimation framework based on multispectral analysis from AGRI onboard Chinese FY-4 series satellites. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  65. Zhu, S.; Ma, Z. Does AGRI of FY4A have the ability to capture the motions of precipitation? IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  66. Xu, W. Precipitation and convective characteristics of summer deep convection over East Asia observed by TRMM. Mon. Weather Rev. 2013, 141, 1577–1592. [Google Scholar] [CrossRef]
  67. Murakami, M. Analysis of the deep convective activity over the Western Pacific and Southeast Asia part II: Seasonal and intraseasonal variations during Northern Summer. J. Meteorol. Soc. Jpn. 1984, 62, 88–108. [Google Scholar] [CrossRef]
  68. Su, B.; Chen, A.; Liu, M.; Kong, L.; Zhang, A.; Tian, Z.; Liu, B.; Wang, X.; Wang, W.; Zhang, X.; et al. Ground Calibration and In-Flight Performance of the Low Energy Particle Analyzer on FY-4B. Atmosphere 2023, 14, 1834. [Google Scholar] [CrossRef]
  69. Huang, Y.; Bao, Y.; Petropoulos, G.P.; Lu, Q.; Huo, Y.; Wang, F.J. Precipitation Estimation Using FY-4B/AGRI Satellite Data Based on Random Forest. Remote Sens. 2024, 16, 1267. [Google Scholar] [CrossRef]
  70. Li, X.; Cao, Q.; Zhou, S.; Qian, J.; Wang, B.; Zou, Y.; Wang, J.; Shen, X.; Han, C.; Wang, L.; et al. Prelaunch Radiometric Characterization and Calibration for Long Wave Infrared Band of FY-4B GHI. Acta Opt. Sin. 2023, 43, 1212005. [Google Scholar]
  71. Wenqiang, L.; Yi, Q.; Lei, Y.; Yong, H. Analysis of the ranging systematic error of the FY-4 geostationary satellite and its influence on orbit determination. Chin. J. Sci. Instrum. 2022, 43, 73–83. [Google Scholar]
  72. Zhang, J.; Liu, P.; Zhang, F.; Song, Q. CloudNet: Ground-based cloud classification with deep convolutional neural network. Geophys. Res. Lett. 2018, 45, 8665–8672. [Google Scholar] [CrossRef]
  73. Li, S.; Wang, M.; Sun, S.; Wu, J.; Zhuang, Z. CloudDenseNet: Lightweight Ground-Based Cloud Classification Method for Large-Scale Datasets Based on Reconstructed DenseNet. Sensors 2023, 23, 7957. [Google Scholar] [CrossRef] [PubMed]
  74. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 6, 679–698. [Google Scholar] [CrossRef]
  75. Deriche, R. Using Canny’s criteria to derive a recursively implemented optimal edge detector. Int. J. Comput. Vis. 1987, 1, 167–187. [Google Scholar] [CrossRef]
  76. Shih, F.Y.; Cheng, S. Automatic seeded region growing for color image segmentation. Image Vis. Comput. 2005, 23, 877–886. [Google Scholar] [CrossRef]
  77. Bao, P.; Zhang, L.; Wu, X. Canny edge detection enhancement by scale multiplication. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1485–1490. [Google Scholar] [CrossRef] [PubMed]
  78. Lapušinskij, A.; Suzdalev, I.; Goranin, N.; Janulevičius, J.; Ramanauskaitė, S.; Stankūnavičius, G. The application of Hough transform and Canny edge detector methods for the visual detection of cumuliform clouds. Sensors 2021, 21, 5821. [Google Scholar] [CrossRef] [PubMed]
  79. Ansel, J.; Yang, E.; He, H.; Gimelshein, N.; Jain, A.; Voznesensky, M.; Bao, B.; Bell, P.; Berard, D.; Burovski, E. PyTorch 2: Faster Machine Learning Through Dynamic Python Bytecode Transformation and Graph Compilation. In Proceedings of the 29th ACM International Conference on Architectural Support for Programming Languages and Operating Systems (ASPLOS’24), La Jolla, CA, USA, 27 April–1 May 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 317–335. [Google Scholar]
  80. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  81. De Boer, P.-T.; Kroese, D.P.; Mannor, S.; Rubinstein, R.Y. A tutorial on the cross-entropy method. Ann. Oper. Res. 2005, 134, 19–67. [Google Scholar] [CrossRef]
Figure 1. Research area located within 0–45°N, 100–160°E (red dashed box). In the shaded relief image on the map, the colors correspond to elevation.
Figure 1. Research area located within 0–45°N, 100–160°E (red dashed box). In the shaded relief image on the map, the colors correspond to elevation.
Remotesensing 16 02070 g001
Figure 2. Severe convective cloud dataset example. The edges of the severe convective cloud clusters are shown by blue lines.
Figure 2. Severe convective cloud dataset example. The edges of the severe convective cloud clusters are shown by blue lines.
Remotesensing 16 02070 g002
Figure 3. Methodology framework for severe convective cloud identification using lightweight neural network model ensembling. In the three-channel example figure, the colors correspond to grayscale values; in the Ground Truth figure, the colors represent pixels labeled as strong convective clouds; in the sample result figure, colors signify prediction accuracy: green for severe convective clouds incorrectly predicted as non-severe, red for non-severe clouds incorrectly predicted as severe deep convection, white for correctly predicted severe convective clouds, and black for correctly predicted non-severe convective clouds.
Figure 3. Methodology framework for severe convective cloud identification using lightweight neural network model ensembling. In the three-channel example figure, the colors correspond to grayscale values; in the Ground Truth figure, the colors represent pixels labeled as strong convective clouds; in the sample result figure, colors signify prediction accuracy: green for severe convective clouds incorrectly predicted as non-severe, red for non-severe clouds incorrectly predicted as severe deep convection, white for correctly predicted severe convective clouds, and black for correctly predicted non-severe convective clouds.
Remotesensing 16 02070 g003
Figure 4. Training loss and validation loss.
Figure 4. Training loss and validation loss.
Remotesensing 16 02070 g004
Figure 5. Cloud prediction results of the five lightweight models (ENet, ESPNet, Fast-SCNN, ICNet, and MobineNetV2) and their model ensembling ( C 5,1 + C 5,2 = 15 kinds of single- and double-model combinations) in ablation study. Subfigures (ac) display results for channels 1–3 of the image at wavelengths of 6.25 μm, BTD (6.25–10.8 μm), and BTD (10.8–12 μm) respectively. Subfigure (d) presents the labeled data, and subfigures (es) show results from the following model configurations: ENet, ESPNet, Fast-SCNN, ICNet, MobileNetV2, ENet + ESPNet, ENet + Fast-SCNN, ENet + ICNet, ENet + MobileNetV2, ESPNet + Fast-SCNN, ESPNet + ICNet, ESPNet + MobileNetV2, Fast-SCNN + ICNet, Fast-SCNN + MobileNetV2, and ICNet + MobileNetV2, respectively. The ablation study of these 10 images took 46.79 s to complete, including calculation, plotting and saving. In subfigures (es), colors signify prediction accuracy: green for severe convective clouds incorrectly predicted as non-severe, red for non-severe clouds incorrectly predicted as severe deep convection, white for correctly predicted severe convective clouds, and black for correctly predicted non-severe convective clouds.
Figure 5. Cloud prediction results of the five lightweight models (ENet, ESPNet, Fast-SCNN, ICNet, and MobineNetV2) and their model ensembling ( C 5,1 + C 5,2 = 15 kinds of single- and double-model combinations) in ablation study. Subfigures (ac) display results for channels 1–3 of the image at wavelengths of 6.25 μm, BTD (6.25–10.8 μm), and BTD (10.8–12 μm) respectively. Subfigure (d) presents the labeled data, and subfigures (es) show results from the following model configurations: ENet, ESPNet, Fast-SCNN, ICNet, MobileNetV2, ENet + ESPNet, ENet + Fast-SCNN, ENet + ICNet, ENet + MobileNetV2, ESPNet + Fast-SCNN, ESPNet + ICNet, ESPNet + MobileNetV2, Fast-SCNN + ICNet, Fast-SCNN + MobileNetV2, and ICNet + MobileNetV2, respectively. The ablation study of these 10 images took 46.79 s to complete, including calculation, plotting and saving. In subfigures (es), colors signify prediction accuracy: green for severe convective clouds incorrectly predicted as non-severe, red for non-severe clouds incorrectly predicted as severe deep convection, white for correctly predicted severe convective clouds, and black for correctly predicted non-severe convective clouds.
Remotesensing 16 02070 g005
Figure 6. Same as Figure 5, but for results of the other model combinations ( C 5,3 + C 5,4 + C 5,5 = 16 kinds). Displayed (ap) configurations include ENet + ESPNet + Fast-SCNN, ENet + ESPNet + ICNet, ENet + ESPNet + MobileNetV2, ENet + Fast-SCNN + ICNet, ENet + Fast-SCNN + MobileNetV2, ENet + ICNet + MobileNetV2, ESPNet + Fast-SCNN + ICNet, ESPNet + Fast-SCNN + MobileNetV2, ESPNet + ICNet + MobileNetV2, Fast-SCNN + ICNet + MobileNetV2, ENet + ESPNet + Fast-SCNN + ICNet, ENet + ESPNet + Fast-SCNN + MobileNetV2, ENet + ESPNet + ICNet + MobileNetV2, ENet + Fast-SCNN + ICNet + MobileNetV2, ESPNet + Fast-SCNN + ICNet + MobileNetV2, and ENet + ESPNet + Fast-SCNN + ICNet + MobileNetV2, respectively. The ablation study of these configurations on 10 images took 46.67 s to complete, including calculations, plotting, and saving.
Figure 6. Same as Figure 5, but for results of the other model combinations ( C 5,3 + C 5,4 + C 5,5 = 16 kinds). Displayed (ap) configurations include ENet + ESPNet + Fast-SCNN, ENet + ESPNet + ICNet, ENet + ESPNet + MobileNetV2, ENet + Fast-SCNN + ICNet, ENet + Fast-SCNN + MobileNetV2, ENet + ICNet + MobileNetV2, ESPNet + Fast-SCNN + ICNet, ESPNet + Fast-SCNN + MobileNetV2, ESPNet + ICNet + MobileNetV2, Fast-SCNN + ICNet + MobileNetV2, ENet + ESPNet + Fast-SCNN + ICNet, ENet + ESPNet + Fast-SCNN + MobileNetV2, ENet + ESPNet + ICNet + MobileNetV2, ENet + Fast-SCNN + ICNet + MobileNetV2, ESPNet + Fast-SCNN + ICNet + MobileNetV2, and ENet + ESPNet + Fast-SCNN + ICNet + MobileNetV2, respectively. The ablation study of these configurations on 10 images took 46.67 s to complete, including calculations, plotting, and saving.
Remotesensing 16 02070 g006
Figure 7. Same as Figure 5, but for (ac) channels 1–3 of the image, (d) the label, and (el) results of EFNet, U-Net, DeepLabV3, ENet, ESPNet, Fast-SCNN, ICNet, and MobileNetV2, respectively. The test of 10 images took 29.92 s to run, including calculation, plotting, and saving.
Figure 7. Same as Figure 5, but for (ac) channels 1–3 of the image, (d) the label, and (el) results of EFNet, U-Net, DeepLabV3, ENet, ESPNet, Fast-SCNN, ICNet, and MobileNetV2, respectively. The test of 10 images took 29.92 s to run, including calculation, plotting, and saving.
Remotesensing 16 02070 g007
Table 1. FY-4B AGRI parameters by channel *.
Table 1. FY-4B AGRI parameters by channel *.
Channel TypeBandCenter Wavelength (μm)Bandwidth (μm)Spatial Resolution (km)Main Applications
VIS/NIR10.470.45–0.491Small-particle aerosols, true color synthesis
20.650.55–0.750.5Vegetation, image navigation registration, stellar observations
30.8250.75–0.901Vegetation, aerosols over water surfaces
Shortwave IR41.3791.371–1.3862Cirrus clouds
51.611.58–1.642Low cloud/snow identification, water/ice cloud discrimination
62.252.10–2.352Cirrus, aerosols, particle size
Midwave IR73.753.50–4.02Clouds and high albedo targets, fire spots
83.753.50–4.04Low albedo targets, surface
Water vapor96.255.80–6.704Upper-level water vapor
106.956.75–7.154Mid-level water vapor
117.427.24–7.604Low-level water vapor
Longwave IR128.558.3–8.84Clouds
1310.8010.30–11.304Clouds, surface temperature, etc.
1412.0011.50–12.504Clouds, total water vapor, surface temperature
1513.313.00–13.604Clouds, water vapor
* Available online: https://www.nsmc.org.cn/nsmc/cn/instrument/AGRI.html (accessed on 20 April 2024).
Table 2. Training details.
Table 2. Training details.
ModelAvg
Train
Time
(s/Epoch)
Avg
Val
Time
(s/Epoch)
Number of ParametersModel Size
(MB)
Batch SizeInitial Learning Rate
ENet75.2113.63214,4650.9580.001
ESPNet26.8012.99103,7840.4280.001
Fast-SCNN118.4513.521,112,3504.1280.0005
ICNet21.1113.4337410.0240.0005
MobileNetV2221.2415.653,363,60211.39160.0005
DeepLabV3536.4828.9141,999,209160.5640.0001
U-Net483.1618.577,383,23428.2240.001
Table 3. Performance parameters of 31 model combinations of ENet (E), ESPNet (ES), Fast-SCNN (F), ICNet (I), and MobileNetV2 (M). Results are arranged in descending order according to the OP metric.
Table 3. Performance parameters of 31 model combinations of ENet (E), ESPNet (ES), Fast-SCNN (F), ICNet (I), and MobileNetV2 (M). Results are arranged in descending order according to the OP metric.
Model CombinationTest
Time
(s)
Test
Loss
AccuracyPrecisionRecallF1IoUOP
E + F18.650.01560.99410.93910.92010.92950.86844.6512
E + F + M25.270.02020.9940.94420.91220.92790.86554.6438
E + ES + F + M24.950.02580.99390.96060.89110.92460.85974.6299
E + ES + F19.630.02150.99380.9590.890.92320.85744.6234
ES + F + M21.240.02150.99350.95380.88910.92030.85244.6105
E + F + I + M25.050.02580.99360.96280.88140.92030.85244.6091
E + M19.510.01840.99350.95530.88690.91990.85164.6072
ES + F17.230.01760.99330.95110.88710.9180.84854.5983
E + F + I20.760.02150.99340.96230.8770.91770.84794.598
E + ES + F + I + M25.760.03370.99320.97040.86590.91510.84364.5882
F + I + M21.560.02280.99280.95380.87260.91140.83724.5678
F + M20.730.02030.99250.90830.91390.91110.83674.5666
E + ES + M22.440.02650.99290.96730.86010.91050.83584.5646
E15.290.01730.99270.94370.87940.91050.83564.5625
E + ES + F + I20.670.030.99280.96980.85710.910.83494.5619
ES + F + I + M23.180.0290.99280.96940.8560.90920.83354.5609
F + I16.980.01990.99220.95030.85910.90240.82224.5281
E + I + M19.280.02710.99230.97120.84160.90180.82124.5262
ES + F + I18.090.02590.99210.96940.83940.89970.81774.5183
ES + M17.150.02250.9920.96560.84090.8990.81654.514
F18.370.02080.99120.86860.93130.89890.81644.5064
M16.720.01990.99150.91940.87620.89730.81374.4988
E + ES16.310.02520.99170.96120.83830.89560.81094.4981
E + ES + I + M20.590.0370.99180.97520.82690.8950.80994.4977
E + I15.860.02710.99050.96620.8030.87710.78114.4179
ES + I + M17.930.03380.99050.97710.79370.87590.77924.4164
I + M17.210.02570.99040.96790.79820.87490.77774.4091
E + ES + I17.870.03730.99030.97090.7930.8730.77464.4018
ES16.390.02720.98840.95020.76590.84810.73634.2889
ES + I14.510.04020.98630.96840.69790.81120.68244.1462
I11.950.04640.97860.91560.54420.68260.51823.6392
Table 4. Performance parameters of EFNet, U-Net, DeepLabV3, ENet (E), ESPNet (ES), Fast-SCNN (F), ICNet (I), and MobileNetV2 (M). Results are arranged in descending order according to the OP metric.
Table 4. Performance parameters of EFNet, U-Net, DeepLabV3, ENet (E), ESPNet (ES), Fast-SCNN (F), ICNet (I), and MobileNetV2 (M). Results are arranged in descending order according to the OP metric.
Model CombinationTest
Time
(s)
Test
Loss
AccuracyPrecisionRecallF1IoUOP
E + F18.650.01560.99410.93910.92010.92950.86844.6512
E15.290.01730.99270.94370.87940.91050.83564.5625
U-Net20.410.01730.99260.9550.8540.90170.82094.5242
F18.370.02080.99120.86860.93130.89890.81644.5064
M16.720.01990.99150.91940.87620.89730.81374.4988
DeepLabV343.150.02560.99120.94120.83230.88340.79114.4392
ES16.390.02720.98840.95020.76590.84810.73634.2889
I11.950.04640.97860.91560.54420.68260.51823.6392
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; He, M. Methodology for Severe Convective Cloud Identification Using Lightweight Neural Network Model Ensembling. Remote Sens. 2024, 16, 2070. https://doi.org/10.3390/rs16122070

AMA Style

Zhang J, He M. Methodology for Severe Convective Cloud Identification Using Lightweight Neural Network Model Ensembling. Remote Sensing. 2024; 16(12):2070. https://doi.org/10.3390/rs16122070

Chicago/Turabian Style

Zhang, Jie, and Mingyuan He. 2024. "Methodology for Severe Convective Cloud Identification Using Lightweight Neural Network Model Ensembling" Remote Sensing 16, no. 12: 2070. https://doi.org/10.3390/rs16122070

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop